Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
609,100 | 18,853,997,419 | IssuesEvent | 2021-11-12 02:10:37 | lokka30/Treasury | https://api.github.com/repos/lokka30/Treasury | closed | Address the difficulty of use with CompletableFutures, etc. | help wanted priority: high approved | Please see #37 which is an in-progress resolution to this issue. | 1.0 | Address the difficulty of use with CompletableFutures, etc. - Please see #37 which is an in-progress resolution to this issue. | priority | address the difficulty of use with completablefutures etc please see which is an in progress resolution to this issue | 1 |
73,830 | 3,421,786,014 | IssuesEvent | 2015-12-08 20:12:42 | tipsybear/actors-simulation | https://api.github.com/repos/tipsybear/actors-simulation | closed | Baseline Simulation | priority: high type: feature | Here is the baseline simulation that we need to show that our simulation model is realistic and serves as a proof of concept about how our cluster compares to the actor simulators.
The simulation is composed of programs (1 per processor) that do the following:
1. wait for a message
2. sleep for a random amount of time (work)
3. send message(s) of varying sizes to other nodes
The input (parameters of the simulation are as follows:
1. # of nodes that initially receive messages (to get things started)
2. (min, max) sleep time (amount of work()
3. (min, max) message size (This is the key as we're measuring latency in a fully utilized cluster.)
Output: a time series of the following aggregate information:
1. # of messages in flight
2. total (average?) message size in flight (e.g. the bandwidth capacity)
3. current average latency (which means we have to measure how long it takes a message to arrive)
4. available bandwidth though this might be the same as 2.
The visualization of this time series will be reported in the presentation. | 1.0 | Baseline Simulation - Here is the baseline simulation that we need to show that our simulation model is realistic and serves as a proof of concept about how our cluster compares to the actor simulators.
The simulation is composed of programs (1 per processor) that do the following:
1. wait for a message
2. sleep for a random amount of time (work)
3. send message(s) of varying sizes to other nodes
The input (parameters of the simulation are as follows:
1. # of nodes that initially receive messages (to get things started)
2. (min, max) sleep time (amount of work()
3. (min, max) message size (This is the key as we're measuring latency in a fully utilized cluster.)
Output: a time series of the following aggregate information:
1. # of messages in flight
2. total (average?) message size in flight (e.g. the bandwidth capacity)
3. current average latency (which means we have to measure how long it takes a message to arrive)
4. available bandwidth though this might be the same as 2.
The visualization of this time series will be reported in the presentation. | priority | baseline simulation here is the baseline simulation that we need to show that our simulation model is realistic and serves as a proof of concept about how our cluster compares to the actor simulators the simulation is composed of programs per processor that do the following wait for a message sleep for a random amount of time work send message s of varying sizes to other nodes the input parameters of the simulation are as follows of nodes that initially receive messages to get things started min max sleep time amount of work min max message size this is the key as we re measuring latency in a fully utilized cluster output a time series of the following aggregate information of messages in flight total average message size in flight e g the bandwidth capacity current average latency which means we have to measure how long it takes a message to arrive available bandwidth though this might be the same as the visualization of this time series will be reported in the presentation | 1 |
378,719 | 11,206,730,695 | IssuesEvent | 2020-01-05 23:34:39 | akiraux/Akira | https://api.github.com/repos/akiraux/Akira | closed | Split the Canvas | Not an Issue Priority: High | The `Canvas.vala` file is growing exponentially at every PR. Since that's the main core of our application, we should define how to divide it into multiple classes in order to keep our core library as scalable and modular as possible.
Let's throw a bunch of ideas on how to approach this, and open some draft PRs to prototype solutions. | 1.0 | Split the Canvas - The `Canvas.vala` file is growing exponentially at every PR. Since that's the main core of our application, we should define how to divide it into multiple classes in order to keep our core library as scalable and modular as possible.
Let's throw a bunch of ideas on how to approach this, and open some draft PRs to prototype solutions. | priority | split the canvas the canvas vala file is growing exponentially at every pr since that s the main core of our application we should define how to divide it into multiple classes in order to keep our core library as scalable and modular as possible let s throw a bunch of ideas on how to approach this and open some draft prs to prototype solutions | 1 |
414,168 | 12,100,100,661 | IssuesEvent | 2020-04-20 13:18:13 | cilium/cilium | https://api.github.com/repos/cilium/cilium | closed | clustermesh issue: io.cilium/global-service=true is not working for an existing service | area/clustermesh kind/community-report priority/high | <!--
If you have usage questions, please try the [slack
channel](http://cilium.io/slack) and see the [FAQ](https://goo.gl/qG2YmU)
first.
Choose either "Proposal" or "Bug report"
-->
## Bug report
Here's what I did
1. I took your clustermesh/rebel-base example and downloaded your cluster1.yml and cluster2.yml.
2. Removed the "io.cilium/global-service=true" from the rebel-base service.
3. Renamed rebel-base to bipin-base as I was having your original service running to test the environment.
alias kl='kubectl config get-contexts'
alias ka='kubectl config use-context a.xxxxxxxxxxxx'
alias kb='kubectl config use-context b.xxxxxxxxxxx'
alias ks='kubectl -n kube-system'
alias k='kubectl'
ka
k apply -f k8s-yaml/nginx/cluster1.yml
kb
k apply -f k8s-yaml/nginx/cluster2.yml
ka
kubectl annotate svc bipin-base io.cilium/global-service=true
kb
kubectl annotate svc bipin-base io.cilium/global-service=true
k get svc
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
bipin-base ClusterIP 100.64.9.182 <none> 80/TCP 10m
kubernetes ClusterIP 100.64.0.1 <none> 443/TCP 4h
ks exec cilium-cq9vw -- cilium service list
ID Frontend Backend
1 100.64.0.1:443 1 => 10.0.1.206:443
3 100.64.0.10:53 1 => 100.64.129.212:53
2 => 100.64.129.245:53
4 100.64.26.37:2379 1 => 100.64.129.101:2379
2 => 100.64.129.193:2379
3 => 100.64.129.27:2379
5 100.64.18.136:2379 1 => 100.64.129.101:2379
2 => 100.64.129.193:2379
3 => 100.64.129.27:2379
8 100.64.9.182:80 1 => 100.64.129.102:80
2 => 100.64.129.128:80
Please look at the service-id=8
It show only the local ips.
The cluster1 and cluster2 YAML files which came with your examples works file.
That proves that this global-service annotation is not being picked for an existing service.
As a workaround if you can tell me how to update the service definition manually for time being for demo'ing this to team...it would help greatly
<!--
Important: For security related issues: We strongly encourage you to report
security vulnerabilities to our private security mailing list:
security@cilium.io - first, before disclosing them in any public forums.
-->
**General Information**
- Cilium version (run `cilium version`)
ks exec cilium-cq9vw -- cilium version
Client: 1.4.90 3adbd7f 2019-02-12T19:27:16-08:00 go version go1.11.1 linux/amd64
Daemon: 1.4.90 3adbd7f 2019-02-12T19:27:16-08:00 go version go1.11.1 linux/amd64
- Kernel version (run `uname -a`)
4.14.96-coreos
- Orchestration system version in use (e.g. `kubectl version`, Mesos, ...)
kubectl version
Client Version: version.Info{Major:"1", Minor:"12", GitVersion:"v1.12.2", GitCommit:"17c77c7898218073f14c8d573582e8d2313dc740", GitTreeState:"clean", BuildDate:"2018-10-24T06:54:59Z", GoVersion:"go1.10.4", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.6", GitCommit:"b1d75deca493a24a2f87eb1efde1a569e52fc8d9", GitTreeState:"clean", BuildDate:"2018-12-16T04:30:10Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"}
- Link to relevant artifacts (policies, deployments scripts, ...)
[nginx.zip](https://github.com/cilium/cilium/files/2860741/nginx.zip)
- Upload a system dump (run `curl -sLO
[cluster-diagnosis.zip](https://github.com/cilium/cilium/files/2860756/cluster-diagnosis.zip)
releases.cilium.io/tools/cluster-diagnosis.zip &&
python cluster-diagnosis.zip sysdump` and then attach the generated zip file)
[cluster-diagnosis.zip](https://github.com/cilium/cilium/files/2860757/cluster-diagnosis.zip)
**How to reproduce the issue**
1. instruction 1
2. instruction 2
| 1.0 | clustermesh issue: io.cilium/global-service=true is not working for an existing service - <!--
If you have usage questions, please try the [slack
channel](http://cilium.io/slack) and see the [FAQ](https://goo.gl/qG2YmU)
first.
Choose either "Proposal" or "Bug report"
-->
## Bug report
Here's what I did
1. I took your clustermesh/rebel-base example and downloaded your cluster1.yml and cluster2.yml.
2. Removed the "io.cilium/global-service=true" from the rebel-base service.
3. Renamed rebel-base to bipin-base as I was having your original service running to test the environment.
alias kl='kubectl config get-contexts'
alias ka='kubectl config use-context a.xxxxxxxxxxxx'
alias kb='kubectl config use-context b.xxxxxxxxxxx'
alias ks='kubectl -n kube-system'
alias k='kubectl'
ka
k apply -f k8s-yaml/nginx/cluster1.yml
kb
k apply -f k8s-yaml/nginx/cluster2.yml
ka
kubectl annotate svc bipin-base io.cilium/global-service=true
kb
kubectl annotate svc bipin-base io.cilium/global-service=true
k get svc
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
bipin-base ClusterIP 100.64.9.182 <none> 80/TCP 10m
kubernetes ClusterIP 100.64.0.1 <none> 443/TCP 4h
ks exec cilium-cq9vw -- cilium service list
ID Frontend Backend
1 100.64.0.1:443 1 => 10.0.1.206:443
3 100.64.0.10:53 1 => 100.64.129.212:53
2 => 100.64.129.245:53
4 100.64.26.37:2379 1 => 100.64.129.101:2379
2 => 100.64.129.193:2379
3 => 100.64.129.27:2379
5 100.64.18.136:2379 1 => 100.64.129.101:2379
2 => 100.64.129.193:2379
3 => 100.64.129.27:2379
8 100.64.9.182:80 1 => 100.64.129.102:80
2 => 100.64.129.128:80
Please look at the service-id=8
It show only the local ips.
The cluster1 and cluster2 YAML files which came with your examples works file.
That proves that this global-service annotation is not being picked for an existing service.
As a workaround if you can tell me how to update the service definition manually for time being for demo'ing this to team...it would help greatly
<!--
Important: For security related issues: We strongly encourage you to report
security vulnerabilities to our private security mailing list:
security@cilium.io - first, before disclosing them in any public forums.
-->
**General Information**
- Cilium version (run `cilium version`)
ks exec cilium-cq9vw -- cilium version
Client: 1.4.90 3adbd7f 2019-02-12T19:27:16-08:00 go version go1.11.1 linux/amd64
Daemon: 1.4.90 3adbd7f 2019-02-12T19:27:16-08:00 go version go1.11.1 linux/amd64
- Kernel version (run `uname -a`)
4.14.96-coreos
- Orchestration system version in use (e.g. `kubectl version`, Mesos, ...)
kubectl version
Client Version: version.Info{Major:"1", Minor:"12", GitVersion:"v1.12.2", GitCommit:"17c77c7898218073f14c8d573582e8d2313dc740", GitTreeState:"clean", BuildDate:"2018-10-24T06:54:59Z", GoVersion:"go1.10.4", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.6", GitCommit:"b1d75deca493a24a2f87eb1efde1a569e52fc8d9", GitTreeState:"clean", BuildDate:"2018-12-16T04:30:10Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"}
- Link to relevant artifacts (policies, deployments scripts, ...)
[nginx.zip](https://github.com/cilium/cilium/files/2860741/nginx.zip)
- Upload a system dump (run `curl -sLO
[cluster-diagnosis.zip](https://github.com/cilium/cilium/files/2860756/cluster-diagnosis.zip)
releases.cilium.io/tools/cluster-diagnosis.zip &&
python cluster-diagnosis.zip sysdump` and then attach the generated zip file)
[cluster-diagnosis.zip](https://github.com/cilium/cilium/files/2860757/cluster-diagnosis.zip)
**How to reproduce the issue**
1. instruction 1
2. instruction 2
| priority | clustermesh issue io cilium global service true is not working for an existing service if you have usage questions please try the slack channel and see the first choose either proposal or bug report bug report here s what i did i took your clustermesh rebel base example and downloaded your yml and yml removed the io cilium global service true from the rebel base service renamed rebel base to bipin base as i was having your original service running to test the environment alias kl kubectl config get contexts alias ka kubectl config use context a xxxxxxxxxxxx alias kb kubectl config use context b xxxxxxxxxxx alias ks kubectl n kube system alias k kubectl ka k apply f yaml nginx yml kb k apply f yaml nginx yml ka kubectl annotate svc bipin base io cilium global service true kb kubectl annotate svc bipin base io cilium global service true k get svc name type cluster ip external ip port s age bipin base clusterip tcp kubernetes clusterip tcp ks exec cilium cilium service list id frontend backend please look at the service id it show only the local ips the and yaml files which came with your examples works file that proves that this global service annotation is not being picked for an existing service as a workaround if you can tell me how to update the service definition manually for time being for demo ing this to team it would help greatly important for security related issues we strongly encourage you to report security vulnerabilities to our private security mailing list security cilium io first before disclosing them in any public forums general information cilium version run cilium version ks exec cilium cilium version client go version linux daemon go version linux kernel version run uname a coreos orchestration system version in use e g kubectl version mesos kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux link to relevant artifacts policies deployments scripts upload a system dump run curl slo releases cilium io tools cluster diagnosis zip python cluster diagnosis zip sysdump and then attach the generated zip file how to reproduce the issue instruction instruction | 1 |
643,398 | 20,956,223,631 | IssuesEvent | 2022-03-27 05:56:35 | AY2122S2-CS2103T-T09-2/tp | https://api.github.com/repos/AY2122S2-CS2103T-T09-2/tp | closed | As a Recruiter, I want to be able to search applicants by name | type.Story priority.High | so that I can view their job applications if they have applied for more than 1 job | 1.0 | As a Recruiter, I want to be able to search applicants by name - so that I can view their job applications if they have applied for more than 1 job | priority | as a recruiter i want to be able to search applicants by name so that i can view their job applications if they have applied for more than job | 1 |
306,515 | 9,395,935,249 | IssuesEvent | 2019-04-08 05:16:07 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Outbound provisioning to Salesforce does not work as expected | Affected/5.8.0-Alpha2 Complexity/Medium Component/Extensions Priority/High Resolution/Cannot Reproduce Severity/Blocker Type/Bug | **Suggested Labels**
Affected: 5.8.0-alpha2
Severity: Major
Priority: High
Component: Salesforce Outbound Provisioning
Type-Bug
**Steps to reproduce**
- Follow the instructions in the [documentation](https://docs.wso2.com/display/IS580/Outbound+Provisioning+with+Salesforce)
**Actual Behavior**
BadRequestException is thrown from the backend and the user is not created at Salesforce. | 1.0 | Outbound provisioning to Salesforce does not work as expected - **Suggested Labels**
Affected: 5.8.0-alpha2
Severity: Major
Priority: High
Component: Salesforce Outbound Provisioning
Type-Bug
**Steps to reproduce**
- Follow the instructions in the [documentation](https://docs.wso2.com/display/IS580/Outbound+Provisioning+with+Salesforce)
**Actual Behavior**
BadRequestException is thrown from the backend and the user is not created at Salesforce. | priority | outbound provisioning to salesforce does not work as expected suggested labels affected severity major priority high component salesforce outbound provisioning type bug steps to reproduce follow the instructions in the actual behavior badrequestexception is thrown from the backend and the user is not created at salesforce | 1 |
439,462 | 12,683,080,874 | IssuesEvent | 2020-06-19 18:53:02 | Helioviewer-Project/helioviewer.org | https://api.github.com/repos/Helioviewer-Project/helioviewer.org | closed | Many of the Images Are Darker and Darker in Comparison with Previous Releases and with the SDO Page | 3.3 movies priority: high reported by user screenshots | _From @live4right on July 30, 2016 1:28_
Hi,
I really enjoy using the jhelioviewer for monitoring the sun. I have actually determined there is a connection between the solar flares and severe weather events... Anyway I noticed that with each progressive release the images are becoming darker and darker. While on the SDO website they are as bright as ever. I have tried working with the various gamma settings and such but to no avail. What has happened to the coloring? I will include 2 screen shots same settings with jhelioviewer just 6 years apart. Maybe you can shed some light on this as I would like the coloring of 2010 back in the current version
Thanks,
Carl


_Copied from original issue: Helioviewer-Project/JHelioviewer#290_
| 1.0 | Many of the Images Are Darker and Darker in Comparison with Previous Releases and with the SDO Page - _From @live4right on July 30, 2016 1:28_
Hi,
I really enjoy using the jhelioviewer for monitoring the sun. I have actually determined there is a connection between the solar flares and severe weather events... Anyway I noticed that with each progressive release the images are becoming darker and darker. While on the SDO website they are as bright as ever. I have tried working with the various gamma settings and such but to no avail. What has happened to the coloring? I will include 2 screen shots same settings with jhelioviewer just 6 years apart. Maybe you can shed some light on this as I would like the coloring of 2010 back in the current version
Thanks,
Carl


_Copied from original issue: Helioviewer-Project/JHelioviewer#290_
| priority | many of the images are darker and darker in comparison with previous releases and with the sdo page from on july hi i really enjoy using the jhelioviewer for monitoring the sun i have actually determined there is a connection between the solar flares and severe weather events anyway i noticed that with each progressive release the images are becoming darker and darker while on the sdo website they are as bright as ever i have tried working with the various gamma settings and such but to no avail what has happened to the coloring i will include screen shots same settings with jhelioviewer just years apart maybe you can shed some light on this as i would like the coloring of back in the current version thanks carl copied from original issue helioviewer project jhelioviewer | 1 |
311,498 | 9,534,133,545 | IssuesEvent | 2019-04-29 23:57:57 | hackla-engage/engage-backend | https://api.github.com/repos/hackla-engage/engage-backend | opened | PDF Report enhancement - clickable ToC | Priority: High Status: Help Wanted Type: Enhancement | Per hackla-engage/Product-Documentation#3, city would like to be able to hyper to the feedback for each agenda item. A clickable Table of Contents on the 2nd page of the report should address this request. | 1.0 | PDF Report enhancement - clickable ToC - Per hackla-engage/Product-Documentation#3, city would like to be able to hyper to the feedback for each agenda item. A clickable Table of Contents on the 2nd page of the report should address this request. | priority | pdf report enhancement clickable toc per hackla engage product documentation city would like to be able to hyper to the feedback for each agenda item a clickable table of contents on the page of the report should address this request | 1 |
663,264 | 22,171,631,093 | IssuesEvent | 2022-06-06 01:50:01 | ctm/mb2-doc | https://api.github.com/repos/ctm/mb2-doc | closed | UI hang after final cards shown on final hand | bug high priority easy | Fix whatever caused the UI to not update at the very end of yesterday's evening PRP tournament.
> kidZee: […] the UI did hang once the final hole cards were shown.
This may be related to #947, but I haven't even looked at the timestamps yet since I need to get out on the trail at daylight. | 1.0 | UI hang after final cards shown on final hand - Fix whatever caused the UI to not update at the very end of yesterday's evening PRP tournament.
> kidZee: […] the UI did hang once the final hole cards were shown.
This may be related to #947, but I haven't even looked at the timestamps yet since I need to get out on the trail at daylight. | priority | ui hang after final cards shown on final hand fix whatever caused the ui to not update at the very end of yesterday s evening prp tournament kidzee the ui did hang once the final hole cards were shown this may be related to but i haven t even looked at the timestamps yet since i need to get out on the trail at daylight | 1 |
122,105 | 4,827,520,196 | IssuesEvent | 2016-11-07 13:53:33 | CS2103AUG2016-F10-C2/main | https://api.github.com/repos/CS2103AUG2016-F10-C2/main | closed | Things to complete for V0.5 submission | priority.high | - [x] Create a xml file with sample data at \src\test\data\ManualTesting\SampleData.xml
File should contain at least 50+ entries
- [x] Create a .md file outlining the steps to test the program at src\test\data\ManualTesting\TestScript.md
The test should contain expected results (screenshots not needed) and the whole test should not exceed 20 minutes. The test should cover all functionality of the app.
- [x] Submit source code by tagging v0.5 on github. Make sure all collated codes are up to date because they cannot be updated after the deadline
- [x] Upload the fat jar file to IVLE. File name: [team][project name].jar e.g. [T09-C1][ToDoo Pro].jar
- [x] Update documentation on Github.
- [x] Copy and paste contents of documentation into a .docx file following the format.
[[TeamID][ProjectName].docx](https://github.com/CS2103AUG2016-F10-C2/main/files/573190/TeamID.ProjectName.docx)
- [x] Convert the said .docx into pdf with name [TEAM_ID][Project Name].pdf e.g. [T09-C1][ToDoo Pro].pdf and upload it to IVLE
#### Deadline: Monday 2359
| 1.0 | Things to complete for V0.5 submission - - [x] Create a xml file with sample data at \src\test\data\ManualTesting\SampleData.xml
File should contain at least 50+ entries
- [x] Create a .md file outlining the steps to test the program at src\test\data\ManualTesting\TestScript.md
The test should contain expected results (screenshots not needed) and the whole test should not exceed 20 minutes. The test should cover all functionality of the app.
- [x] Submit source code by tagging v0.5 on github. Make sure all collated codes are up to date because they cannot be updated after the deadline
- [x] Upload the fat jar file to IVLE. File name: [team][project name].jar e.g. [T09-C1][ToDoo Pro].jar
- [x] Update documentation on Github.
- [x] Copy and paste contents of documentation into a .docx file following the format.
[[TeamID][ProjectName].docx](https://github.com/CS2103AUG2016-F10-C2/main/files/573190/TeamID.ProjectName.docx)
- [x] Convert the said .docx into pdf with name [TEAM_ID][Project Name].pdf e.g. [T09-C1][ToDoo Pro].pdf and upload it to IVLE
#### Deadline: Monday 2359
| priority | things to complete for submission create a xml file with sample data at src test data manualtesting sampledata xml file should contain at least entries create a md file outlining the steps to test the program at src test data manualtesting testscript md the test should contain expected results screenshots not needed and the whole test should not exceed minutes the test should cover all functionality of the app submit source code by tagging on github make sure all collated codes are up to date because they cannot be updated after the deadline upload the fat jar file to ivle file name jar e g jar update documentation on github copy and paste contents of documentation into a docx file following the format docx convert the said docx into pdf with name pdf e g pdf and upload it to ivle deadline monday | 1 |
166,497 | 6,305,814,709 | IssuesEvent | 2017-07-21 19:19:01 | DashboardHub/PipelineDashboard | https://api.github.com/repos/DashboardHub/PipelineDashboard | closed | Auto deploy to dashboardhub.io (travis -> release branches) | BLOCKED help wanted priority: high | Using custom deploy on travis get this error when adding `pem` file:
```
$ ssh-add .travis/deploy.pem
Could not open a connection to your authentication agent.
```
Posted in Gitter Travis chat room for support.
| 1.0 | Auto deploy to dashboardhub.io (travis -> release branches) - Using custom deploy on travis get this error when adding `pem` file:
```
$ ssh-add .travis/deploy.pem
Could not open a connection to your authentication agent.
```
Posted in Gitter Travis chat room for support.
| priority | auto deploy to dashboardhub io travis release branches using custom deploy on travis get this error when adding pem file ssh add travis deploy pem could not open a connection to your authentication agent posted in gitter travis chat room for support | 1 |
242,765 | 7,846,620,978 | IssuesEvent | 2018-06-19 15:57:42 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | opened | [studio] dependency resolution error causes create site to fail | CI bug priority: highest! | ### Expected behavior
Malformed paths should generate warnings/errors in the log and warnings to the user but should not fail the action.
### Actual behavior
Encountering bad content such as `/static-assets/documents/someproduct/*` bubbles up an exception and terminates the user's actions. In this case, a create site.
### Steps to reproduce the problem
working on a simple test case
### Log/stack trace (use https://gist.github.com)
https://gist.github.com/russdanner/382c88b03650d19d7710cdb075b7e0da
### Specs
#### Version
Studio Version Number: 3.0.14-SNAPSHOT-4838ba
Build Number: 4838bae0b3586ea0edabb4e8d57e3b1019f4e04e
Build Date/Time: 06-14-2018 14:17:47 -0400
#### OS
Any
#### Browser
N/A | 1.0 | [studio] dependency resolution error causes create site to fail - ### Expected behavior
Malformed paths should generate warnings/errors in the log and warnings to the user but should not fail the action.
### Actual behavior
Encountering bad content such as `/static-assets/documents/someproduct/*` bubbles up an exception and terminates the user's actions. In this case, a create site.
### Steps to reproduce the problem
working on a simple test case
### Log/stack trace (use https://gist.github.com)
https://gist.github.com/russdanner/382c88b03650d19d7710cdb075b7e0da
### Specs
#### Version
Studio Version Number: 3.0.14-SNAPSHOT-4838ba
Build Number: 4838bae0b3586ea0edabb4e8d57e3b1019f4e04e
Build Date/Time: 06-14-2018 14:17:47 -0400
#### OS
Any
#### Browser
N/A | priority | dependency resolution error causes create site to fail expected behavior malformed paths should generate warnings errors in the log and warnings to the user but should not fail the action actual behavior encountering bad content such as static assets documents someproduct bubbles up an exception and terminates the user s actions in this case a create site steps to reproduce the problem working on a simple test case log stack trace use specs version studio version number snapshot build number build date time os any browser n a | 1 |
820,446 | 30,772,403,235 | IssuesEvent | 2023-07-31 01:44:12 | steedos/steedos-platform | https://api.github.com/repos/steedos/steedos-platform | closed | [Bug]: 首页刷新后报错 | bug done priority: High | ### Description

### Steps To Reproduce 重现步骤
更新到2.5最新的正式版,进首页后报错
### Version 版本
2.5.9 | 1.0 | [Bug]: 首页刷新后报错 - ### Description

### Steps To Reproduce 重现步骤
更新到2.5最新的正式版,进首页后报错
### Version 版本
2.5.9 | priority | 首页刷新后报错 description steps to reproduce 重现步骤 ,进首页后报错 version 版本 | 1 |
782,591 | 27,500,569,922 | IssuesEvent | 2023-03-05 16:50:28 | Ellivers/WorldTool | https://api.github.com/repos/Ellivers/WorldTool | closed | Undoing is inconsistent | bug priority: high | Undoing changes seems to be very inconsistent in whether it works or not, and which changes it undoes. This will probably require a lot of testing... | 1.0 | Undoing is inconsistent - Undoing changes seems to be very inconsistent in whether it works or not, and which changes it undoes. This will probably require a lot of testing... | priority | undoing is inconsistent undoing changes seems to be very inconsistent in whether it works or not and which changes it undoes this will probably require a lot of testing | 1 |
45,438 | 2,931,442,601 | IssuesEvent | 2015-06-29 12:34:41 | ufal/lindat-dspace | https://api.github.com/repos/ufal/lindat-dspace | opened | Download of zip archive of data larger than 1073741824B is forbidden. | high priority | Its a strange error, I tried to access all the items during the timeframe of the exception, and I am unable to reproduce it. We need to investigate more about this.
org.apache.cocoon.ProcessingException: Download of zip archive of data larger than 1073741824B is forbidden.
at <map:read type="AllBitstreamZipArchiveReader"> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:334:77
at <map:select type="AuthenticatedSelector"> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:332:46
at <map:match> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:331:64
at cz.cuni.mff.ufal.AllBitstreamZipArchiveReader.setup(AllBitstreamZipArchiveReader.java:187)
at sun.reflect.GeneratedMethodAccessor69.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.cocoon.core.container.spring.avalon.PoolableProxyHandler.invoke(PoolableProxyHandler.java:71)
at com.sun.proxy.$Proxy51.setup(Unknown Source)
at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.setupReader(AbstractProcessingPipeline.java:560)
at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.preparePipeline(AbstractProcessingPipeline.java:464)
at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.process(AbstractProcessingPipeline.java:411)
at sun.reflect.GeneratedMethodAccessor195.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.cocoon.core.container.spring.avalon.PoolableProxyHandler.invoke(PoolableProxyHandler.java:71)
at com.sun.proxy.$Proxy29.process(Unknown Source)
at org.apache.cocoon.components.treeprocessor.sitemap.ReadNode.invoke(ReadNode.java:94)
at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78)
at org.apache.cocoon.components.treeprocessor.sitemap.SelectNode.invoke(SelectNode.java:82)
at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:55)
at org.apache.cocoon.components.treeprocessor.sitemap.MatchNode.invoke(MatchNode.java:87)
at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78)
at org.apache.cocoon.components.treeprocessor.sitemap.PipelineNode.invoke(PipelineNode.java:143)
at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78)
at org.apache.cocoon.components.treeprocessor.sitemap.PipelinesNode.invoke(PipelinesNode.java:81)
at org.apache.cocoon.components.treeprocessor.ConcreteTreeProcessor.process(ConcreteTreeProcessor.java:239)
at org.apache.cocoon.components.treeprocessor.ConcreteTreeProcessor.process(ConcreteTreeProcessor.java:171)
at org.apache.cocoon.components.treeprocessor.TreeProcessor.process(TreeProcessor.java:247)
at org.apache.cocoon.servlet.RequestProcessor.process(RequestProcessor.java:351)
at org.apache.cocoon.servlet.RequestProcessor.service(RequestProcessor.java:169)
at org.apache.cocoon.sitemap.SitemapServlet.service(SitemapServlet.java:84)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:729)
at org.apache.cocoon.servletservice.ServletServiceContext$PathDispatcher.forward(ServletServiceContext.java:468)
at org.apache.cocoon.servletservice.ServletServiceContext$PathDispatcher.forward(ServletServiceContext.java:443)
at org.apache.cocoon.servletservice.spring.ServletFactoryBean$ServiceInterceptor.invoke(ServletFactoryBean.java:264)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)
at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:202)
at com.sun.proxy.$Proxy20.service(Unknown Source)
at org.dspace.springmvc.CocoonView.render(CocoonView.java:117)
at org.springframework.web.servlet.DispatcherServlet.render(DispatcherServlet.java:1180)
at org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:950)
at org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:852)
at org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:882)
at org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:778)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:622)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:729)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:291)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.app.xmlui.cocoon.SetCharacterEncodingFilter.doFilter(SetCharacterEncodingFilter.java:111)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.app.xmlui.cocoon.DSpaceCocoonServletFilter.doFilter(DSpaceCocoonServletFilter.java:276)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.app.xmlui.cocoon.servlet.multipart.DSpaceMultipartFilter.doFilter(DSpaceMultipartFilter.java:119)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.rdf.negotiation.NegotiationFilter.doFilter(NegotiationFilter.java:50)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.utils.servlet.DSpaceWebappServletFilter.doFilter(DSpaceWebappServletFilter.java:78)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:219)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106)
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:142)
at com.googlecode.psiprobe.Tomcat80AgentValve.invoke(Tomcat80AgentValve.java:36)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79)
at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:610)
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518)
at org.apache.coyote.ajp.AbstractAjpProcessor.process(AbstractAjpProcessor.java:844)
at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1517)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1474)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
at java.lang.Thread.run(Thread.java:745) | 1.0 | Download of zip archive of data larger than 1073741824B is forbidden. - Its a strange error, I tried to access all the items during the timeframe of the exception, and I am unable to reproduce it. We need to investigate more about this.
org.apache.cocoon.ProcessingException: Download of zip archive of data larger than 1073741824B is forbidden.
at <map:read type="AllBitstreamZipArchiveReader"> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:334:77
at <map:select type="AuthenticatedSelector"> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:332:46
at <map:match> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:331:64
at cz.cuni.mff.ufal.AllBitstreamZipArchiveReader.setup(AllBitstreamZipArchiveReader.java:187)
at sun.reflect.GeneratedMethodAccessor69.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.cocoon.core.container.spring.avalon.PoolableProxyHandler.invoke(PoolableProxyHandler.java:71)
at com.sun.proxy.$Proxy51.setup(Unknown Source)
at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.setupReader(AbstractProcessingPipeline.java:560)
at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.preparePipeline(AbstractProcessingPipeline.java:464)
at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.process(AbstractProcessingPipeline.java:411)
at sun.reflect.GeneratedMethodAccessor195.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.cocoon.core.container.spring.avalon.PoolableProxyHandler.invoke(PoolableProxyHandler.java:71)
at com.sun.proxy.$Proxy29.process(Unknown Source)
at org.apache.cocoon.components.treeprocessor.sitemap.ReadNode.invoke(ReadNode.java:94)
at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78)
at org.apache.cocoon.components.treeprocessor.sitemap.SelectNode.invoke(SelectNode.java:82)
at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:55)
at org.apache.cocoon.components.treeprocessor.sitemap.MatchNode.invoke(MatchNode.java:87)
at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78)
at org.apache.cocoon.components.treeprocessor.sitemap.PipelineNode.invoke(PipelineNode.java:143)
at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78)
at org.apache.cocoon.components.treeprocessor.sitemap.PipelinesNode.invoke(PipelinesNode.java:81)
at org.apache.cocoon.components.treeprocessor.ConcreteTreeProcessor.process(ConcreteTreeProcessor.java:239)
at org.apache.cocoon.components.treeprocessor.ConcreteTreeProcessor.process(ConcreteTreeProcessor.java:171)
at org.apache.cocoon.components.treeprocessor.TreeProcessor.process(TreeProcessor.java:247)
at org.apache.cocoon.servlet.RequestProcessor.process(RequestProcessor.java:351)
at org.apache.cocoon.servlet.RequestProcessor.service(RequestProcessor.java:169)
at org.apache.cocoon.sitemap.SitemapServlet.service(SitemapServlet.java:84)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:729)
at org.apache.cocoon.servletservice.ServletServiceContext$PathDispatcher.forward(ServletServiceContext.java:468)
at org.apache.cocoon.servletservice.ServletServiceContext$PathDispatcher.forward(ServletServiceContext.java:443)
at org.apache.cocoon.servletservice.spring.ServletFactoryBean$ServiceInterceptor.invoke(ServletFactoryBean.java:264)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)
at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:202)
at com.sun.proxy.$Proxy20.service(Unknown Source)
at org.dspace.springmvc.CocoonView.render(CocoonView.java:117)
at org.springframework.web.servlet.DispatcherServlet.render(DispatcherServlet.java:1180)
at org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:950)
at org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:852)
at org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:882)
at org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:778)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:622)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:729)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:291)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.app.xmlui.cocoon.SetCharacterEncodingFilter.doFilter(SetCharacterEncodingFilter.java:111)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.app.xmlui.cocoon.DSpaceCocoonServletFilter.doFilter(DSpaceCocoonServletFilter.java:276)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.app.xmlui.cocoon.servlet.multipart.DSpaceMultipartFilter.doFilter(DSpaceMultipartFilter.java:119)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.rdf.negotiation.NegotiationFilter.doFilter(NegotiationFilter.java:50)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.dspace.utils.servlet.DSpaceWebappServletFilter.doFilter(DSpaceWebappServletFilter.java:78)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:219)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106)
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:142)
at com.googlecode.psiprobe.Tomcat80AgentValve.invoke(Tomcat80AgentValve.java:36)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79)
at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:610)
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518)
at org.apache.coyote.ajp.AbstractAjpProcessor.process(AbstractAjpProcessor.java:844)
at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1517)
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1474)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
at java.lang.Thread.run(Thread.java:745) | priority | download of zip archive of data larger than is forbidden its a strange error i tried to access all the items during the timeframe of the exception and i am unable to reproduce it we need to investigate more about this org apache cocoon processingexception download of zip archive of data larger than is forbidden at file mnt data home dspace installations dspace lindat dspace webapps xmlui sitemap xmap at file mnt data home dspace installations dspace lindat dspace webapps xmlui sitemap xmap at file mnt data home dspace installations dspace lindat dspace webapps xmlui sitemap xmap at cz cuni mff ufal allbitstreamziparchivereader setup allbitstreamziparchivereader java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache cocoon core container spring avalon poolableproxyhandler invoke poolableproxyhandler java at com sun proxy setup unknown source at org apache cocoon components pipeline abstractprocessingpipeline setupreader abstractprocessingpipeline java at org apache cocoon components pipeline abstractprocessingpipeline preparepipeline abstractprocessingpipeline java at org apache cocoon components pipeline abstractprocessingpipeline process abstractprocessingpipeline java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache cocoon core container spring avalon poolableproxyhandler invoke poolableproxyhandler java at com sun proxy process unknown source at org apache cocoon components treeprocessor sitemap readnode invoke readnode java at org apache cocoon components treeprocessor abstractparentprocessingnode invokenodes abstractparentprocessingnode java at org apache cocoon components treeprocessor sitemap selectnode invoke selectnode java at org apache cocoon components treeprocessor abstractparentprocessingnode invokenodes abstractparentprocessingnode java at org apache cocoon components treeprocessor sitemap matchnode invoke matchnode java at org apache cocoon components treeprocessor abstractparentprocessingnode invokenodes abstractparentprocessingnode java at org apache cocoon components treeprocessor sitemap pipelinenode invoke pipelinenode java at org apache cocoon components treeprocessor abstractparentprocessingnode invokenodes abstractparentprocessingnode java at org apache cocoon components treeprocessor sitemap pipelinesnode invoke pipelinesnode java at org apache cocoon components treeprocessor concretetreeprocessor process concretetreeprocessor java at org apache cocoon components treeprocessor concretetreeprocessor process concretetreeprocessor java at org apache cocoon components treeprocessor treeprocessor process treeprocessor java at org apache cocoon servlet requestprocessor process requestprocessor java at org apache cocoon servlet requestprocessor service requestprocessor java at org apache cocoon sitemap sitemapservlet service sitemapservlet java at javax servlet http httpservlet service httpservlet java at org apache cocoon servletservice servletservicecontext pathdispatcher forward servletservicecontext java at org apache cocoon servletservice servletservicecontext pathdispatcher forward servletservicecontext java at org apache cocoon servletservice spring servletfactorybean serviceinterceptor invoke servletfactorybean java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework aop framework jdkdynamicaopproxy invoke jdkdynamicaopproxy java at com sun proxy service unknown source at org dspace springmvc cocoonview render cocoonview java at org springframework web servlet dispatcherservlet render dispatcherservlet java at org springframework web servlet dispatcherservlet dodispatch dispatcherservlet java at org springframework web servlet dispatcherservlet doservice dispatcherservlet java at org springframework web servlet frameworkservlet processrequest frameworkservlet java at org springframework web servlet frameworkservlet doget frameworkservlet java at javax servlet http httpservlet service httpservlet java at javax servlet http httpservlet service httpservlet java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace app xmlui cocoon setcharacterencodingfilter dofilter setcharacterencodingfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace app xmlui cocoon dspacecocoonservletfilter dofilter dspacecocoonservletfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace app xmlui cocoon servlet multipart dspacemultipartfilter dofilter dspacemultipartfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace rdf negotiation negotiationfilter dofilter negotiationfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace utils servlet dspacewebappservletfilter dofilter dspacewebappservletfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at com googlecode psiprobe invoke java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote ajp abstractajpprocessor process abstractajpprocessor java at org apache coyote abstractprotocol abstractconnectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net nioendpoint socketprocessor run nioendpoint java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java | 1 |
137,746 | 5,315,858,848 | IssuesEvent | 2017-02-13 18:23:53 | Valhalla-Gaming/Tracker | https://api.github.com/repos/Valhalla-Gaming/Tracker | closed | Druid talent Warrior of Elune. | Class-Druid Priority-High Type-Spell | **Describe the issue you're having**:
After using Warrior of Elune (talent) it gives you 2 Lunar Strikes without casting it, but instead you can cast it unlimited times.
**Explain how you expect it work**:
You should get only 2 Lunar Strikes without casting
This also works with mage's talent which let's you use Arcane Blast 2 times without casting it.
**Steps to reproduce the problem**:
1. Learn the talent
2. Use it
3. Spam Lunar Strike
**Links to Wowhead, YouTube, etc**:
http://www.wowhead.com/spell=202425/warrior-of-elune | 1.0 | Druid talent Warrior of Elune. - **Describe the issue you're having**:
After using Warrior of Elune (talent) it gives you 2 Lunar Strikes without casting it, but instead you can cast it unlimited times.
**Explain how you expect it work**:
You should get only 2 Lunar Strikes without casting
This also works with mage's talent which let's you use Arcane Blast 2 times without casting it.
**Steps to reproduce the problem**:
1. Learn the talent
2. Use it
3. Spam Lunar Strike
**Links to Wowhead, YouTube, etc**:
http://www.wowhead.com/spell=202425/warrior-of-elune | priority | druid talent warrior of elune describe the issue you re having after using warrior of elune talent it gives you lunar strikes without casting it but instead you can cast it unlimited times explain how you expect it work you should get only lunar strikes without casting this also works with mage s talent which let s you use arcane blast times without casting it steps to reproduce the problem learn the talent use it spam lunar strike links to wowhead youtube etc | 1 |
152,391 | 5,846,278,406 | IssuesEvent | 2017-05-10 15:50:56 | radcampaign/pulitzer2016 | https://api.github.com/repos/radcampaign/pulitzer2016 | closed | event - image caption missing | needs work priority HIGH | Not a huge rush, but something to be sure to fix.
You can see that captions on old events are still appearing, but the field on the backend to enter/edit is missing: http://pulitzercenter.org/event/gallery-opening-god-government-lgbt-rights-russia-india-uganda | 1.0 | event - image caption missing - Not a huge rush, but something to be sure to fix.
You can see that captions on old events are still appearing, but the field on the backend to enter/edit is missing: http://pulitzercenter.org/event/gallery-opening-god-government-lgbt-rights-russia-india-uganda | priority | event image caption missing not a huge rush but something to be sure to fix you can see that captions on old events are still appearing but the field on the backend to enter edit is missing | 1 |
579,016 | 17,170,446,623 | IssuesEvent | 2021-07-15 03:02:49 | TeamDooRiBon/DooRi-iOS | https://api.github.com/repos/TeamDooRiBon/DooRi-iOS | closed | [FEAT] 여행 보드 조회 API 연결 | Feat P1 / Priority High Taehyeon 🐮 | # 👀 이슈 (issue)
여행 보드 조회하는 API를 연결합니다.
# 🚀 to-do
<!-- 진행할 작업에 대해 적어주세요 -->
<img width="250" alt="스크린샷 2021-07-15 오전 3 37 25" src="https://user-images.githubusercontent.com/61109660/125675169-6efbc15d-d20a-42b7-9cd5-afe45f25c4a3.png">
- [ ] Model 생성
- [ ] Service 구현
- [ ] API 연결 및 테스트
| 1.0 | [FEAT] 여행 보드 조회 API 연결 - # 👀 이슈 (issue)
여행 보드 조회하는 API를 연결합니다.
# 🚀 to-do
<!-- 진행할 작업에 대해 적어주세요 -->
<img width="250" alt="스크린샷 2021-07-15 오전 3 37 25" src="https://user-images.githubusercontent.com/61109660/125675169-6efbc15d-d20a-42b7-9cd5-afe45f25c4a3.png">
- [ ] Model 생성
- [ ] Service 구현
- [ ] API 연결 및 테스트
| priority | 여행 보드 조회 api 연결 👀 이슈 issue 여행 보드 조회하는 api를 연결합니다 🚀 to do img width alt 스크린샷 오전 src model 생성 service 구현 api 연결 및 테스트 | 1 |
316,548 | 9,651,538,211 | IssuesEvent | 2019-05-18 08:49:20 | usemoslinux/aprelendo | https://api.github.com/repos/usemoslinux/aprelendo | closed | Optimize static images to reduce their size | Priority: high Type: improvement | Losslessly compressing https://www.aprelendo.com/img/backgrounds/welcome-page-background.jpeg could saves 12.4KiB (3% reduction). | 1.0 | Optimize static images to reduce their size - Losslessly compressing https://www.aprelendo.com/img/backgrounds/welcome-page-background.jpeg could saves 12.4KiB (3% reduction). | priority | optimize static images to reduce their size losslessly compressing could saves reduction | 1 |
210,792 | 7,194,737,769 | IssuesEvent | 2018-02-04 09:14:04 | hpi-swt2/sport-portal | https://api.github.com/repos/hpi-swt2/sport-portal | closed | Notifications: events/matches | epic notifications po-review priority high team kickerz | As User I want to be informed about changes in my events to plan my matches accordingly.
**Acceptance criteria**
- general requirements see #301
- emails are automatically generated when triggered
- the text of emails is predefined and includes placeholders for customization (see texts below)
Triggers are:
- a match where user participates is scheduled
> Hallo _user name_,
> dein Spiel gegen _oponent team_ in _event name_ findet am _datum_ statt.
> Viel Erfolg beim Training!
> _Link to the event_
- a match where user participates is cancelled or postponed
> Hallo _user name_,
> dein Spiel gegen _oponent team_ in _event name_ wurde (auf _datum_ verschoben / abgesagt).
> Viel Erfolg beim Training!
> _Link to the event_
- an event where user is registered is started/cancelled/finished
> Hallo _user name_,
> das Event _event name_ hat gestartet (ist abgesagt/ ist abgeschlossen).
> Viel Erfolg beim Training!
> _Link to the event_
**Important**
Check issues #57 and #149 and communicate with other dev teams
| 1.0 | Notifications: events/matches - As User I want to be informed about changes in my events to plan my matches accordingly.
**Acceptance criteria**
- general requirements see #301
- emails are automatically generated when triggered
- the text of emails is predefined and includes placeholders for customization (see texts below)
Triggers are:
- a match where user participates is scheduled
> Hallo _user name_,
> dein Spiel gegen _oponent team_ in _event name_ findet am _datum_ statt.
> Viel Erfolg beim Training!
> _Link to the event_
- a match where user participates is cancelled or postponed
> Hallo _user name_,
> dein Spiel gegen _oponent team_ in _event name_ wurde (auf _datum_ verschoben / abgesagt).
> Viel Erfolg beim Training!
> _Link to the event_
- an event where user is registered is started/cancelled/finished
> Hallo _user name_,
> das Event _event name_ hat gestartet (ist abgesagt/ ist abgeschlossen).
> Viel Erfolg beim Training!
> _Link to the event_
**Important**
Check issues #57 and #149 and communicate with other dev teams
| priority | notifications events matches as user i want to be informed about changes in my events to plan my matches accordingly acceptance criteria general requirements see emails are automatically generated when triggered the text of emails is predefined and includes placeholders for customization see texts below triggers are a match where user participates is scheduled hallo user name dein spiel gegen oponent team in event name findet am datum statt viel erfolg beim training link to the event a match where user participates is cancelled or postponed hallo user name dein spiel gegen oponent team in event name wurde auf datum verschoben abgesagt viel erfolg beim training link to the event an event where user is registered is started cancelled finished hallo user name das event event name hat gestartet ist abgesagt ist abgeschlossen viel erfolg beim training link to the event important check issues and and communicate with other dev teams | 1 |
343,001 | 10,324,312,113 | IssuesEvent | 2019-09-01 07:58:53 | OpenSRP/opensrp-client-chw-anc | https://api.github.com/repos/OpenSRP/opensrp-client-chw-anc | closed | Accent characters are not rendering anywhere in the app UI | High Priority bug | Inside the forms, the accent characters are rendering, but anywhere else in the app UI (register page, family profile page, etc.) they are not.

| 1.0 | Accent characters are not rendering anywhere in the app UI - Inside the forms, the accent characters are rendering, but anywhere else in the app UI (register page, family profile page, etc.) they are not.

| priority | accent characters are not rendering anywhere in the app ui inside the forms the accent characters are rendering but anywhere else in the app ui register page family profile page etc they are not | 1 |
178,072 | 6,598,684,518 | IssuesEvent | 2017-09-16 09:15:33 | oSoc17/rideaway-frontend | https://api.github.com/repos/oSoc17/rideaway-frontend | opened | Lock top controls when calculating a route | bug high priority | Because the loading screen is on the map (feels more intuitive), we should lock the geocoder inputs because else we get weird behaviour, aka 2 routes being calculated at the same time. | 1.0 | Lock top controls when calculating a route - Because the loading screen is on the map (feels more intuitive), we should lock the geocoder inputs because else we get weird behaviour, aka 2 routes being calculated at the same time. | priority | lock top controls when calculating a route because the loading screen is on the map feels more intuitive we should lock the geocoder inputs because else we get weird behaviour aka routes being calculated at the same time | 1 |
197,797 | 6,964,071,990 | IssuesEvent | 2017-12-08 20:01:56 | uva-slp/meals | https://api.github.com/repos/uva-slp/meals | closed | Improve alerts page | customer-requests priority-high story-points-5 volunteer-database | Currently, the alerts page only shows what routes and jobs are always missing volunteers. It needs to be modified to show which routes and jobs need volunteers because they have a substitution request submitted. | 1.0 | Improve alerts page - Currently, the alerts page only shows what routes and jobs are always missing volunteers. It needs to be modified to show which routes and jobs need volunteers because they have a substitution request submitted. | priority | improve alerts page currently the alerts page only shows what routes and jobs are always missing volunteers it needs to be modified to show which routes and jobs need volunteers because they have a substitution request submitted | 1 |
328,930 | 10,001,617,697 | IssuesEvent | 2019-07-12 16:03:25 | ampproject/amp-wp | https://api.github.com/repos/ampproject/amp-wp | closed | Choosing pre-defined font size causing error | AMP Stories Sprint 10 [Priority] High [Type] Bug | Steps to reproduce:
- Create a text block.
- Add some text.
- Turn off the setting for amp-fit-text ("Automatically fit text to container").
- Switch the font size to one of the pre-defined sizes, e.g. to “Large”
<img width="1374" alt="Screenshot 2019-07-11 at 13 26 47" src="https://user-images.githubusercontent.com/3294597/61049084-df411980-a3e3-11e9-8f53-b4d2e0c46989.png">
AC: User should be able to set a custom font size and choose between different font sizes. | 1.0 | Choosing pre-defined font size causing error - Steps to reproduce:
- Create a text block.
- Add some text.
- Turn off the setting for amp-fit-text ("Automatically fit text to container").
- Switch the font size to one of the pre-defined sizes, e.g. to “Large”
<img width="1374" alt="Screenshot 2019-07-11 at 13 26 47" src="https://user-images.githubusercontent.com/3294597/61049084-df411980-a3e3-11e9-8f53-b4d2e0c46989.png">
AC: User should be able to set a custom font size and choose between different font sizes. | priority | choosing pre defined font size causing error steps to reproduce create a text block add some text turn off the setting for amp fit text automatically fit text to container switch the font size to one of the pre defined sizes e g to “large” img width alt screenshot at src ac user should be able to set a custom font size and choose between different font sizes | 1 |
689,470 | 23,621,834,462 | IssuesEvent | 2022-08-24 21:26:29 | choderalab/perses | https://api.github.com/repos/choderalab/perses | opened | `RelativeFEPSetup` cannot deal with list of ligand input files | bug :bug: priority: high effort: low | In PR #1073 we included a way to deal with files from different cloud providers using `cloudpathlib`. But when doing this we generated a bug where if we pass a list of filenames for the `RelativeFEPSetup` object as the `ligand_input` parameter, we end up with an error like the following:
```python
Traceback (most recent call last):
File "/home/user/workdir/debugging/perses/perses-new-issue/test_RelativeFEPSetup_API.py", line 19, in <module>
fe_setup = RelativeFEPSetup(
File "/home/user/miniconda3/envs/perses-1092-omm-nightly-test/lib/python3.9/site-packages/perses/app/relative_setup.py", line 165, in __init__
self._ligand_input = AnyPath(ligand_input)
File "/home/user/miniconda3/envs/perses-1092-omm-nightly-test/lib/python3.9/site-packages/cloudpathlib/anypath.py", line 38, in __new__
raise AnyPathTypeError(
cloudpathlib.exceptions.AnyPathTypeError: Invalid input for both CloudPath and Path. CloudPath exception: InvalidPrefixError("Path ['./a1.sybyl.mol2', './a2.sybyl.mol2'] does not begin with a known prefix ['az://', 's3://', 'gs://'].") Path exception: TypeError('expected str, bytes or os.PathLike object, not list')
(perses-1092-omm-nightly-test)
```
This is caused by trying to convert the list to an `cloudpathlib.AnyPath` object [here](https://github.com/choderalab/perses/blob/c96e65572b26cfdbce2d3ecaec7dae708e01568c/perses/app/relative_setup.py#L164-L165).
A quick fix would be to just convert the given `input_ligand` (or even `self._input_ligand`) to an `AnyPath` object after we check it isn't a list in [line 242](https://github.com/choderalab/perses/blob/c96e65572b26cfdbce2d3ecaec7dae708e01568c/perses/app/relative_setup.py#L242).
A longer term solution would be to refactor how we are dealing with this and have special methods for each case (list or str path) instead of having the huge if-else statements we currently have. | 1.0 | `RelativeFEPSetup` cannot deal with list of ligand input files - In PR #1073 we included a way to deal with files from different cloud providers using `cloudpathlib`. But when doing this we generated a bug where if we pass a list of filenames for the `RelativeFEPSetup` object as the `ligand_input` parameter, we end up with an error like the following:
```python
Traceback (most recent call last):
File "/home/user/workdir/debugging/perses/perses-new-issue/test_RelativeFEPSetup_API.py", line 19, in <module>
fe_setup = RelativeFEPSetup(
File "/home/user/miniconda3/envs/perses-1092-omm-nightly-test/lib/python3.9/site-packages/perses/app/relative_setup.py", line 165, in __init__
self._ligand_input = AnyPath(ligand_input)
File "/home/user/miniconda3/envs/perses-1092-omm-nightly-test/lib/python3.9/site-packages/cloudpathlib/anypath.py", line 38, in __new__
raise AnyPathTypeError(
cloudpathlib.exceptions.AnyPathTypeError: Invalid input for both CloudPath and Path. CloudPath exception: InvalidPrefixError("Path ['./a1.sybyl.mol2', './a2.sybyl.mol2'] does not begin with a known prefix ['az://', 's3://', 'gs://'].") Path exception: TypeError('expected str, bytes or os.PathLike object, not list')
(perses-1092-omm-nightly-test)
```
This is caused by trying to convert the list to an `cloudpathlib.AnyPath` object [here](https://github.com/choderalab/perses/blob/c96e65572b26cfdbce2d3ecaec7dae708e01568c/perses/app/relative_setup.py#L164-L165).
A quick fix would be to just convert the given `input_ligand` (or even `self._input_ligand`) to an `AnyPath` object after we check it isn't a list in [line 242](https://github.com/choderalab/perses/blob/c96e65572b26cfdbce2d3ecaec7dae708e01568c/perses/app/relative_setup.py#L242).
A longer term solution would be to refactor how we are dealing with this and have special methods for each case (list or str path) instead of having the huge if-else statements we currently have. | priority | relativefepsetup cannot deal with list of ligand input files in pr we included a way to deal with files from different cloud providers using cloudpathlib but when doing this we generated a bug where if we pass a list of filenames for the relativefepsetup object as the ligand input parameter we end up with an error like the following python traceback most recent call last file home user workdir debugging perses perses new issue test relativefepsetup api py line in fe setup relativefepsetup file home user envs perses omm nightly test lib site packages perses app relative setup py line in init self ligand input anypath ligand input file home user envs perses omm nightly test lib site packages cloudpathlib anypath py line in new raise anypathtypeerror cloudpathlib exceptions anypathtypeerror invalid input for both cloudpath and path cloudpath exception invalidprefixerror path does not begin with a known prefix path exception typeerror expected str bytes or os pathlike object not list perses omm nightly test this is caused by trying to convert the list to an cloudpathlib anypath object a quick fix would be to just convert the given input ligand or even self input ligand to an anypath object after we check it isn t a list in a longer term solution would be to refactor how we are dealing with this and have special methods for each case list or str path instead of having the huge if else statements we currently have | 1 |
822,586 | 30,878,991,082 | IssuesEvent | 2023-08-03 16:06:43 | dhowe/AdNauseam | https://api.github.com/repos/dhowe/AdNauseam | closed | YouTube AdBlock ban bypass? | PRIORITY: High Ads Visible | Hello, could you guys please see if this can be replicated on your end?

Reddit Page:
https://www.reddit.com/r/youtube/comments/13cfdbi/apparently_ad_blockers_are_not_allowed_on_youtube/
Thought i'd point this out to you guys to prepare Adnauseum when this more widely rolls out. Hopefully we'll continue blocking youtube ads. | 1.0 | YouTube AdBlock ban bypass? - Hello, could you guys please see if this can be replicated on your end?

Reddit Page:
https://www.reddit.com/r/youtube/comments/13cfdbi/apparently_ad_blockers_are_not_allowed_on_youtube/
Thought i'd point this out to you guys to prepare Adnauseum when this more widely rolls out. Hopefully we'll continue blocking youtube ads. | priority | youtube adblock ban bypass hello could you guys please see if this can be replicated on your end reddit page thought i d point this out to you guys to prepare adnauseum when this more widely rolls out hopefully we ll continue blocking youtube ads | 1 |
783,568 | 27,537,077,991 | IssuesEvent | 2023-03-07 04:54:05 | AY2223S2-CS2113-W15-1/tp | https://api.github.com/repos/AY2223S2-CS2113-W15-1/tp | opened | Calculate Calories | priority.High type.Story | As a user, I can calculate my caloric needs based on my height and weight so that I will know how much calories I can take in, in a day | 1.0 | Calculate Calories - As a user, I can calculate my caloric needs based on my height and weight so that I will know how much calories I can take in, in a day | priority | calculate calories as a user i can calculate my caloric needs based on my height and weight so that i will know how much calories i can take in in a day | 1 |
524,527 | 15,215,882,267 | IssuesEvent | 2021-02-17 14:54:33 | bexis/Module_ResourceManagement | https://api.github.com/repos/bexis/Module_ResourceManagement | opened | Edit resource time : end-date changes unexpected | Priority: High bug | If you change the "from-date" the "to-date" by itself to "to-date"+1 day
Should stay as it is. | 1.0 | Edit resource time : end-date changes unexpected - If you change the "from-date" the "to-date" by itself to "to-date"+1 day
Should stay as it is. | priority | edit resource time end date changes unexpected if you change the from date the to date by itself to to date day should stay as it is | 1 |
198,040 | 6,968,970,438 | IssuesEvent | 2017-12-11 01:49:30 | madeline-bauer/backend-web-dev | https://api.github.com/repos/madeline-bauer/backend-web-dev | closed | Make authorization system | Priority: High Status: Accepted Type: Enhancement | @ccourtney121, I've made a file `authorization.js` that gets passed the request every time an API call is made. It should return `true` or `false` for if the request is authorized. Right now I've just hard-coded `return true`.
When you make security happen, this is where it should happen. If it returns `false` then the request will get an `Error 401 (Unauthorized)` instead of the data they were after.
I think you're planning on using session tokens? So it would be something like `if (req.query.token == /*is in authorized tokens db*/){return true;}` | 1.0 | Make authorization system - @ccourtney121, I've made a file `authorization.js` that gets passed the request every time an API call is made. It should return `true` or `false` for if the request is authorized. Right now I've just hard-coded `return true`.
When you make security happen, this is where it should happen. If it returns `false` then the request will get an `Error 401 (Unauthorized)` instead of the data they were after.
I think you're planning on using session tokens? So it would be something like `if (req.query.token == /*is in authorized tokens db*/){return true;}` | priority | make authorization system i ve made a file authorization js that gets passed the request every time an api call is made it should return true or false for if the request is authorized right now i ve just hard coded return true when you make security happen this is where it should happen if it returns false then the request will get an error unauthorized instead of the data they were after i think you re planning on using session tokens so it would be something like if req query token is in authorized tokens db return true | 1 |
655,786 | 21,709,209,033 | IssuesEvent | 2022-05-10 12:31:30 | wazuh/wazuh-documentation | https://api.github.com/repos/wazuh/wazuh-documentation | closed | Create Release Notes for Release 4.3.0 | priority: high type: refactor | Hi team,
This issue aims to create the Release Notes for 4.3.0 that will be published on the GA date.
The commits of this issue can be seen here: https://github.com/wazuh/wazuh-documentation/commits/4.3-rc
### Stage 1:
- [x] Create branch and template
- [x] Check Changelogs and add issues from `wazuh/wazuh rc1`:
- Highlights
- [x] Manager
- [x] Agent
- [x] Wazuh Kibana plugin
- [x] RESTful API
- What’s new
- [x] Manager
- [x] Agent
- [x] RESTful API
- [x] Ruleset
- [x] Wazuh Kibana plugin
- [x] Others
- Resolved issues
- [x] Manager
- [x] Agent
- [x] RESTful API
- [x] Ruleset
- [x] Wazuh Kibana plugin
- [x] Others
- [x] Changelogs
- [x] Edit the content and check that the new added issues are consistent
- [x] Check correspondence between issue number and number in link
- [x] Make necessary linguistic adjustments
- [x] Check Changelogs and add issues from `wazuh/wazuh rc2`
- [x] Check Changelogs and add issues from `wazuh-kibana-app/CHANGELOG.md`
- [x] Check Changelogs and add issues from `wazuh-splunk/CHANGELOG.md`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc3`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc4`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc5`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc6`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc7`
- [x] Check Changelogs and add issues from `wazuh-packages/CHANGELOG.md`
- [x] Edit the content and check that the new added issues are consistent
- [x] Check correspondence between issue number and number in link
- [x] Make necessary linguistic adjustments
- [x] Review from the content team
### Stage 2:
- [x] Work on the branch `4.3.0-release-notes-update`
- [x] Make a review, add updates and make fixes
- [x] Review from the content team
Further steps will be added if additional revisions are needed for new release candidates.
Regards,
Mariel and Damián | 1.0 | Create Release Notes for Release 4.3.0 - Hi team,
This issue aims to create the Release Notes for 4.3.0 that will be published on the GA date.
The commits of this issue can be seen here: https://github.com/wazuh/wazuh-documentation/commits/4.3-rc
### Stage 1:
- [x] Create branch and template
- [x] Check Changelogs and add issues from `wazuh/wazuh rc1`:
- Highlights
- [x] Manager
- [x] Agent
- [x] Wazuh Kibana plugin
- [x] RESTful API
- What’s new
- [x] Manager
- [x] Agent
- [x] RESTful API
- [x] Ruleset
- [x] Wazuh Kibana plugin
- [x] Others
- Resolved issues
- [x] Manager
- [x] Agent
- [x] RESTful API
- [x] Ruleset
- [x] Wazuh Kibana plugin
- [x] Others
- [x] Changelogs
- [x] Edit the content and check that the new added issues are consistent
- [x] Check correspondence between issue number and number in link
- [x] Make necessary linguistic adjustments
- [x] Check Changelogs and add issues from `wazuh/wazuh rc2`
- [x] Check Changelogs and add issues from `wazuh-kibana-app/CHANGELOG.md`
- [x] Check Changelogs and add issues from `wazuh-splunk/CHANGELOG.md`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc3`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc4`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc5`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc6`
- [x] Check Changelogs and add issues from `wazuh/wazuh rc7`
- [x] Check Changelogs and add issues from `wazuh-packages/CHANGELOG.md`
- [x] Edit the content and check that the new added issues are consistent
- [x] Check correspondence between issue number and number in link
- [x] Make necessary linguistic adjustments
- [x] Review from the content team
### Stage 2:
- [x] Work on the branch `4.3.0-release-notes-update`
- [x] Make a review, add updates and make fixes
- [x] Review from the content team
Further steps will be added if additional revisions are needed for new release candidates.
Regards,
Mariel and Damián | priority | create release notes for release hi team this issue aims to create the release notes for that will be published on the ga date the commits of this issue can be seen here stage create branch and template check changelogs and add issues from wazuh wazuh highlights manager agent wazuh kibana plugin restful api what’s new manager agent restful api ruleset wazuh kibana plugin others resolved issues manager agent restful api ruleset wazuh kibana plugin others changelogs edit the content and check that the new added issues are consistent check correspondence between issue number and number in link make necessary linguistic adjustments check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh kibana app changelog md check changelogs and add issues from wazuh splunk changelog md check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh packages changelog md edit the content and check that the new added issues are consistent check correspondence between issue number and number in link make necessary linguistic adjustments review from the content team stage work on the branch release notes update make a review add updates and make fixes review from the content team further steps will be added if additional revisions are needed for new release candidates regards mariel and damián | 1 |
463,323 | 13,263,311,977 | IssuesEvent | 2020-08-21 00:08:53 | radcampaign/allaboveall2020 | https://api.github.com/repos/radcampaign/allaboveall2020 | opened | [OG Image] Get working | Priority: High | I think this is still unsolved? Or I missed what's up on slack. Putting an image out there, and prioritizing it high, as they need to go through all their pages and make sure the right images are loading. | 1.0 | [OG Image] Get working - I think this is still unsolved? Or I missed what's up on slack. Putting an image out there, and prioritizing it high, as they need to go through all their pages and make sure the right images are loading. | priority | get working i think this is still unsolved or i missed what s up on slack putting an image out there and prioritizing it high as they need to go through all their pages and make sure the right images are loading | 1 |
233,217 | 7,695,554,208 | IssuesEvent | 2018-05-18 12:44:19 | bitshares/bitshares-ui | https://api.github.com/repos/bitshares/bitshares-ui | closed | [3][happyconcepts] Order Expiration Date bug | bug high priority | OSX, latest, Chrome Browser
## Problem
It appears that the order expiration is off by the timezone difference. In This case, it is 7:18 pm. I chose a 12 hour expiration. The confirmation modal displays 12:18 AM CDT.

## Expected Behavior
My current time + offset chosen = Time displayed on modal. | 1.0 | [3][happyconcepts] Order Expiration Date bug - OSX, latest, Chrome Browser
## Problem
It appears that the order expiration is off by the timezone difference. In This case, it is 7:18 pm. I chose a 12 hour expiration. The confirmation modal displays 12:18 AM CDT.

## Expected Behavior
My current time + offset chosen = Time displayed on modal. | priority | order expiration date bug osx latest chrome browser problem it appears that the order expiration is off by the timezone difference in this case it is pm i chose a hour expiration the confirmation modal displays am cdt expected behavior my current time offset chosen time displayed on modal | 1 |
713,277 | 24,523,456,429 | IssuesEvent | 2022-10-11 11:21:18 | opencrvs/opencrvs-core | https://api.github.com/repos/opencrvs/opencrvs-core | opened | OpenHIM Channel security improvements | Priority: high | **CORS is not setup in OpenHIM**
It seems to be also possible to access OpenHIM and specifically the wildcard DNS exposes unused traefik routes "api."
The workflow passthrough channel regexp essentially forwards any URL suffix to it.
We need to strictly set the workflow passthrough regexp and update the backend where appropriate when the workflow passthrough is called. See screenshots ...



Run full cypress suite when done | 1.0 | OpenHIM Channel security improvements - **CORS is not setup in OpenHIM**
It seems to be also possible to access OpenHIM and specifically the wildcard DNS exposes unused traefik routes "api."
The workflow passthrough channel regexp essentially forwards any URL suffix to it.
We need to strictly set the workflow passthrough regexp and update the backend where appropriate when the workflow passthrough is called. See screenshots ...



Run full cypress suite when done | priority | openhim channel security improvements cors is not setup in openhim it seems to be also possible to access openhim and specifically the wildcard dns exposes unused traefik routes api the workflow passthrough channel regexp essentially forwards any url suffix to it we need to strictly set the workflow passthrough regexp and update the backend where appropriate when the workflow passthrough is called see screenshots run full cypress suite when done | 1 |
567,405 | 16,857,863,513 | IssuesEvent | 2021-06-21 09:09:36 | PlaceOS/user-interfaces | https://api.github.com/repos/PlaceOS/user-interfaces | closed | Workplace template errors when booking a desk from explore view | Priority: High Type: Bug focus: front end | When attempting to book a desk from explore view, the following error is hit.
https://github.com/PlaceOS/user-interfaces/blob/81f167ff1753f94903168b07c4e5bf0c0594cca5/libs/bookings/src/lib/desk.service.ts#L44
At no point during this flow is the ability to select a host provided to the user. | 1.0 | Workplace template errors when booking a desk from explore view - When attempting to book a desk from explore view, the following error is hit.
https://github.com/PlaceOS/user-interfaces/blob/81f167ff1753f94903168b07c4e5bf0c0594cca5/libs/bookings/src/lib/desk.service.ts#L44
At no point during this flow is the ability to select a host provided to the user. | priority | workplace template errors when booking a desk from explore view when attempting to book a desk from explore view the following error is hit at no point during this flow is the ability to select a host provided to the user | 1 |
584,510 | 17,456,798,026 | IssuesEvent | 2021-08-06 03:18:35 | opsramp/openapi-specs | https://api.github.com/repos/opsramp/openapi-specs | closed | Confirm use of query string object for patching endpoint | bug high priority | The endpoint:
`/api/v2/tenants/{tenantId}/patches`
Starts here:
https://github.com/opsramp/openapi-specs/blob/5ca943366fe3e717050f9ce0ea9df001339c5e76/v2/opsramp-patching.v2.yaml#L1861
Parameter in question:
https://github.com/opsramp/openapi-specs/blob/5ca943366fe3e717050f9ce0ea9df001339c5e76/v2/opsramp-patching.v2.yaml#L1936
This seems to imply a request would look similar to:
`/api/v2/tenants/{tenantId}/patches?pageNo=1&pageSize=100&queryString[name]=TheName&querystring[uid]=00101&queryString[startReleaseDate]=2018-08-19T08%3A30%3A20%2B0000`
Can you confirm what a properly formatted request for this endpoint would look like?
| 1.0 | Confirm use of query string object for patching endpoint - The endpoint:
`/api/v2/tenants/{tenantId}/patches`
Starts here:
https://github.com/opsramp/openapi-specs/blob/5ca943366fe3e717050f9ce0ea9df001339c5e76/v2/opsramp-patching.v2.yaml#L1861
Parameter in question:
https://github.com/opsramp/openapi-specs/blob/5ca943366fe3e717050f9ce0ea9df001339c5e76/v2/opsramp-patching.v2.yaml#L1936
This seems to imply a request would look similar to:
`/api/v2/tenants/{tenantId}/patches?pageNo=1&pageSize=100&queryString[name]=TheName&querystring[uid]=00101&queryString[startReleaseDate]=2018-08-19T08%3A30%3A20%2B0000`
Can you confirm what a properly formatted request for this endpoint would look like?
| priority | confirm use of query string object for patching endpoint the endpoint api tenants tenantid patches starts here parameter in question this seems to imply a request would look similar to api tenants tenantid patches pageno pagesize querystring thename querystring querystring can you confirm what a properly formatted request for this endpoint would look like | 1 |
4,009 | 2,544,669,644 | IssuesEvent | 2015-01-29 11:53:33 | IntellectualCrafters/PlotSquared | https://api.github.com/repos/IntellectualCrafters/PlotSquared | closed | Plot protection system | enhancement high priority | ```
Enhancement
Another well known and quite similar plot management plugin offers in this configuration to define which blocks or items like buttons or levers are protected on plots and which are public accessible.
I really liked this feature because for example you could make buttons made from wood usable for everyone meanwhile stone buttons can only be used from authorized players. So there could be public and private areas.
Another example would be chests and trapped chests. If players want to have a shared chest with everyone on the server without having added each and every player to a plot there could be trapped chests unprotected and common chests protected.
Maybe you wanna give this feature a chance. I would love to see that :).
```
```
Enhancement
Users which have no helper or trust status on a plot can not use quickshop chests to buy or sell items. Pex debug tells me that the plugin asks for plots.admin.interact.other which is obviously not present on common players without any special rights.
Therefore QuickShops cannot be used in a Plotworld seriously.
It would be very nice if this could be fixed.
Maybe this could be combined with #94
``` | 1.0 | Plot protection system - ```
Enhancement
Another well known and quite similar plot management plugin offers in this configuration to define which blocks or items like buttons or levers are protected on plots and which are public accessible.
I really liked this feature because for example you could make buttons made from wood usable for everyone meanwhile stone buttons can only be used from authorized players. So there could be public and private areas.
Another example would be chests and trapped chests. If players want to have a shared chest with everyone on the server without having added each and every player to a plot there could be trapped chests unprotected and common chests protected.
Maybe you wanna give this feature a chance. I would love to see that :).
```
```
Enhancement
Users which have no helper or trust status on a plot can not use quickshop chests to buy or sell items. Pex debug tells me that the plugin asks for plots.admin.interact.other which is obviously not present on common players without any special rights.
Therefore QuickShops cannot be used in a Plotworld seriously.
It would be very nice if this could be fixed.
Maybe this could be combined with #94
``` | priority | plot protection system enhancement another well known and quite similar plot management plugin offers in this configuration to define which blocks or items like buttons or levers are protected on plots and which are public accessible i really liked this feature because for example you could make buttons made from wood usable for everyone meanwhile stone buttons can only be used from authorized players so there could be public and private areas another example would be chests and trapped chests if players want to have a shared chest with everyone on the server without having added each and every player to a plot there could be trapped chests unprotected and common chests protected maybe you wanna give this feature a chance i would love to see that enhancement users which have no helper or trust status on a plot can not use quickshop chests to buy or sell items pex debug tells me that the plugin asks for plots admin interact other which is obviously not present on common players without any special rights therefore quickshops cannot be used in a plotworld seriously it would be very nice if this could be fixed maybe this could be combined with | 1 |
648,744 | 21,193,020,857 | IssuesEvent | 2022-04-08 19:50:05 | NOAA-GSL/MATS | https://api.github.com/repos/NOAA-GSL/MATS | closed | Fix Deploy action failure | Type: Bug Priority: High CI/CD | The deploy action associated with the v4.4.1 release failed due to a syntax error with `GITHUB_REF`. This was believed to have been fixed in #716.
The failing job is here: https://github.com/NOAA-GSL/MATS/runs/5875625209?check_suite_focus=true
According to the GHA docs, a release event should populate the `GITHUB_REF` with the tag: https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#release. Current thoughts are that the variable is quoted incorrectly and/or we need to trim the result from the variable as it returns like so: `refs/tags/<tag_name>`.
This info also exists in the `github.ref` context so we could look at using that instead if it'd be advantageous. https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#release | 1.0 | Fix Deploy action failure - The deploy action associated with the v4.4.1 release failed due to a syntax error with `GITHUB_REF`. This was believed to have been fixed in #716.
The failing job is here: https://github.com/NOAA-GSL/MATS/runs/5875625209?check_suite_focus=true
According to the GHA docs, a release event should populate the `GITHUB_REF` with the tag: https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#release. Current thoughts are that the variable is quoted incorrectly and/or we need to trim the result from the variable as it returns like so: `refs/tags/<tag_name>`.
This info also exists in the `github.ref` context so we could look at using that instead if it'd be advantageous. https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#release | priority | fix deploy action failure the deploy action associated with the release failed due to a syntax error with github ref this was believed to have been fixed in the failing job is here according to the gha docs a release event should populate the github ref with the tag current thoughts are that the variable is quoted incorrectly and or we need to trim the result from the variable as it returns like so refs tags this info also exists in the github ref context so we could look at using that instead if it d be advantageous | 1 |
388,100 | 11,474,539,802 | IssuesEvent | 2020-02-10 04:44:25 | openmsupply/mobile | https://api.github.com/repos/openmsupply/mobile | closed | Supplier Credits straight from a Supplier Invoice | Docs: not needed Effort: small Feature Module: dispensary Priority: high | ## Is your feature request related to a problem? Please describe.
Can create a supplier credit from the current stock page. Should be able to create one straight from an SI
## Describe the solution you'd like
As above
## Implementation
N/A
## Describe alternatives you've considered
N/A
## Additional context
N/A
| 1.0 | Supplier Credits straight from a Supplier Invoice - ## Is your feature request related to a problem? Please describe.
Can create a supplier credit from the current stock page. Should be able to create one straight from an SI
## Describe the solution you'd like
As above
## Implementation
N/A
## Describe alternatives you've considered
N/A
## Additional context
N/A
| priority | supplier credits straight from a supplier invoice is your feature request related to a problem please describe can create a supplier credit from the current stock page should be able to create one straight from an si describe the solution you d like as above implementation n a describe alternatives you ve considered n a additional context n a | 1 |
622,501 | 19,640,366,255 | IssuesEvent | 2022-01-09 00:02:24 | SourceWriters/SmoothTimber | https://api.github.com/repos/SourceWriters/SmoothTimber | closed | [REVISION] Remove the permission handle of the SmoothTimber command from the plugin.yml | Done Priority: High Revision | **Is your revision request related to a problem? Please describe.**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
Currently bukkit is handling the first permission check which resolves in the "You are not permitted to execute this command" permission message.
However SmoothTimber itself does also a permission check before redirecting to any subcommand and if the permission isn't set it would send a custom permission message which it currently doesn't as it is set in the plugin.yml
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
Update the plugin.yml to not contain the permission node for the "smoothtimber" command
**Additional context**
Only the plugin.yml is changed
Related to: #32
| 1.0 | [REVISION] Remove the permission handle of the SmoothTimber command from the plugin.yml - **Is your revision request related to a problem? Please describe.**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
Currently bukkit is handling the first permission check which resolves in the "You are not permitted to execute this command" permission message.
However SmoothTimber itself does also a permission check before redirecting to any subcommand and if the permission isn't set it would send a custom permission message which it currently doesn't as it is set in the plugin.yml
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
Update the plugin.yml to not contain the permission node for the "smoothtimber" command
**Additional context**
Only the plugin.yml is changed
Related to: #32
| priority | remove the permission handle of the smoothtimber command from the plugin yml is your revision request related to a problem please describe currently bukkit is handling the first permission check which resolves in the you are not permitted to execute this command permission message however smoothtimber itself does also a permission check before redirecting to any subcommand and if the permission isn t set it would send a custom permission message which it currently doesn t as it is set in the plugin yml describe the solution you d like update the plugin yml to not contain the permission node for the smoothtimber command additional context only the plugin yml is changed related to | 1 |
219,687 | 7,344,969,801 | IssuesEvent | 2018-03-07 16:06:23 | prettier/prettier | https://api.github.com/repos/prettier/prettier | closed | In Markdown Files, Nested Code Block/Raw HTML Indent For Each Format on Save | lang:markdown priority:high status:has pr type:bug | I apologize if this is already an open ticket, I didn't see anything for this in particular. The issue is appearing in markdown files. When attempting to nest code blocks or raw HTML the format on save feature will add more spacing each time the file is saved which breaks the indentation. When using code blocks the issue is intermittent where once the code is nested correctly it will not break again but with raw HTML the only way to prevent the issues is to turn off format on save or ignore markdown files, neither of which is a solution. Any thoughts on a setting that may help are appreciated.
p.s. I'm using Prettier in VS Code.

**Input:**
1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync.
<table class="table table-striped">
<tr>
<th>Test</th>
<th>Table</th>
</tr>
<tbody>
<tr>
<td>will</td>
<td>be</td>
</tr>
<tr>
<td>pushed</td>
<td>When</td>
</tr>
<tr>
<td>Format on</td>
<td>Save</td>
</tr>
</tbody>
</table>
**Output on first save:**
1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync.
<table class="table table-striped">
<tr>
<th>Test</th>
<th>Table</th>
</tr>
<tbody>
<tr>
<td>will</td>
<td>be</td>
</tr>
<tr>
<td>pushed</td>
<td>When</td>
</tr>
<tr>
<td>Format on</td>
<td>Save</td>
</tr>
</tbody>
</table>
**Output on fourth save:**
1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync.
<table class="table table-striped">
<tr>
<th>Test</th>
<th>Table</th>
</tr>
<tbody>
<tr>
<td>will</td>
<td>be</td>
</tr>
<tr>
<td>pushed</td>
<td>When</td>
</tr>
<tr>
<td>Format on</td>
<td>Save</td>
</tr>
</tbody>
</table>
**Expected behavior:**
1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync.
<table class="table table-striped">
<tr>
<th>Test</th>
<th>Table</th>
</tr>
<tbody>
<tr>
<td>will</td>
<td>be</td>
</tr>
<tr>
<td>pushed</td>
<td>When</td>
</tr>
<tr>
<td>Format on</td>
<td>Save</td>
</tr>
</tbody>
</table>
| 1.0 | In Markdown Files, Nested Code Block/Raw HTML Indent For Each Format on Save - I apologize if this is already an open ticket, I didn't see anything for this in particular. The issue is appearing in markdown files. When attempting to nest code blocks or raw HTML the format on save feature will add more spacing each time the file is saved which breaks the indentation. When using code blocks the issue is intermittent where once the code is nested correctly it will not break again but with raw HTML the only way to prevent the issues is to turn off format on save or ignore markdown files, neither of which is a solution. Any thoughts on a setting that may help are appreciated.
p.s. I'm using Prettier in VS Code.

**Input:**
1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync.
<table class="table table-striped">
<tr>
<th>Test</th>
<th>Table</th>
</tr>
<tbody>
<tr>
<td>will</td>
<td>be</td>
</tr>
<tr>
<td>pushed</td>
<td>When</td>
</tr>
<tr>
<td>Format on</td>
<td>Save</td>
</tr>
</tbody>
</table>
**Output on first save:**
1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync.
<table class="table table-striped">
<tr>
<th>Test</th>
<th>Table</th>
</tr>
<tbody>
<tr>
<td>will</td>
<td>be</td>
</tr>
<tr>
<td>pushed</td>
<td>When</td>
</tr>
<tr>
<td>Format on</td>
<td>Save</td>
</tr>
</tbody>
</table>
**Output on fourth save:**
1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync.
<table class="table table-striped">
<tr>
<th>Test</th>
<th>Table</th>
</tr>
<tbody>
<tr>
<td>will</td>
<td>be</td>
</tr>
<tr>
<td>pushed</td>
<td>When</td>
</tr>
<tr>
<td>Format on</td>
<td>Save</td>
</tr>
</tbody>
</table>
**Expected behavior:**
1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync.
<table class="table table-striped">
<tr>
<th>Test</th>
<th>Table</th>
</tr>
<tbody>
<tr>
<td>will</td>
<td>be</td>
</tr>
<tr>
<td>pushed</td>
<td>When</td>
</tr>
<tr>
<td>Format on</td>
<td>Save</td>
</tr>
</tbody>
</table>
| priority | in markdown files nested code block raw html indent for each format on save i apologize if this is already an open ticket i didn t see anything for this in particular the issue is appearing in markdown files when attempting to nest code blocks or raw html the format on save feature will add more spacing each time the file is saved which breaks the indentation when using code blocks the issue is intermittent where once the code is nested correctly it will not break again but with raw html the only way to prevent the issues is to turn off format on save or ignore markdown files neither of which is a solution any thoughts on a setting that may help are appreciated p s i m using prettier in vs code input some test text the goal is to have the html table below nested within this number when formatting on save prettier will continue to add an indent each time pushing the table further and further out of sync test table will be pushed when format on save output on first save some test text the goal is to have the html table below nested within this number when formatting on save prettier will continue to add an indent each time pushing the table further and further out of sync test table will be pushed when format on save output on fourth save some test text the goal is to have the html table below nested within this number when formatting on save prettier will continue to add an indent each time pushing the table further and further out of sync test table will be pushed when format on save expected behavior some test text the goal is to have the html table below nested within this number when formatting on save prettier will continue to add an indent each time pushing the table further and further out of sync test table will be pushed when format on save | 1 |
310,696 | 9,523,024,288 | IssuesEvent | 2019-04-27 13:57:38 | NOVACProject/NOVACProgram | https://api.github.com/repos/NOVACProject/NOVACProgram | closed | Add Support for SFTP | enhancement high priority | The new upload server at Chalmers will use SFTP protocol once implemented. This change is required before we can switch to the new server. | 1.0 | Add Support for SFTP - The new upload server at Chalmers will use SFTP protocol once implemented. This change is required before we can switch to the new server. | priority | add support for sftp the new upload server at chalmers will use sftp protocol once implemented this change is required before we can switch to the new server | 1 |
1,313 | 2,511,569,656 | IssuesEvent | 2015-01-14 09:19:04 | olga-jane/prizm | https://api.github.com/repos/olga-jane/prizm | opened | Export/Import set flag "to export" in proper places | Coding Coding - POOL COMMON HIGH priority | Google document PRIZMA Design contains 2 diagrams in Data export/import section.
These diagrams show, when and at what conditions should be set the flag "to export".
DB column exists already. | 1.0 | Export/Import set flag "to export" in proper places - Google document PRIZMA Design contains 2 diagrams in Data export/import section.
These diagrams show, when and at what conditions should be set the flag "to export".
DB column exists already. | priority | export import set flag to export in proper places google document prizma design contains diagrams in data export import section these diagrams show when and at what conditions should be set the flag to export db column exists already | 1 |
582,212 | 17,356,087,333 | IssuesEvent | 2021-07-29 14:33:19 | staynomad/Nomad-Back | https://api.github.com/repos/staynomad/Nomad-Back | closed | Cannot Create a Listing | dev:bug difficulty:medium priority:high | # Background
<!--- Put any relevant background information here. --->
This error pops up when I try to save the listing I'm creating as a draft

Same error when I submit:

# Task
<!--- Put the task here (ideally bullet points). --->
* Fix the route and make it correctly save the draft
* Fix the route and make it correctly submit the listing
# Done When
<!--- Put the completion criteria for the issue here. --->
* Loading spinner doesn't spin indefinitely
| 1.0 | Cannot Create a Listing - # Background
<!--- Put any relevant background information here. --->
This error pops up when I try to save the listing I'm creating as a draft

Same error when I submit:

# Task
<!--- Put the task here (ideally bullet points). --->
* Fix the route and make it correctly save the draft
* Fix the route and make it correctly submit the listing
# Done When
<!--- Put the completion criteria for the issue here. --->
* Loading spinner doesn't spin indefinitely
| priority | cannot create a listing background this error pops up when i try to save the listing i m creating as a draft same error when i submit task fix the route and make it correctly save the draft fix the route and make it correctly submit the listing done when loading spinner doesn t spin indefinitely | 1 |
533,634 | 15,595,677,285 | IssuesEvent | 2021-03-18 15:06:55 | HDFGroup/hermes | https://api.github.com/repos/HDFGroup/hermes | closed | Go to neighborhood Targets if a Put fails on local Targets | priority-high | Add tests using the cluster simulator. | 1.0 | Go to neighborhood Targets if a Put fails on local Targets - Add tests using the cluster simulator. | priority | go to neighborhood targets if a put fails on local targets add tests using the cluster simulator | 1 |
307,263 | 9,415,131,507 | IssuesEvent | 2019-04-10 11:55:11 | meumobi/sitebuilder | https://api.github.com/repos/meumobi/sitebuilder | closed | Enable apache log rotation on it.c-se | high priority | ### Expected behaviour
On it.c-se env
#### Enable apache log rotation;
on */etc/apache2/sites-available/{site}*
```
CustomLog "/var/log/apache2/{site}/access.%Y-%m-%d" combined
```
```$ a2ensite {site}```
#### Track log on scalyr
on */etc/scalyr-agent-2/agent.json*
```
{
path: "/var/log/apache2/{site}/access.*",
attributes: {parser: "accessLog"},
rename_logfile: {
match: "/var/log/apache2/{site}/access.([0-9]{4}-[0-9]{2}-[0-9]{2})",
replacement: "/scalyr/access.log"
}
}
```
```$ sudo scalyr-agent-2 restart```
### Actual behaviour
- There is one single log file
### How to test
- open it app
- perform some requests
- certify if the requests were sent to scalyr
| 1.0 | Enable apache log rotation on it.c-se - ### Expected behaviour
On it.c-se env
#### Enable apache log rotation;
on */etc/apache2/sites-available/{site}*
```
CustomLog "/var/log/apache2/{site}/access.%Y-%m-%d" combined
```
```$ a2ensite {site}```
#### Track log on scalyr
on */etc/scalyr-agent-2/agent.json*
```
{
path: "/var/log/apache2/{site}/access.*",
attributes: {parser: "accessLog"},
rename_logfile: {
match: "/var/log/apache2/{site}/access.([0-9]{4}-[0-9]{2}-[0-9]{2})",
replacement: "/scalyr/access.log"
}
}
```
```$ sudo scalyr-agent-2 restart```
### Actual behaviour
- There is one single log file
### How to test
- open it app
- perform some requests
- certify if the requests were sent to scalyr
| priority | enable apache log rotation on it c se expected behaviour on it c se env enable apache log rotation on etc sites available site customlog var log site access y m d combined site track log on scalyr on etc scalyr agent agent json path var log site access attributes parser accesslog rename logfile match var log site access replacement scalyr access log sudo scalyr agent restart actual behaviour there is one single log file how to test open it app perform some requests certify if the requests were sent to scalyr | 1 |
803,315 | 29,171,645,475 | IssuesEvent | 2023-05-19 02:44:25 | mlcommons/medperf | https://api.github.com/repos/mlcommons/medperf | opened | [FEATURE] Integration tests need to consider every possible command | enhancement High Priority | We should add integration tests for all possible CLI commands. | 1.0 | [FEATURE] Integration tests need to consider every possible command - We should add integration tests for all possible CLI commands. | priority | integration tests need to consider every possible command we should add integration tests for all possible cli commands | 1 |
220,929 | 7,372,243,961 | IssuesEvent | 2018-03-13 14:16:44 | SANBIBiodiversityforLife/species | https://api.github.com/repos/SANBIBiodiversityforLife/species | closed | List of taxa without images and maps within each group | high-priority | Dewidine wants a list of taxa which don't have images, and a list of taxa which don't have any distribution info/maps. | 1.0 | List of taxa without images and maps within each group - Dewidine wants a list of taxa which don't have images, and a list of taxa which don't have any distribution info/maps. | priority | list of taxa without images and maps within each group dewidine wants a list of taxa which don t have images and a list of taxa which don t have any distribution info maps | 1 |
349,884 | 10,475,226,755 | IssuesEvent | 2019-09-23 15:54:26 | electrala/documentation | https://api.github.com/repos/electrala/documentation | opened | AAD - I want to create the official db for electra. | Database backend high priority | We need to go into postgres and make the following tables to use in our final project.
Critiques table
User table
comments table
Gallery(optional not high priority)
Be extra careful and double check everything when creating this, it will be our final db. | 1.0 | AAD - I want to create the official db for electra. - We need to go into postgres and make the following tables to use in our final project.
Critiques table
User table
comments table
Gallery(optional not high priority)
Be extra careful and double check everything when creating this, it will be our final db. | priority | aad i want to create the official db for electra we need to go into postgres and make the following tables to use in our final project critiques table user table comments table gallery optional not high priority be extra careful and double check everything when creating this it will be our final db | 1 |
509,328 | 14,728,460,751 | IssuesEvent | 2021-01-06 09:59:25 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | web.facebook.com - desktop site instead of mobile site | browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical | <!-- @browser: Firefox 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:85.0) Gecko/20100101 Firefox/85.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/65044 -->
**URL**: https://web.facebook.com/
**Browser / Version**: Firefox 85.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/207c5eaf-995d-4723-9933-b73d1e38d86a.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210105185604</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/e967b02c-d117-45df-95d9-a3c9125e0413)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | web.facebook.com - desktop site instead of mobile site - <!-- @browser: Firefox 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:85.0) Gecko/20100101 Firefox/85.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/65044 -->
**URL**: https://web.facebook.com/
**Browser / Version**: Firefox 85.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/207c5eaf-995d-4723-9933-b73d1e38d86a.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210105185604</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/e967b02c-d117-45df-95d9-a3c9125e0413)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | web facebook com desktop site instead of mobile site url browser version firefox operating system windows tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
788,410 | 27,752,094,174 | IssuesEvent | 2023-03-15 21:41:25 | nasa/prog_models | https://api.github.com/repos/nasa/prog_models | opened | Divide by Zero in calc_error | bug Priority: High | Identified by @aqitya
PrognosticsModel.calc_error() throws a divide by zero exception when the model is configured in a way that is unstable (i.e., a state becomes NaN) in the first step (e.g., BatteryElectroChemEOD with a qMax of 4000). See https://github.com/nasa/prog_models/blob/f0891d59e4afd726f12812d273f6d97e7abc7a9f/src/prog_models/prognostics_model.py#L1176-L1178
This is of course undesired behavior. I see three options here:
1. if the model goes unstable at any point return NaN error (or Inf if NaN isn't treated correctly by estimate_params)
2. Of the model goes unstable before some cutoff (e.g., 90% of the points have passed), return NaN error. If it goes unstable at any other point, warn and return an estimate of error from the points to that point.
3. Only if the model goes unstable at the first timestep, return NaN error. If it goes unstable at any other point, warn and return an estimate of error from the points to that point.
We will want to make sure consider the following cases in our decision:
1. The case where this was first identified, the model is never stable so we get NaN on the first timestep, then we definitely want to return Nan (or Inf, see message above)
2. Case where we're running estimate_params and the model goes unstable midway through. Not sure what we want to do here
3. Case where we're running estimate_params and the model goes unstable at the very last timestep. This is frequently the case when training on run-to-failure data. In this case the parameters may be very close to the "correct" or best values.
4. Case where we're running estimate_params and the model never goes unstable- this one we obviously want to work
We also should keep in mind that calc_error can be used outside of estimate_params | 1.0 | Divide by Zero in calc_error - Identified by @aqitya
PrognosticsModel.calc_error() throws a divide by zero exception when the model is configured in a way that is unstable (i.e., a state becomes NaN) in the first step (e.g., BatteryElectroChemEOD with a qMax of 4000). See https://github.com/nasa/prog_models/blob/f0891d59e4afd726f12812d273f6d97e7abc7a9f/src/prog_models/prognostics_model.py#L1176-L1178
This is of course undesired behavior. I see three options here:
1. if the model goes unstable at any point return NaN error (or Inf if NaN isn't treated correctly by estimate_params)
2. Of the model goes unstable before some cutoff (e.g., 90% of the points have passed), return NaN error. If it goes unstable at any other point, warn and return an estimate of error from the points to that point.
3. Only if the model goes unstable at the first timestep, return NaN error. If it goes unstable at any other point, warn and return an estimate of error from the points to that point.
We will want to make sure consider the following cases in our decision:
1. The case where this was first identified, the model is never stable so we get NaN on the first timestep, then we definitely want to return Nan (or Inf, see message above)
2. Case where we're running estimate_params and the model goes unstable midway through. Not sure what we want to do here
3. Case where we're running estimate_params and the model goes unstable at the very last timestep. This is frequently the case when training on run-to-failure data. In this case the parameters may be very close to the "correct" or best values.
4. Case where we're running estimate_params and the model never goes unstable- this one we obviously want to work
We also should keep in mind that calc_error can be used outside of estimate_params | priority | divide by zero in calc error identified by aqitya prognosticsmodel calc error throws a divide by zero exception when the model is configured in a way that is unstable i e a state becomes nan in the first step e g batteryelectrochemeod with a qmax of see this is of course undesired behavior i see three options here if the model goes unstable at any point return nan error or inf if nan isn t treated correctly by estimate params of the model goes unstable before some cutoff e g of the points have passed return nan error if it goes unstable at any other point warn and return an estimate of error from the points to that point only if the model goes unstable at the first timestep return nan error if it goes unstable at any other point warn and return an estimate of error from the points to that point we will want to make sure consider the following cases in our decision the case where this was first identified the model is never stable so we get nan on the first timestep then we definitely want to return nan or inf see message above case where we re running estimate params and the model goes unstable midway through not sure what we want to do here case where we re running estimate params and the model goes unstable at the very last timestep this is frequently the case when training on run to failure data in this case the parameters may be very close to the correct or best values case where we re running estimate params and the model never goes unstable this one we obviously want to work we also should keep in mind that calc error can be used outside of estimate params | 1 |
265,125 | 8,337,590,821 | IssuesEvent | 2018-09-28 11:41:58 | ThmmyNoLife/mTHMMY | https://api.github.com/repos/ThmmyNoLife/mTHMMY | closed | crash on orientation change in topic | bug high priority | ### Steps to reproduce the problem
1. Open a topic
2. Try to edit a post or write a reply
3. Change the orientation to landscape
4. Change the orientation back to portrait
5 App crashes
### What causes the crash
[AutoFitGridLayout](https://github.com/ThmmyNoLife/mTHMMY/blob/develop/app/src/main/java/gr/thmmy/mthmmy/editorview/AutoFitGridLayout.java#L44) auto-changes its columnCount to fit as many items as it can in each row. However, when the app goes from portrait to landscape, this value gets smaller, while some of the layout's children have the old layout params which may have larger column index, causeing the crash
### How to fix this
Replace [AutoFitGridLayout](https://github.com/ThmmyNoLife/mTHMMY/blob/develop/app/src/main/java/gr/thmmy/mthmmy/editorview/AutoFitGridLayout.java) with a recyclerview that uses a GridLayoutManager | 1.0 | crash on orientation change in topic - ### Steps to reproduce the problem
1. Open a topic
2. Try to edit a post or write a reply
3. Change the orientation to landscape
4. Change the orientation back to portrait
5 App crashes
### What causes the crash
[AutoFitGridLayout](https://github.com/ThmmyNoLife/mTHMMY/blob/develop/app/src/main/java/gr/thmmy/mthmmy/editorview/AutoFitGridLayout.java#L44) auto-changes its columnCount to fit as many items as it can in each row. However, when the app goes from portrait to landscape, this value gets smaller, while some of the layout's children have the old layout params which may have larger column index, causeing the crash
### How to fix this
Replace [AutoFitGridLayout](https://github.com/ThmmyNoLife/mTHMMY/blob/develop/app/src/main/java/gr/thmmy/mthmmy/editorview/AutoFitGridLayout.java) with a recyclerview that uses a GridLayoutManager | priority | crash on orientation change in topic steps to reproduce the problem open a topic try to edit a post or write a reply change the orientation to landscape change the orientation back to portrait app crashes what causes the crash auto changes its columncount to fit as many items as it can in each row however when the app goes from portrait to landscape this value gets smaller while some of the layout s children have the old layout params which may have larger column index causeing the crash how to fix this replace with a recyclerview that uses a gridlayoutmanager | 1 |
113,385 | 4,553,199,043 | IssuesEvent | 2016-09-13 03:12:48 | FelipeRosa/clojure-gsgp | https://api.github.com/repos/FelipeRosa/clojure-gsgp | closed | Implement data exporters | Priority: High Type: Feature Request | We need data exporters (.txt, .csv, etc) in order to plot prediction results. | 1.0 | Implement data exporters - We need data exporters (.txt, .csv, etc) in order to plot prediction results. | priority | implement data exporters we need data exporters txt csv etc in order to plot prediction results | 1 |
83,688 | 3,640,755,389 | IssuesEvent | 2016-02-13 03:55:17 | RestComm/RestComm-Core | https://api.github.com/repos/RestComm/RestComm-Core | opened | Add support for Japanese language in RVD Say verb | High-Priority Task | Add Japanese option to list of languages in RVD Say verb. | 1.0 | Add support for Japanese language in RVD Say verb - Add Japanese option to list of languages in RVD Say verb. | priority | add support for japanese language in rvd say verb add japanese option to list of languages in rvd say verb | 1 |
392,098 | 11,583,522,021 | IssuesEvent | 2020-02-22 11:37:40 | WoWManiaUK/Blackwing-Lair | https://api.github.com/repos/WoWManiaUK/Blackwing-Lair | closed | [Boss] Omnotron Defense System - Blackwing Descent | Dungeon/Raid Fixed in Dev Priority-High | **Links:**
https://www.wowhead.com/npc=42180/toxitron#comments
**What is happening:**
Nothing worked with this fight
1) the mobs healed faster than we could do damage
2) only two were killable with GM commands
3) they never moved
4)tank and spank didn't do any ability's even when we tried to kill them before using GM commands
5) **it stops the progression of the raid** because the gates wouldn't open without gm on
**What should happen:**
When the fight starts, only one mob will be active, and three stone golems will be frozen near it. Like most council fights, there is a shared HP pool. you start by attacking the active one and when it reaches 50% energy a second one becomes active with 100% energy when this one hits 50% it will activate a shield and the next one become targetable shutting the first one down. and so on until the end of the fight
the activation is random, but it will always be one of the two that is is inactive
Golems have 100 Energy when first activated, and the way the Energy works is the same for all of them:
100 Energy= Golem just activated
50 Energy= Next Golem is activated
45ish Energy= Golem puts up his Shield
25ish Energy = Golem's shield falls
0 Energy = Golem is shut down, becomes stone and untargetable.
(Zero energy on your current target means 50 energy on the other active golem, so a new golem will be active).
**Magmatron: abilities** -
Incineration Security Measure: Basic fire AOE.
Acquiring Target: Picks a target, with a giant red laser to let you know who it is. After a few seconds, Magmatron Turns and blasts a cone of Fire AOE at the target.
Barrier: This is his Shield, absorbs damage then nukes the raid if it takes too much.
https://www.wowhead.com/spell=78740 -activated
https://www.wowhead.com/spell=78726 -inactive
https://www.wowhead.com/spell=79023 -security measures
https://www.wowhead.com/spell=78746 shutting down
**Arcanotron: Abilities** -
Power Generator: Puts this down on the ground. Anyone in it gets a damage buff and mana regen buff. Don't let the boss stand in it. Think Steelbreaker from Ulduar. The fight being long, these are essential to healers getting mana back.
Arcane Annihlator: Fast basic cast on anyone. Can be interrupted but is fast enough to be difficult to do so.
Power Conversion: This is his Shield. Each attack gives him a buff increasing his damage. Spell Stealabl
https://www.wowhead.com/spell=78740 - activated
https://www.wowhead.com/spell=79710 -arcane anhillitaor
https://www.wowhead.com/spell=78726 - inactive
https://www.wowhead.com/spell=78746 -shutting down
**Toxitron abilities:**
Poison Cloud: anyone in it takes 50 percent damage, including the boss
Poison Protocol: summons three slimes. These seek out random targets, and if they hit them, explode for massive damage. This can include Melee and the offtank, If it hits a target, it leaves a poison pool on the ground. The adds don't have much HP and are pretty slow.
Poison Soaked Shell: This is his Shield.Each time you attack you get a stacking dot. However each stack causes you to inflict extra damage.
https://www.wowhead.com/spell=78740 -activates
https://www.wowhead.com/spell=80157 -chemical boss
https://www.wowhead.com/spell=78726 -inactive
https://www.wowhead.com/spell=80053 -posion protocol
https://www.wowhead.com/spell=78746 -shutting down
**Electron abilties**
Single Target Ability - (Electrical Discharge) - Randomly targets a raid member and zaps them with a chain lightning for ~30k damage and jumps to up to three targets for +20% damage each jump.
AoE Ability - (Lightning Conductor) - This is an ability that looks like an Ele Shaman's Thunderstorm. It will be cast on a random raid member... Including both tanks and any DPS. . It hits everyone within range for ~20k damage per second.
Shield Ability - (Unstable Shield) - This is his shield ability, and if too much damage is done to him while this is active it will do an empowered Electrical Discharge type thing. It's a raid wiper if this hurts anyone. It's ~60k damage per second to anyone in range.
https://www.wowhead.com/spell=78740 -activate
https://www.wowhead.com/spell=79879 -electrical discharge
https://www.wowhead.com/spell=78726 -inactive
https://www.wowhead.com/spell=79888 -lighnigh conductor
https://www.wowhead.com/spell=78746 - shutting down
| 1.0 | [Boss] Omnotron Defense System - Blackwing Descent - **Links:**
https://www.wowhead.com/npc=42180/toxitron#comments
**What is happening:**
Nothing worked with this fight
1) the mobs healed faster than we could do damage
2) only two were killable with GM commands
3) they never moved
4)tank and spank didn't do any ability's even when we tried to kill them before using GM commands
5) **it stops the progression of the raid** because the gates wouldn't open without gm on
**What should happen:**
When the fight starts, only one mob will be active, and three stone golems will be frozen near it. Like most council fights, there is a shared HP pool. you start by attacking the active one and when it reaches 50% energy a second one becomes active with 100% energy when this one hits 50% it will activate a shield and the next one become targetable shutting the first one down. and so on until the end of the fight
the activation is random, but it will always be one of the two that is is inactive
Golems have 100 Energy when first activated, and the way the Energy works is the same for all of them:
100 Energy= Golem just activated
50 Energy= Next Golem is activated
45ish Energy= Golem puts up his Shield
25ish Energy = Golem's shield falls
0 Energy = Golem is shut down, becomes stone and untargetable.
(Zero energy on your current target means 50 energy on the other active golem, so a new golem will be active).
**Magmatron: abilities** -
Incineration Security Measure: Basic fire AOE.
Acquiring Target: Picks a target, with a giant red laser to let you know who it is. After a few seconds, Magmatron Turns and blasts a cone of Fire AOE at the target.
Barrier: This is his Shield, absorbs damage then nukes the raid if it takes too much.
https://www.wowhead.com/spell=78740 -activated
https://www.wowhead.com/spell=78726 -inactive
https://www.wowhead.com/spell=79023 -security measures
https://www.wowhead.com/spell=78746 shutting down
**Arcanotron: Abilities** -
Power Generator: Puts this down on the ground. Anyone in it gets a damage buff and mana regen buff. Don't let the boss stand in it. Think Steelbreaker from Ulduar. The fight being long, these are essential to healers getting mana back.
Arcane Annihlator: Fast basic cast on anyone. Can be interrupted but is fast enough to be difficult to do so.
Power Conversion: This is his Shield. Each attack gives him a buff increasing his damage. Spell Stealabl
https://www.wowhead.com/spell=78740 - activated
https://www.wowhead.com/spell=79710 -arcane anhillitaor
https://www.wowhead.com/spell=78726 - inactive
https://www.wowhead.com/spell=78746 -shutting down
**Toxitron abilities:**
Poison Cloud: anyone in it takes 50 percent damage, including the boss
Poison Protocol: summons three slimes. These seek out random targets, and if they hit them, explode for massive damage. This can include Melee and the offtank, If it hits a target, it leaves a poison pool on the ground. The adds don't have much HP and are pretty slow.
Poison Soaked Shell: This is his Shield.Each time you attack you get a stacking dot. However each stack causes you to inflict extra damage.
https://www.wowhead.com/spell=78740 -activates
https://www.wowhead.com/spell=80157 -chemical boss
https://www.wowhead.com/spell=78726 -inactive
https://www.wowhead.com/spell=80053 -posion protocol
https://www.wowhead.com/spell=78746 -shutting down
**Electron abilties**
Single Target Ability - (Electrical Discharge) - Randomly targets a raid member and zaps them with a chain lightning for ~30k damage and jumps to up to three targets for +20% damage each jump.
AoE Ability - (Lightning Conductor) - This is an ability that looks like an Ele Shaman's Thunderstorm. It will be cast on a random raid member... Including both tanks and any DPS. . It hits everyone within range for ~20k damage per second.
Shield Ability - (Unstable Shield) - This is his shield ability, and if too much damage is done to him while this is active it will do an empowered Electrical Discharge type thing. It's a raid wiper if this hurts anyone. It's ~60k damage per second to anyone in range.
https://www.wowhead.com/spell=78740 -activate
https://www.wowhead.com/spell=79879 -electrical discharge
https://www.wowhead.com/spell=78726 -inactive
https://www.wowhead.com/spell=79888 -lighnigh conductor
https://www.wowhead.com/spell=78746 - shutting down
| priority | omnotron defense system blackwing descent links what is happening nothing worked with this fight the mobs healed faster than we could do damage only two were killable with gm commands they never moved tank and spank didn t do any ability s even when we tried to kill them before using gm commands it stops the progression of the raid because the gates wouldn t open without gm on what should happen when the fight starts only one mob will be active and three stone golems will be frozen near it like most council fights there is a shared hp pool you start by attacking the active one and when it reaches energy a second one becomes active with energy when this one hits it will activate a shield and the next one become targetable shutting the first one down and so on until the end of the fight the activation is random but it will always be one of the two that is is inactive golems have energy when first activated and the way the energy works is the same for all of them energy golem just activated energy next golem is activated energy golem puts up his shield energy golem s shield falls energy golem is shut down becomes stone and untargetable zero energy on your current target means energy on the other active golem so a new golem will be active magmatron abilities incineration security measure basic fire aoe acquiring target picks a target with a giant red laser to let you know who it is after a few seconds magmatron turns and blasts a cone of fire aoe at the target barrier this is his shield absorbs damage then nukes the raid if it takes too much activated inactive security measures shutting down arcanotron abilities power generator puts this down on the ground anyone in it gets a damage buff and mana regen buff don t let the boss stand in it think steelbreaker from ulduar the fight being long these are essential to healers getting mana back arcane annihlator fast basic cast on anyone can be interrupted but is fast enough to be difficult to do so power conversion this is his shield each attack gives him a buff increasing his damage spell stealabl activated arcane anhillitaor inactive shutting down toxitron abilities poison cloud anyone in it takes percent damage including the boss poison protocol summons three slimes these seek out random targets and if they hit them explode for massive damage this can include melee and the offtank if it hits a target it leaves a poison pool on the ground the adds don t have much hp and are pretty slow poison soaked shell this is his shield each time you attack you get a stacking dot however each stack causes you to inflict extra damage activates chemical boss inactive posion protocol shutting down electron abilties single target ability electrical discharge randomly targets a raid member and zaps them with a chain lightning for damage and jumps to up to three targets for damage each jump aoe ability lightning conductor this is an ability that looks like an ele shaman s thunderstorm it will be cast on a random raid member including both tanks and any dps it hits everyone within range for damage per second shield ability unstable shield this is his shield ability and if too much damage is done to him while this is active it will do an empowered electrical discharge type thing it s a raid wiper if this hurts anyone it s damage per second to anyone in range activate electrical discharge inactive lighnigh conductor shutting down | 1 |
396,680 | 11,712,175,692 | IssuesEvent | 2020-03-09 07:40:35 | AY1920S2-CS2103T-T10-2/main | https://api.github.com/repos/AY1920S2-CS2103T-T10-2/main | closed | Add ability to save user answers to corresponding flashcards | priority.High type.Task | For use later, when user needs to view their answer history/statistics | 1.0 | Add ability to save user answers to corresponding flashcards - For use later, when user needs to view their answer history/statistics | priority | add ability to save user answers to corresponding flashcards for use later when user needs to view their answer history statistics | 1 |
209,473 | 7,176,521,177 | IssuesEvent | 2018-01-31 10:18:56 | VALIDproject/netflower | https://api.github.com/repos/VALIDproject/netflower | closed | BUG with Filters and update time | bug high priority | - Filters update not appropriately
- Filters are not applying to all and some are to restrictive
- The data isn't fully correct
- The UI should update as well appropiately | 1.0 | BUG with Filters and update time - - Filters update not appropriately
- Filters are not applying to all and some are to restrictive
- The data isn't fully correct
- The UI should update as well appropiately | priority | bug with filters and update time filters update not appropriately filters are not applying to all and some are to restrictive the data isn t fully correct the ui should update as well appropiately | 1 |
524,070 | 15,195,403,564 | IssuesEvent | 2021-02-16 06:16:10 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.3.0 beta staging-1930]Industrial Elevator missing properties in .csv | Category: Balance Priority: High Regression Status: Fixed Type: Bug | Industrial elevator does not have a recipe accompanied as its blank in .csv but populated in techtree.xlsm
There is an inconsistency here if someone didnt properly pull an up to date techtree.csv when updating? | 1.0 | [0.9.3.0 beta staging-1930]Industrial Elevator missing properties in .csv - Industrial elevator does not have a recipe accompanied as its blank in .csv but populated in techtree.xlsm
There is an inconsistency here if someone didnt properly pull an up to date techtree.csv when updating? | priority | industrial elevator missing properties in csv industrial elevator does not have a recipe accompanied as its blank in csv but populated in techtree xlsm there is an inconsistency here if someone didnt properly pull an up to date techtree csv when updating | 1 |
494,675 | 14,263,003,799 | IssuesEvent | 2020-11-20 13:50:25 | reichlab/covidHubUtils | https://api.github.com/repos/reichlab/covidHubUtils | opened | create facet-ready plot_forecasts() function | High Priority | specs via google doc design.
#' @param forecast_data data frame with truth and forecasts from load_forecasts()
#' @param truth_data optional data frame with forecasts in the format returned
#' by load_truth().
#' @param models model_abbr specifying model to plot. Optional if there is only
#' one model available in forecast data.
#' @param target_variable string specifying target type. It should be one of
#' "cum death", "inc case", "inc death"
#' @param locations string for fips code or 'US'. Optional if there is only one
#' location available in forecast data.
#' @param facet interpretable facet option for ggplot
#’ @param facet_scales argument for scales in ggplot2::facet_wrap
#’ @param forecast_dates forecast dates to plot, defaults to forecast_dates present
#’ in the data
#' @param intervals values indicating which central prediction interval levels
#' to plot, defaults to c(.5, .8, .95). NULL means only plotting point forecasts.
#' If not provided, it will default to all available intervals in forecast data.
#' @param horizon forecasts are plotted for the horizon time steps after the
#' forecast date. Default to all available horizons in forecast data.
#' @param truth_source character specifying where the truth data will
#' be loaded from if truth_data is not provided.
#' Otherwise, this character specifies the data source to plot.
#' Currently support "JHU","USAFacts" and "NYTimes".
#' @param truth_as_of the plot includes the truth data that would have been
#' in real time as of the truth_as_of date (not using this parameter when truth data
#' is from github repo)
#' @param title optional text for the title of the plot. if left as "default",
#' the title will be automatically generated. if "none", no title will be plotted.
#' @param show.caption logical, if TRUE, caption will be included showing data sources
#’ @param plot logical; if TRUE, the plot is printed
#'
#' @return invisible ggplot object
#'
#' @export
plot_forecasts <- function(forecast_data,
truth_data = NULL,
models,
target_variable,
locations,
forecast_dates,
facet,
intervals = c(.5, .8, .95),
horizon,
truth_source = "JHU",
truth_as_of = NULL,
title = "default",
show.caption = TRUE,
plot = TRUE) | 1.0 | create facet-ready plot_forecasts() function - specs via google doc design.
#' @param forecast_data data frame with truth and forecasts from load_forecasts()
#' @param truth_data optional data frame with forecasts in the format returned
#' by load_truth().
#' @param models model_abbr specifying model to plot. Optional if there is only
#' one model available in forecast data.
#' @param target_variable string specifying target type. It should be one of
#' "cum death", "inc case", "inc death"
#' @param locations string for fips code or 'US'. Optional if there is only one
#' location available in forecast data.
#' @param facet interpretable facet option for ggplot
#’ @param facet_scales argument for scales in ggplot2::facet_wrap
#’ @param forecast_dates forecast dates to plot, defaults to forecast_dates present
#’ in the data
#' @param intervals values indicating which central prediction interval levels
#' to plot, defaults to c(.5, .8, .95). NULL means only plotting point forecasts.
#' If not provided, it will default to all available intervals in forecast data.
#' @param horizon forecasts are plotted for the horizon time steps after the
#' forecast date. Default to all available horizons in forecast data.
#' @param truth_source character specifying where the truth data will
#' be loaded from if truth_data is not provided.
#' Otherwise, this character specifies the data source to plot.
#' Currently support "JHU","USAFacts" and "NYTimes".
#' @param truth_as_of the plot includes the truth data that would have been
#' in real time as of the truth_as_of date (not using this parameter when truth data
#' is from github repo)
#' @param title optional text for the title of the plot. if left as "default",
#' the title will be automatically generated. if "none", no title will be plotted.
#' @param show.caption logical, if TRUE, caption will be included showing data sources
#’ @param plot logical; if TRUE, the plot is printed
#'
#' @return invisible ggplot object
#'
#' @export
plot_forecasts <- function(forecast_data,
truth_data = NULL,
models,
target_variable,
locations,
forecast_dates,
facet,
intervals = c(.5, .8, .95),
horizon,
truth_source = "JHU",
truth_as_of = NULL,
title = "default",
show.caption = TRUE,
plot = TRUE) | priority | create facet ready plot forecasts function specs via google doc design param forecast data data frame with truth and forecasts from load forecasts param truth data optional data frame with forecasts in the format returned by load truth param models model abbr specifying model to plot optional if there is only one model available in forecast data param target variable string specifying target type it should be one of cum death inc case inc death param locations string for fips code or us optional if there is only one location available in forecast data param facet interpretable facet option for ggplot ’ param facet scales argument for scales in facet wrap ’ param forecast dates forecast dates to plot defaults to forecast dates present ’ in the data param intervals values indicating which central prediction interval levels to plot defaults to c null means only plotting point forecasts if not provided it will default to all available intervals in forecast data param horizon forecasts are plotted for the horizon time steps after the forecast date default to all available horizons in forecast data param truth source character specifying where the truth data will be loaded from if truth data is not provided otherwise this character specifies the data source to plot currently support jhu usafacts and nytimes param truth as of the plot includes the truth data that would have been in real time as of the truth as of date not using this parameter when truth data is from github repo param title optional text for the title of the plot if left as default the title will be automatically generated if none no title will be plotted param show caption logical if true caption will be included showing data sources ’ param plot logical if true the plot is printed return invisible ggplot object export plot forecasts function forecast data truth data null models target variable locations forecast dates facet intervals c horizon truth source jhu truth as of null title default show caption true plot true | 1 |
613,561 | 19,093,797,039 | IssuesEvent | 2021-11-29 14:47:39 | yukiHaga/regex-hunting | https://api.github.com/repos/yukiHaga/regex-hunting | closed | seed_fuを使ってDBにデータを追加する。 | Priority: high | ## 概要
seed_fuを使ってDBにデータを追加します。
## やること
- [x] `gem 'seed-fu'`をGemfileに記入してインストールする。
- [x] `mkdir -p db/fixtures/{development, production}`を実行して、ディレクトリを作成する。
- [x] `touch db/fixtures/development/ファイル名.rb`で開発DB用のシードファイルを作成する。
- [x] テーブルの数だけシードファイルを作成する。
- [x] `rails db:seed_fu`を実行して、DBにデータを追加する。
## 受け入れ条件
- [x] seedデータがちゃんとDBに存在する。
- [x] `rails db:seed_fu`を2回実行しても、重複したデータがDBに保存されない。
## 参考記事
- [【Rails】seed_fuまとめ](https://qiita.com/d0ne1s/items/11d4e51fa11e65653e12)
- [railsで初期データを入れる(seed-fuの使い方)](https://qiita.com/ko2ic/items/be96e450a33d631e0059) | 1.0 | seed_fuを使ってDBにデータを追加する。 - ## 概要
seed_fuを使ってDBにデータを追加します。
## やること
- [x] `gem 'seed-fu'`をGemfileに記入してインストールする。
- [x] `mkdir -p db/fixtures/{development, production}`を実行して、ディレクトリを作成する。
- [x] `touch db/fixtures/development/ファイル名.rb`で開発DB用のシードファイルを作成する。
- [x] テーブルの数だけシードファイルを作成する。
- [x] `rails db:seed_fu`を実行して、DBにデータを追加する。
## 受け入れ条件
- [x] seedデータがちゃんとDBに存在する。
- [x] `rails db:seed_fu`を2回実行しても、重複したデータがDBに保存されない。
## 参考記事
- [【Rails】seed_fuまとめ](https://qiita.com/d0ne1s/items/11d4e51fa11e65653e12)
- [railsで初期データを入れる(seed-fuの使い方)](https://qiita.com/ko2ic/items/be96e450a33d631e0059) | priority | seed fuを使ってdbにデータを追加する。 概要 seed fuを使ってdbにデータを追加します。 やること gem seed fu をgemfileに記入してインストールする。 mkdir p db fixtures development production を実行して、ディレクトリを作成する。 touch db fixtures development ファイル名 rb で開発db用のシードファイルを作成する。 テーブルの数だけシードファイルを作成する。 rails db seed fu を実行して、dbにデータを追加する。 受け入れ条件 seedデータがちゃんとdbに存在する。 rails db seed fu 、重複したデータがdbに保存されない。 参考記事 | 1 |
303,306 | 9,305,140,688 | IssuesEvent | 2019-03-25 05:02:32 | richmondrcmp/mobileapp | https://api.github.com/repos/richmondrcmp/mobileapp | closed | Weather high and lows reversed with negative temperatures | Issue Type - Bug Priority - High | If the highs and/or lows have a negative value the high and lows are reversed (i.e., the low (-4oC) is the high and the high (1oC) is the low):

and there seems to be a disconnect between the current, highs and lows (and maybe the weather icon) between the main menu and the weather page:

| 1.0 | Weather high and lows reversed with negative temperatures - If the highs and/or lows have a negative value the high and lows are reversed (i.e., the low (-4oC) is the high and the high (1oC) is the low):

and there seems to be a disconnect between the current, highs and lows (and maybe the weather icon) between the main menu and the weather page:

| priority | weather high and lows reversed with negative temperatures if the highs and or lows have a negative value the high and lows are reversed i e the low is the high and the high is the low and there seems to be a disconnect between the current highs and lows and maybe the weather icon between the main menu and the weather page | 1 |
244,397 | 7,874,877,070 | IssuesEvent | 2018-06-25 18:30:24 | socialappslab/appcivist-platform | https://api.github.com/repos/socialappslab/appcivist-platform | closed | Logout from profile page | Priority: High duplicate | Logging out from the profile page goes to a non-LDAP window because you are at a page that has no information about any particular assembly or campaign and so AppCivist has no way of knowing which specific sign-in to show.
The problem is that logout redirects to the same page, and when it redirects to the profile page without a logged in user, you have all the data empty.
Change logout from profile page to redirect to the last page visited.
| 1.0 | Logout from profile page - Logging out from the profile page goes to a non-LDAP window because you are at a page that has no information about any particular assembly or campaign and so AppCivist has no way of knowing which specific sign-in to show.
The problem is that logout redirects to the same page, and when it redirects to the profile page without a logged in user, you have all the data empty.
Change logout from profile page to redirect to the last page visited.
| priority | logout from profile page logging out from the profile page goes to a non ldap window because you are at a page that has no information about any particular assembly or campaign and so appcivist has no way of knowing which specific sign in to show the problem is that logout redirects to the same page and when it redirects to the profile page without a logged in user you have all the data empty change logout from profile page to redirect to the last page visited | 1 |
206,758 | 7,121,060,314 | IssuesEvent | 2018-01-19 05:23:12 | wso2/product-iots | https://api.github.com/repos/wso2/product-iots | closed | [iOS] Device vitals are not getting updated | Component/IOS Priority/High Type/Bug | **Description:**
iOS device vitals are not getting updated after the initial enrolment. At the time of the enrolment, device publishes its initial vitals and then after, it will not get changed.
**Affected Product Version:**
IoT-3.1.0
**OS, DB, other environment details and versions:**
Mac OSX 10.13.2
MySQL 5.7
**Steps to reproduce:**
Enrol an iOS device and wait for few minutes until vital are changed from the device.
Then check the device details page. | 1.0 | [iOS] Device vitals are not getting updated - **Description:**
iOS device vitals are not getting updated after the initial enrolment. At the time of the enrolment, device publishes its initial vitals and then after, it will not get changed.
**Affected Product Version:**
IoT-3.1.0
**OS, DB, other environment details and versions:**
Mac OSX 10.13.2
MySQL 5.7
**Steps to reproduce:**
Enrol an iOS device and wait for few minutes until vital are changed from the device.
Then check the device details page. | priority | device vitals are not getting updated description ios device vitals are not getting updated after the initial enrolment at the time of the enrolment device publishes its initial vitals and then after it will not get changed affected product version iot os db other environment details and versions mac osx mysql steps to reproduce enrol an ios device and wait for few minutes until vital are changed from the device then check the device details page | 1 |
651,192 | 21,468,973,302 | IssuesEvent | 2022-04-26 07:46:37 | labsquare/cutevariant | https://api.github.com/repos/labsquare/cutevariant | closed | [Validation] Not refresh when new project/DB | bug high-priority CRITICAL | Please edit this template:
**Describe the bug**
When a new project/DB is created or loaded, list of samples in Validation widget is not refreshed.
**To Reproduce**
Steps to reproduce the behavior, for example:
1. Create/Open a project/DB
2. Select Samples in Validation widget
3. Create/Open a NEW project/DB
4. See that sample list is still the same
**Expected behavior**
List of samples cleared in Validation widget when create/open a new project/DB.
**Screenshots**

**Desktop (please complete the following information):**
- OS: MacOs
- Version of Cutevariant: devel
| 1.0 | [Validation] Not refresh when new project/DB - Please edit this template:
**Describe the bug**
When a new project/DB is created or loaded, list of samples in Validation widget is not refreshed.
**To Reproduce**
Steps to reproduce the behavior, for example:
1. Create/Open a project/DB
2. Select Samples in Validation widget
3. Create/Open a NEW project/DB
4. See that sample list is still the same
**Expected behavior**
List of samples cleared in Validation widget when create/open a new project/DB.
**Screenshots**

**Desktop (please complete the following information):**
- OS: MacOs
- Version of Cutevariant: devel
| priority | not refresh when new project db please edit this template describe the bug when a new project db is created or loaded list of samples in validation widget is not refreshed to reproduce steps to reproduce the behavior for example create open a project db select samples in validation widget create open a new project db see that sample list is still the same expected behavior list of samples cleared in validation widget when create open a new project db screenshots desktop please complete the following information os macos version of cutevariant devel | 1 |
611,080 | 18,944,377,345 | IssuesEvent | 2021-11-18 08:33:44 | finos/legend | https://api.github.com/repos/finos/legend | opened | Legend Studio fails with 401 (Unauthorized) | bug high priority | [Opening issue on behalf of Shane Morris (RBC)]
Hi! We are attempting to install Legend Studio locally so that we can begin to contribute some code. We were thinking that a good starting point would be to try some of the good first issues as recommended in the documentation. When we were attempting to install using the yarn commands, we ran into some issues. We have SDLC, Engine, Gitlab, and mongodb running in a Linux box and were planning on using this for development. We wanted to run yarn on our local machine and connect to the Linux box for the Engine and SDLC backbends. To do this we change the config.json file located at `legend-studio/packages/legend-studio-deployment/dev/config.json` to the following:
```
{
"appName": "studio",
"env": "local",
"sdlc": {
"url": "http://<linux-box-ip>:7070/api"
},
"engine": {
"url": "http://<linux-box-ip>:6060/api"
},
"depot": {
"url": "http://<linux-box-ip>:9090/depot/api"
},
"documentation": {
"url": "https://legend.finos.org"
}
}
```
After doing this we received the following error: (I have blocked out the Linux box IP in the console logs)

We were wondering if this type of setup is possible, and why this error might be occurring.
We understand that this issue is known to be caused by SDLC and Engine not being authenticated, but we already attempted to:
1. Grab the authorization tokens from SDLC and Engine
2. Hard refresh Chrome window and delete cookies
3. Use Chrome Incognito
4. Check the api/currentUser endpoint on Engine (and it returns user data as expected)
Before doing more experiments (ie, running SDLC and Engine locally, and see if that solves the issue), could someone from the Legend team confirm that it is possible to run a Legend Studio instance in a local environment, but having SDLC and Engine on a remote server? Was this configuration ever tested?
Thanks in advance! | 1.0 | Legend Studio fails with 401 (Unauthorized) - [Opening issue on behalf of Shane Morris (RBC)]
Hi! We are attempting to install Legend Studio locally so that we can begin to contribute some code. We were thinking that a good starting point would be to try some of the good first issues as recommended in the documentation. When we were attempting to install using the yarn commands, we ran into some issues. We have SDLC, Engine, Gitlab, and mongodb running in a Linux box and were planning on using this for development. We wanted to run yarn on our local machine and connect to the Linux box for the Engine and SDLC backbends. To do this we change the config.json file located at `legend-studio/packages/legend-studio-deployment/dev/config.json` to the following:
```
{
"appName": "studio",
"env": "local",
"sdlc": {
"url": "http://<linux-box-ip>:7070/api"
},
"engine": {
"url": "http://<linux-box-ip>:6060/api"
},
"depot": {
"url": "http://<linux-box-ip>:9090/depot/api"
},
"documentation": {
"url": "https://legend.finos.org"
}
}
```
After doing this we received the following error: (I have blocked out the Linux box IP in the console logs)

We were wondering if this type of setup is possible, and why this error might be occurring.
We understand that this issue is known to be caused by SDLC and Engine not being authenticated, but we already attempted to:
1. Grab the authorization tokens from SDLC and Engine
2. Hard refresh Chrome window and delete cookies
3. Use Chrome Incognito
4. Check the api/currentUser endpoint on Engine (and it returns user data as expected)
Before doing more experiments (ie, running SDLC and Engine locally, and see if that solves the issue), could someone from the Legend team confirm that it is possible to run a Legend Studio instance in a local environment, but having SDLC and Engine on a remote server? Was this configuration ever tested?
Thanks in advance! | priority | legend studio fails with unauthorized hi we are attempting to install legend studio locally so that we can begin to contribute some code we were thinking that a good starting point would be to try some of the good first issues as recommended in the documentation when we were attempting to install using the yarn commands we ran into some issues we have sdlc engine gitlab and mongodb running in a linux box and were planning on using this for development we wanted to run yarn on our local machine and connect to the linux box for the engine and sdlc backbends to do this we change the config json file located at legend studio packages legend studio deployment dev config json to the following appname studio env local sdlc url engine url depot url documentation url after doing this we received the following error i have blocked out the linux box ip in the console logs we were wondering if this type of setup is possible and why this error might be occurring we understand that this issue is known to be caused by sdlc and engine not being authenticated but we already attempted to grab the authorization tokens from sdlc and engine hard refresh chrome window and delete cookies use chrome incognito check the api currentuser endpoint on engine and it returns user data as expected before doing more experiments ie running sdlc and engine locally and see if that solves the issue could someone from the legend team confirm that it is possible to run a legend studio instance in a local environment but having sdlc and engine on a remote server was this configuration ever tested thanks in advance | 1 |
57,857 | 3,086,120,241 | IssuesEvent | 2015-08-25 00:08:51 | google/paco | https://api.github.com/repos/google/paco | closed | web ui: Participation Stats for Joined should append who query param | Component-Server Component-UI Priority-High | when clicking on 'Stats' under the joined list it should return just that user's data by appending the "who" query param. For example,
/participantStats?experimentId=1234&who=user1@gmail.com | 1.0 | web ui: Participation Stats for Joined should append who query param - when clicking on 'Stats' under the joined list it should return just that user's data by appending the "who" query param. For example,
/participantStats?experimentId=1234&who=user1@gmail.com | priority | web ui participation stats for joined should append who query param when clicking on stats under the joined list it should return just that user s data by appending the who query param for example participantstats experimentid who gmail com | 1 |
783,159 | 27,520,739,603 | IssuesEvent | 2023-03-06 14:53:41 | AY2223S2-CS2103-W16-3/tp | https://api.github.com/repos/AY2223S2-CS2103-W16-3/tp | closed | As an event planner, I can view all my events | type.Story priority.high | ... so that I can track all the upcoming events I have. | 1.0 | As an event planner, I can view all my events - ... so that I can track all the upcoming events I have. | priority | as an event planner i can view all my events so that i can track all the upcoming events i have | 1 |
138,655 | 5,345,255,790 | IssuesEvent | 2017-02-17 16:31:13 | CoderDojo/community-platform | https://api.github.com/repos/CoderDojo/community-platform | closed | Add timezones to Events | bug events high priority top priority | All event times are being inputted in GMT by default and there are no options to change this.
This is problematic as the event gets removed at the date & GMT time on the event rather than local time, particularly in American timezones which are behind GMT meaning the event gets removed too early.
The quick fix for this would be for events to only remove +10 hours after the event but in the long term we will want a drop down or ability to add custom timezones by the user when creating the event.
| 2.0 | Add timezones to Events - All event times are being inputted in GMT by default and there are no options to change this.
This is problematic as the event gets removed at the date & GMT time on the event rather than local time, particularly in American timezones which are behind GMT meaning the event gets removed too early.
The quick fix for this would be for events to only remove +10 hours after the event but in the long term we will want a drop down or ability to add custom timezones by the user when creating the event.
| priority | add timezones to events all event times are being inputted in gmt by default and there are no options to change this this is problematic as the event gets removed at the date gmt time on the event rather than local time particularly in american timezones which are behind gmt meaning the event gets removed too early the quick fix for this would be for events to only remove hours after the event but in the long term we will want a drop down or ability to add custom timezones by the user when creating the event | 1 |
59,468 | 3,113,786,047 | IssuesEvent | 2015-09-03 02:25:56 | onyxfish/agate | https://api.github.com/repos/onyxfish/agate | closed | In tutorial Formula generates DecimalColumn even though text_type is specified? | bug priority-high | See traceback on #201 | 1.0 | In tutorial Formula generates DecimalColumn even though text_type is specified? - See traceback on #201 | priority | in tutorial formula generates decimalcolumn even though text type is specified see traceback on | 1 |
88,619 | 3,783,250,645 | IssuesEvent | 2016-03-19 01:30:01 | durhamatletico/durhamatletico-cms | https://api.github.com/repos/durhamatletico/durhamatletico-cms | closed | registration instructions in Spanish | content high priority | This text should appear early in the registration instructions. It says, "if you want to register an entire team, the cost is $600. Call or send a text to xxx... for more information."
The number is Alfonso's Google Voice number.
Si quieren registrar al equipo entero, el coste es de $600. Llame o escribe un texto al 919-937-2109 para mas información"
| 1.0 | registration instructions in Spanish - This text should appear early in the registration instructions. It says, "if you want to register an entire team, the cost is $600. Call or send a text to xxx... for more information."
The number is Alfonso's Google Voice number.
Si quieren registrar al equipo entero, el coste es de $600. Llame o escribe un texto al 919-937-2109 para mas información"
| priority | registration instructions in spanish this text should appear early in the registration instructions it says if you want to register an entire team the cost is call or send a text to xxx for more information the number is alfonso s google voice number si quieren registrar al equipo entero el coste es de llame o escribe un texto al para mas información | 1 |
165,588 | 6,278,762,520 | IssuesEvent | 2017-07-18 14:58:05 | qutebrowser/qutebrowser | https://api.github.com/repos/qutebrowser/qutebrowser | closed | FileNotFoundError in ipc.update_atime | bug needs triage priority: 0 - high | I can reproduce this somewhat reliably by running the tests in a loop on the Debian Jessie and Ubuntu Trusty buildbots:
```
============================= ERRORS =============================
________ ERROR at teardown of TestListen.test_atime_update _______
Qt exceptions in virtual methods:
________________________________________________________________________________
File "/home/buildbotx/qutebrowser/qutebrowser/misc/ipc.py", line 368, in update_atime
os.utime(path)
FileNotFoundError: [Errno 2] No such file or directory
________________________________________________________________________________
============================ FAILURES ============================
__________________ TestListen.test_atime_update __________________
self = <test_ipc.TestListen object at 0x7f08de08cac8>, qtbot = <pytestqt.qtbot.QtBot object at 0x7f08ddf01a90>, ipc_server = <qutebrowser.misc.ipc.IPCServer object at 0x7f08dec18288>
@pytest.mark.posix
def test_atime_update(self, qtbot, ipc_server):
ipc_server._atime_timer.setInterval(500) # We don't want to wait 6h
ipc_server.listen()
old_atime = os.stat(ipc_server._server.fullServerName()).st_atime_ns
with qtbot.waitSignal(ipc_server._atime_timer.timeout, timeout=2000,
raising=True):
pass
# Make sure the timer is not singleShot
with qtbot.waitSignal(ipc_server._atime_timer.timeout, timeout=2000,
raising=True):
pass
> new_atime = os.stat(ipc_server._server.fullServerName()).st_atime_ns
E FileNotFoundError: [Errno 2] No such file or directory: '/tmp/qute-test'
tests/unit/misc/test_ipc.py:357: FileNotFoundError
-------------------------- Captured log --------------------------
ipc.py 193 DEBUG Not calling setSocketOptions
ipc.py 204 DEBUG Listening as qute-test
ipc.py 367 DEBUG Touching /tmp/qute-test
ipc.py 367 DEBUG Touching /tmp/qute-test
```
| 1.0 | FileNotFoundError in ipc.update_atime - I can reproduce this somewhat reliably by running the tests in a loop on the Debian Jessie and Ubuntu Trusty buildbots:
```
============================= ERRORS =============================
________ ERROR at teardown of TestListen.test_atime_update _______
Qt exceptions in virtual methods:
________________________________________________________________________________
File "/home/buildbotx/qutebrowser/qutebrowser/misc/ipc.py", line 368, in update_atime
os.utime(path)
FileNotFoundError: [Errno 2] No such file or directory
________________________________________________________________________________
============================ FAILURES ============================
__________________ TestListen.test_atime_update __________________
self = <test_ipc.TestListen object at 0x7f08de08cac8>, qtbot = <pytestqt.qtbot.QtBot object at 0x7f08ddf01a90>, ipc_server = <qutebrowser.misc.ipc.IPCServer object at 0x7f08dec18288>
@pytest.mark.posix
def test_atime_update(self, qtbot, ipc_server):
ipc_server._atime_timer.setInterval(500) # We don't want to wait 6h
ipc_server.listen()
old_atime = os.stat(ipc_server._server.fullServerName()).st_atime_ns
with qtbot.waitSignal(ipc_server._atime_timer.timeout, timeout=2000,
raising=True):
pass
# Make sure the timer is not singleShot
with qtbot.waitSignal(ipc_server._atime_timer.timeout, timeout=2000,
raising=True):
pass
> new_atime = os.stat(ipc_server._server.fullServerName()).st_atime_ns
E FileNotFoundError: [Errno 2] No such file or directory: '/tmp/qute-test'
tests/unit/misc/test_ipc.py:357: FileNotFoundError
-------------------------- Captured log --------------------------
ipc.py 193 DEBUG Not calling setSocketOptions
ipc.py 204 DEBUG Listening as qute-test
ipc.py 367 DEBUG Touching /tmp/qute-test
ipc.py 367 DEBUG Touching /tmp/qute-test
```
| priority | filenotfounderror in ipc update atime i can reproduce this somewhat reliably by running the tests in a loop on the debian jessie and ubuntu trusty buildbots errors error at teardown of testlisten test atime update qt exceptions in virtual methods file home buildbotx qutebrowser qutebrowser misc ipc py line in update atime os utime path filenotfounderror no such file or directory failures testlisten test atime update self qtbot ipc server pytest mark posix def test atime update self qtbot ipc server ipc server atime timer setinterval we don t want to wait ipc server listen old atime os stat ipc server server fullservername st atime ns with qtbot waitsignal ipc server atime timer timeout timeout raising true pass make sure the timer is not singleshot with qtbot waitsignal ipc server atime timer timeout timeout raising true pass new atime os stat ipc server server fullservername st atime ns e filenotfounderror no such file or directory tmp qute test tests unit misc test ipc py filenotfounderror captured log ipc py debug not calling setsocketoptions ipc py debug listening as qute test ipc py debug touching tmp qute test ipc py debug touching tmp qute test | 1 |
626,973 | 19,848,795,619 | IssuesEvent | 2022-01-21 09:56:11 | TeamSparker/Spark-iOS | https://api.github.com/repos/TeamSparker/Spark-iOS | opened | [FEAT] 스파크 보내기 토스트메세지 | Feat 🦹t없e맑은水빈 P1 / Priority High | ## 📌 Issue
<!-- 이슈에 대해 간략하게 설명해주세요 -->
스파크 보내기 토스트메세지 만들겠습니다
## 📝 To-do
<!-- 진행할 작업에 대해 적어주세요 -->
- [ ] 토스트메세지 extension 빼기
- [ ] 토스트메세지 띄우기
| 1.0 | [FEAT] 스파크 보내기 토스트메세지 - ## 📌 Issue
<!-- 이슈에 대해 간략하게 설명해주세요 -->
스파크 보내기 토스트메세지 만들겠습니다
## 📝 To-do
<!-- 진행할 작업에 대해 적어주세요 -->
- [ ] 토스트메세지 extension 빼기
- [ ] 토스트메세지 띄우기
| priority | 스파크 보내기 토스트메세지 📌 issue 스파크 보내기 토스트메세지 만들겠습니다 📝 to do 토스트메세지 extension 빼기 토스트메세지 띄우기 | 1 |
449,781 | 12,974,972,083 | IssuesEvent | 2020-07-21 16:13:50 | MatthewSpofford/Multiscale-Statistical-Analysis | https://api.github.com/repos/MatthewSpofford/Multiscale-Statistical-Analysis | opened | PlotData open_file2 does not correctly handle scale discrepencies | bug priority high | If there are scales that may not be contained within all of the opened datasets, then these scales will continue to be appended to the `results_scale` list. Since the order of the scales is based on the files being imported, the `results_scale` list is completely out of order. The `open_file2` function then continues to append on to the `relative_area` and `complexity` lists, causing there to be more scale values than actual area/complexity values, completely disrupting the data table.
Blocking: #86 | 1.0 | PlotData open_file2 does not correctly handle scale discrepencies - If there are scales that may not be contained within all of the opened datasets, then these scales will continue to be appended to the `results_scale` list. Since the order of the scales is based on the files being imported, the `results_scale` list is completely out of order. The `open_file2` function then continues to append on to the `relative_area` and `complexity` lists, causing there to be more scale values than actual area/complexity values, completely disrupting the data table.
Blocking: #86 | priority | plotdata open does not correctly handle scale discrepencies if there are scales that may not be contained within all of the opened datasets then these scales will continue to be appended to the results scale list since the order of the scales is based on the files being imported the results scale list is completely out of order the open function then continues to append on to the relative area and complexity lists causing there to be more scale values than actual area complexity values completely disrupting the data table blocking | 1 |
220,987 | 7,372,672,110 | IssuesEvent | 2018-03-13 15:18:39 | springboilerplate/spring-rest-security-boilerplate | https://api.github.com/repos/springboilerplate/spring-rest-security-boilerplate | closed | RabbitMQ integration for email processes | High Priority enhancement | Currently, We send emails as normal Java methods, which is bad way. To understand why it is bad and what to do, you can read following links.
In this task, we basically use RabbitMQ for our email sending services.
What is RabbitMQ and why do we need it?
- https://www.cloudamqp.com/blog/2015-05-18-part1-rabbitmq-for-beginners-what-is-rabbitmq.html
How to make Spring and RabbitMQ work together?
- http://www.bahadirakin.com/amqp-ve-rabbitmq/
- http://www.bahadirakin.com/spring-amqp-ve-rabbitmq/
Official Tutorial
- https://www.rabbitmq.com/getstarted.html | 1.0 | RabbitMQ integration for email processes - Currently, We send emails as normal Java methods, which is bad way. To understand why it is bad and what to do, you can read following links.
In this task, we basically use RabbitMQ for our email sending services.
What is RabbitMQ and why do we need it?
- https://www.cloudamqp.com/blog/2015-05-18-part1-rabbitmq-for-beginners-what-is-rabbitmq.html
How to make Spring and RabbitMQ work together?
- http://www.bahadirakin.com/amqp-ve-rabbitmq/
- http://www.bahadirakin.com/spring-amqp-ve-rabbitmq/
Official Tutorial
- https://www.rabbitmq.com/getstarted.html | priority | rabbitmq integration for email processes currently we send emails as normal java methods which is bad way to understand why it is bad and what to do you can read following links in this task we basically use rabbitmq for our email sending services what is rabbitmq and why do we need it how to make spring and rabbitmq work together official tutorial | 1 |
472,290 | 13,622,330,550 | IssuesEvent | 2020-09-24 03:18:34 | SE761Team4/jabref | https://api.github.com/repos/SE761Team4/jabref | closed | [4pt] Make BibTeX to MindMap parser | backend blocker high priority | **User Story**
<!--As a [user role], I'd like to [do something], so that [some goal].-->
As a developer, I would like to map BibTex to JSON so that I can parse information to JabMap
**Acceptance Criteria**
- BibTeX data can be parsed via HTTP to JabMap
**Notes**
---
**Why is this feature needed? Please describe the problem your requested feature wants to solve**
<!-- Describe what the problem is. Ex. I'm always frustrated when ... -->
**Describe the solution you'd like**
<!--Describe what you want to happen -->
| 1.0 | [4pt] Make BibTeX to MindMap parser - **User Story**
<!--As a [user role], I'd like to [do something], so that [some goal].-->
As a developer, I would like to map BibTex to JSON so that I can parse information to JabMap
**Acceptance Criteria**
- BibTeX data can be parsed via HTTP to JabMap
**Notes**
---
**Why is this feature needed? Please describe the problem your requested feature wants to solve**
<!-- Describe what the problem is. Ex. I'm always frustrated when ... -->
**Describe the solution you'd like**
<!--Describe what you want to happen -->
| priority | make bibtex to mindmap parser user story as a developer i would like to map bibtex to json so that i can parse information to jabmap acceptance criteria bibtex data can be parsed via http to jabmap notes why is this feature needed please describe the problem your requested feature wants to solve describe the solution you d like | 1 |
171,803 | 6,494,608,673 | IssuesEvent | 2017-08-21 22:48:57 | cranndarach/lifetracker | https://api.github.com/repos/cranndarach/lifetracker | closed | [docs] forms.cson needs documenting | priority/high task-size/medium | There is a wiki page for making and editing forms, but it is empty. As soon as possible, it should be explained in plain language so that LifeTracker is still accessible to non-programmers. | 1.0 | [docs] forms.cson needs documenting - There is a wiki page for making and editing forms, but it is empty. As soon as possible, it should be explained in plain language so that LifeTracker is still accessible to non-programmers. | priority | forms cson needs documenting there is a wiki page for making and editing forms but it is empty as soon as possible it should be explained in plain language so that lifetracker is still accessible to non programmers | 1 |
86,427 | 3,715,118,741 | IssuesEvent | 2016-03-03 00:02:35 | UCHIC/CKANDev | https://api.github.com/repos/UCHIC/CKANDev | closed | Photo Upload | HIGH PRIORITY question | Are there ways to upload a picture other than the Gravatar website? Users may want to add photos without using that website functionality. | 1.0 | Photo Upload - Are there ways to upload a picture other than the Gravatar website? Users may want to add photos without using that website functionality. | priority | photo upload are there ways to upload a picture other than the gravatar website users may want to add photos without using that website functionality | 1 |
535,893 | 15,700,641,130 | IssuesEvent | 2021-03-26 10:07:13 | Systems-Learning-and-Development-Lab/MMM | https://api.github.com/repos/Systems-Learning-and-Development-Lab/MMM | opened | Remove mark not always works | priority-high | . המחיקה לא תמיד עובדת, אני לא ממש מצליחה לשחזר בדיוק. הפעולות שאני עושה הן לשים סמן, להריץ, למחוק, לשנות סמן, להריץ, למחוק את הסמן. ברוב המקרים המחיקה לא עובדת.
@Ron-Teller | 1.0 | Remove mark not always works - . המחיקה לא תמיד עובדת, אני לא ממש מצליחה לשחזר בדיוק. הפעולות שאני עושה הן לשים סמן, להריץ, למחוק, לשנות סמן, להריץ, למחוק את הסמן. ברוב המקרים המחיקה לא עובדת.
@Ron-Teller | priority | remove mark not always works המחיקה לא תמיד עובדת אני לא ממש מצליחה לשחזר בדיוק הפעולות שאני עושה הן לשים סמן להריץ למחוק לשנות סמן להריץ למחוק את הסמן ברוב המקרים המחיקה לא עובדת ron teller | 1 |
558,834 | 16,543,398,181 | IssuesEvent | 2021-05-27 19:59:30 | ampproject/amphtml | https://api.github.com/repos/ampproject/amphtml | opened | AMPHTML ads using gwd-animation are broken | Component: AMPHTML ads P1: High Priority Type: Bug | ### Description
Regression introduced with the launch of no-signing rendering path.
### Reproduction Steps
Example page that will render as a blank ad: https://gist.github.com/calebcordry/8194102522f2b50821ea48d70bc2583e
### Relevant Logs
_No response_
### Browser(s) Affected
Chrome, Firefox, Safari, Edge, UC Browser
### OS(s) Affected
All
### Device(s) Affected
All
### AMP Version Affected
2104170104000 and after | 1.0 | AMPHTML ads using gwd-animation are broken - ### Description
Regression introduced with the launch of no-signing rendering path.
### Reproduction Steps
Example page that will render as a blank ad: https://gist.github.com/calebcordry/8194102522f2b50821ea48d70bc2583e
### Relevant Logs
_No response_
### Browser(s) Affected
Chrome, Firefox, Safari, Edge, UC Browser
### OS(s) Affected
All
### Device(s) Affected
All
### AMP Version Affected
2104170104000 and after | priority | amphtml ads using gwd animation are broken description regression introduced with the launch of no signing rendering path reproduction steps example page that will render as a blank ad relevant logs no response browser s affected chrome firefox safari edge uc browser os s affected all device s affected all amp version affected and after | 1 |
675,854 | 23,109,762,005 | IssuesEvent | 2022-07-27 12:03:00 | fadeinside/s3air-achievements-plus | https://api.github.com/repos/fadeinside/s3air-achievements-plus | closed | Achievement Guides | Type: Feature Priority: Highest | **Description**
\-
**Screenshots/References**
\-
**Additional context**
\- | 1.0 | Achievement Guides - **Description**
\-
**Screenshots/References**
\-
**Additional context**
\- | priority | achievement guides description screenshots references additional context | 1 |
679,503 | 23,235,340,714 | IssuesEvent | 2022-08-03 11:16:13 | chloebrett/mlvet | https://api.github.com/repos/chloebrett/mlvet | closed | Browsing for a file to import and then cancelling causes a blank selected file to appear | high priority | <img width="685" alt="Screen Shot 2022-05-28 at 3 23 14 pm" src="https://user-images.githubusercontent.com/6735055/170811415-e13b0d83-2935-4afb-a77a-51ba4bd069bc.png">
The transcription button then gets enabled, but clicking it doesn't do anything | 1.0 | Browsing for a file to import and then cancelling causes a blank selected file to appear - <img width="685" alt="Screen Shot 2022-05-28 at 3 23 14 pm" src="https://user-images.githubusercontent.com/6735055/170811415-e13b0d83-2935-4afb-a77a-51ba4bd069bc.png">
The transcription button then gets enabled, but clicking it doesn't do anything | priority | browsing for a file to import and then cancelling causes a blank selected file to appear img width alt screen shot at pm src the transcription button then gets enabled but clicking it doesn t do anything | 1 |
748,851 | 26,140,449,085 | IssuesEvent | 2022-12-29 17:37:26 | netdata/netdata-cloud | https://api.github.com/repos/netdata/netdata-cloud | closed | View all Active alerts in a user friendly format (table) from a mobile device. | priority/high Customer Success cloud-frontend alerts-team Q1 GOAL | ##Issue
As a user of Netdata Cloud, I would like to view all active alerts in my Space/Room/Node on my mobile device.
##Use-case
On receiving a notification on the mobile, the user needs to be able to see the specific alert and also other active alerts on a mobile device.
| 1.0 | View all Active alerts in a user friendly format (table) from a mobile device. - ##Issue
As a user of Netdata Cloud, I would like to view all active alerts in my Space/Room/Node on my mobile device.
##Use-case
On receiving a notification on the mobile, the user needs to be able to see the specific alert and also other active alerts on a mobile device.
| priority | view all active alerts in a user friendly format table from a mobile device issue as a user of netdata cloud i would like to view all active alerts in my space room node on my mobile device use case on receiving a notification on the mobile the user needs to be able to see the specific alert and also other active alerts on a mobile device | 1 |
155,819 | 5,961,329,165 | IssuesEvent | 2017-05-29 16:56:13 | WatzekDigitalInitiatives/fitbit-ror | https://api.github.com/repos/WatzekDigitalInitiatives/fitbit-ror | closed | Cannot join private teams (extreme high priority) | front end high priority | Users should see `user_teams/new.html.erb` when joining a private team they are not part of | 1.0 | Cannot join private teams (extreme high priority) - Users should see `user_teams/new.html.erb` when joining a private team they are not part of | priority | cannot join private teams extreme high priority users should see user teams new html erb when joining a private team they are not part of | 1 |
793,842 | 28,013,023,569 | IssuesEvent | 2023-03-27 20:08:49 | RadeonOpenCompute/ROCm-CompilerSupport | https://api.github.com/repos/RadeonOpenCompute/ROCm-CompilerSupport | closed | comgr lacks MT safety | enhancement priority_high | It seems like comgr is not working properly in multithreaded contexts. This prevents it from full-blown usage in MIOpen.
## How to reproduce
Prerequisites: A linux machine with gfx900/906/908 GPU card installed (Radeon VII preferred), vanilla ROCm 3.3.
- (1) Pull the https://github.com/ROCmSoftwarePlatform/MIOpen/tree/comgr-attach branch from MIOpen repo and checkout https://github.com/ROCmSoftwarePlatform/MIOpen/commit/9a3a2e7142ea7651c9e0056ed8a3a1d860a8d702
- (2) Build MIOpen library and MIOpenDriver with HIP backend. Use developer build, `cmake -DBUILD_DEV=On...` (see [here](https://github.com/ROCmSoftwarePlatform/MIOpen/blob/develop/README.md) for detailed instruction), then make `MIOpenDriver -j`. Do not install.
- (3) Optional: modify environment to skip non-relevant kernels in MIOpen. This would speed-up execution and clean MIOpen logs from non-relevant information.
```
export MIOPEN_DEBUG_CONV_GEMM=0
export MIOPEN_DEBUG_CONV_FFT=0
export MIOPEN_DEBUG_CONV_IMPLICIT_GEMM=0
export MIOPEN_DEBUG_CONV_SCGEMM=0
export MIOPEN_DEBUG_CONV_WINOGRAD=0
export MIOPEN_DEBUG_CONV_DIRECT=1
export MIOPEN_DEBUG_GCN_ASM_KERNELS=0
export MIOPEN_DEBUG_OPENCL_CONVOLUTIONS=1
```
- (4) Run the following command and make sure that it works fine:
```
$ ./bin/MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1
MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1
MIOpen Forward Conv. Algorithm: 1, Solution: 9/ConvOclDirectFwdGen
GPU Kernel Time Forward Conv. Elapsed: 0.307048 ms (average)
stats: name, n, c, ho, wo, x, y, k, flopCnt, bytesRead, bytesWritten, GFLOPs, GB/s, timeMs
stats: fwd-conv5x20u2, 8, 1, 79, 341, 5, 20, 32, 1379276800, 3619200, 27585536, 4492, 102, 0.307048
```
- (5) Optional: re-run the same command prefixed with `MIOPEN_LOG_LEVEL=5` and make sure that the following lines present on the console:
```
Info [EvaluateInvokers] ConvOclDirectFwdGen: MIOpenCDFGen4: 0.323367 < 3.40282e+38
Info [EvaluateInvokers] ConvOclDirectFwd: MIOpenConvUni: 1.67044 >= 0.323367
...
Info [FindConvFwdAlgorithm] FW Chosen Algorithm: ConvOclDirectFwdGen , 0, 0.323367
```
These upper two lines tell us that two kernels (MIOpenCDFGen4 and MIOpenConvUni) obtained from two Solutions (ConvOclDirectFwdGen and ConvOclDirectFwd) were built and run by the library. The third line says that ConvOclDirectFwdGen (the fastest) was selected. The library builds the kernels in parallel; you may wish to export `MIOPEN_ENABLE_LOGGING_MPMT=1` and `MIOPEN_LOG_LEVEL=6` and see these details in the console log.
- (6) Now it's time to use comgr. Clean build directory (you can skip cleaning the build directory, but I am not sure if this would rebuild all the necessary components). Add `-DMIOPEN_USE_COMGR=On` to cmake command line (see [here](https://github.com/ROCmSoftwarePlatform/MIOpen/pull/177#issue-408736873) for details) and rebuild library and driver,
- (7) Re-run the command from step (4). Output:
```
$ ./bin/MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1
MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1
clang (LLVM option parsing): for the --amdgpu-early-inline-all option: may only occur zero or one times!
clang (LLVM option parsing): for the --amdgpu-prelink option: may only occur zero or one times!
clang (LLVM option parsing): for the --amdgpu-internalize-symbols option: may only occur zero or one times!
clang (LLVM option parsing): for the --pgo-warn-misexpect option: may only occur zero or one times!
```
### :red_circle: There are two problems:
- Build does not succeed.
- Build kills the driver. Neither error codes returned to MIOpen, nor exceptions thrown.
You can disable MT builds in MIOpen by the following env setting:
```
export MIOPEN_COMPILE_PARALLEL_LEVEL=1
```
and see that driver works normally in this case.
| 1.0 | comgr lacks MT safety - It seems like comgr is not working properly in multithreaded contexts. This prevents it from full-blown usage in MIOpen.
## How to reproduce
Prerequisites: A linux machine with gfx900/906/908 GPU card installed (Radeon VII preferred), vanilla ROCm 3.3.
- (1) Pull the https://github.com/ROCmSoftwarePlatform/MIOpen/tree/comgr-attach branch from MIOpen repo and checkout https://github.com/ROCmSoftwarePlatform/MIOpen/commit/9a3a2e7142ea7651c9e0056ed8a3a1d860a8d702
- (2) Build MIOpen library and MIOpenDriver with HIP backend. Use developer build, `cmake -DBUILD_DEV=On...` (see [here](https://github.com/ROCmSoftwarePlatform/MIOpen/blob/develop/README.md) for detailed instruction), then make `MIOpenDriver -j`. Do not install.
- (3) Optional: modify environment to skip non-relevant kernels in MIOpen. This would speed-up execution and clean MIOpen logs from non-relevant information.
```
export MIOPEN_DEBUG_CONV_GEMM=0
export MIOPEN_DEBUG_CONV_FFT=0
export MIOPEN_DEBUG_CONV_IMPLICIT_GEMM=0
export MIOPEN_DEBUG_CONV_SCGEMM=0
export MIOPEN_DEBUG_CONV_WINOGRAD=0
export MIOPEN_DEBUG_CONV_DIRECT=1
export MIOPEN_DEBUG_GCN_ASM_KERNELS=0
export MIOPEN_DEBUG_OPENCL_CONVOLUTIONS=1
```
- (4) Run the following command and make sure that it works fine:
```
$ ./bin/MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1
MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1
MIOpen Forward Conv. Algorithm: 1, Solution: 9/ConvOclDirectFwdGen
GPU Kernel Time Forward Conv. Elapsed: 0.307048 ms (average)
stats: name, n, c, ho, wo, x, y, k, flopCnt, bytesRead, bytesWritten, GFLOPs, GB/s, timeMs
stats: fwd-conv5x20u2, 8, 1, 79, 341, 5, 20, 32, 1379276800, 3619200, 27585536, 4492, 102, 0.307048
```
- (5) Optional: re-run the same command prefixed with `MIOPEN_LOG_LEVEL=5` and make sure that the following lines present on the console:
```
Info [EvaluateInvokers] ConvOclDirectFwdGen: MIOpenCDFGen4: 0.323367 < 3.40282e+38
Info [EvaluateInvokers] ConvOclDirectFwd: MIOpenConvUni: 1.67044 >= 0.323367
...
Info [FindConvFwdAlgorithm] FW Chosen Algorithm: ConvOclDirectFwdGen , 0, 0.323367
```
These upper two lines tell us that two kernels (MIOpenCDFGen4 and MIOpenConvUni) obtained from two Solutions (ConvOclDirectFwdGen and ConvOclDirectFwd) were built and run by the library. The third line says that ConvOclDirectFwdGen (the fastest) was selected. The library builds the kernels in parallel; you may wish to export `MIOPEN_ENABLE_LOGGING_MPMT=1` and `MIOPEN_LOG_LEVEL=6` and see these details in the console log.
- (6) Now it's time to use comgr. Clean build directory (you can skip cleaning the build directory, but I am not sure if this would rebuild all the necessary components). Add `-DMIOPEN_USE_COMGR=On` to cmake command line (see [here](https://github.com/ROCmSoftwarePlatform/MIOpen/pull/177#issue-408736873) for details) and rebuild library and driver,
- (7) Re-run the command from step (4). Output:
```
$ ./bin/MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1
MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1
clang (LLVM option parsing): for the --amdgpu-early-inline-all option: may only occur zero or one times!
clang (LLVM option parsing): for the --amdgpu-prelink option: may only occur zero or one times!
clang (LLVM option parsing): for the --amdgpu-internalize-symbols option: may only occur zero or one times!
clang (LLVM option parsing): for the --pgo-warn-misexpect option: may only occur zero or one times!
```
### :red_circle: There are two problems:
- Build does not succeed.
- Build kills the driver. Neither error codes returned to MIOpen, nor exceptions thrown.
You can disable MT builds in MIOpen by the following env setting:
```
export MIOPEN_COMPILE_PARALLEL_LEVEL=1
```
and see that driver works normally in this case.
| priority | comgr lacks mt safety it seems like comgr is not working properly in multithreaded contexts this prevents it from full blown usage in miopen how to reproduce prerequisites a linux machine with gpu card installed radeon vii preferred vanilla rocm pull the branch from miopen repo and checkout build miopen library and miopendriver with hip backend use developer build cmake dbuild dev on see for detailed instruction then make miopendriver j do not install optional modify environment to skip non relevant kernels in miopen this would speed up execution and clean miopen logs from non relevant information export miopen debug conv gemm export miopen debug conv fft export miopen debug conv implicit gemm export miopen debug conv scgemm export miopen debug conv winograd export miopen debug conv direct export miopen debug gcn asm kernels export miopen debug opencl convolutions run the following command and make sure that it works fine bin miopendriver conv x y w h c n k p q u v l j g f v s t i miopendriver conv x y w h c n k p q u v l j g f v s t i miopen forward conv algorithm solution convocldirectfwdgen gpu kernel time forward conv elapsed ms average stats name n c ho wo x y k flopcnt bytesread byteswritten gflops gb s timems stats fwd optional re run the same command prefixed with miopen log level and make sure that the following lines present on the console info convocldirectfwdgen info convocldirectfwd miopenconvuni info fw chosen algorithm convocldirectfwdgen these upper two lines tell us that two kernels and miopenconvuni obtained from two solutions convocldirectfwdgen and convocldirectfwd were built and run by the library the third line says that convocldirectfwdgen the fastest was selected the library builds the kernels in parallel you may wish to export miopen enable logging mpmt and miopen log level and see these details in the console log now it s time to use comgr clean build directory you can skip cleaning the build directory but i am not sure if this would rebuild all the necessary components add dmiopen use comgr on to cmake command line see for details and rebuild library and driver re run the command from step output bin miopendriver conv x y w h c n k p q u v l j g f v s t i miopendriver conv x y w h c n k p q u v l j g f v s t i clang llvm option parsing for the amdgpu early inline all option may only occur zero or one times clang llvm option parsing for the amdgpu prelink option may only occur zero or one times clang llvm option parsing for the amdgpu internalize symbols option may only occur zero or one times clang llvm option parsing for the pgo warn misexpect option may only occur zero or one times red circle there are two problems build does not succeed build kills the driver neither error codes returned to miopen nor exceptions thrown you can disable mt builds in miopen by the following env setting export miopen compile parallel level and see that driver works normally in this case | 1 |
755,492 | 26,430,515,971 | IssuesEvent | 2023-01-14 19:04:00 | banco-alimentar/alimentestaideia.pt | https://api.github.com/repos/banco-alimentar/alimentestaideia.pt | opened | Management of donors (outside the site) | High.Priority | The food bank manages all donations that are not made in the site, using a “old” access database.
They register the information about all the donors, and each donation. Then produce a invoice (a word doc) that is now not legal anymore.
I thought that we have “everything” to replace this access with our site.
For that we would need to be able to have a backoffice to enter the donor information, and add donations. One donor might have many donations, so we don’t need to reenter the donor info everytime we add a donation.”
2 times per year we need to export a list of donors, so we can send them a letter.
If we do move the solution, I would need to also bring the past information to the site to provide continuity.

For Each donation:

### Fields:
- Donor
- DonationDate
- DonationAmount
- TypeOfDonation
- Aleatorio
- Espontanio
- Personalizado
- Regular
- Tribunal
- Um amigo
- Site
### PaymentMethods
- Cheque
- Dep. Conta
- Multibanco
- Numerário
- Vale Postal
- MbWay
- Site (the current and existing way!?)
- Paypal
- CreditCard
This feature should be accessed in the Backoffice or it could be implemented in the Backoffice Powerapp. | 1.0 | Management of donors (outside the site) - The food bank manages all donations that are not made in the site, using a “old” access database.
They register the information about all the donors, and each donation. Then produce a invoice (a word doc) that is now not legal anymore.
I thought that we have “everything” to replace this access with our site.
For that we would need to be able to have a backoffice to enter the donor information, and add donations. One donor might have many donations, so we don’t need to reenter the donor info everytime we add a donation.”
2 times per year we need to export a list of donors, so we can send them a letter.
If we do move the solution, I would need to also bring the past information to the site to provide continuity.

For Each donation:

### Fields:
- Donor
- DonationDate
- DonationAmount
- TypeOfDonation
- Aleatorio
- Espontanio
- Personalizado
- Regular
- Tribunal
- Um amigo
- Site
### PaymentMethods
- Cheque
- Dep. Conta
- Multibanco
- Numerário
- Vale Postal
- MbWay
- Site (the current and existing way!?)
- Paypal
- CreditCard
This feature should be accessed in the Backoffice or it could be implemented in the Backoffice Powerapp. | priority | management of donors outside the site the food bank manages all donations that are not made in the site using a “old” access database they register the information about all the donors and each donation then produce a invoice a word doc that is now not legal anymore i thought that we have “everything” to replace this access with our site for that we would need to be able to have a backoffice to enter the donor information and add donations one donor might have many donations so we don’t need to reenter the donor info everytime we add a donation ” times per year we need to export a list of donors so we can send them a letter if we do move the solution i would need to also bring the past information to the site to provide continuity for each donation fields donor donationdate donationamount typeofdonation aleatorio espontanio personalizado regular tribunal um amigo site paymentmethods cheque dep conta multibanco numerário vale postal mbway site the current and existing way paypal creditcard this feature should be accessed in the backoffice or it could be implemented in the backoffice powerapp | 1 |
89,633 | 3,798,164,861 | IssuesEvent | 2016-03-23 11:17:04 | bedita/bedita | https://api.github.com/repos/bedita/bedita | closed | cache: add nickname and object_type_id object cache | Priority - High Topic - Core Topic - Frontend Type - Enhancement | Add cache (using object cache or similar method) for two intensively used methods in frontends:
* `BEObject::getType / findObjectTypeId` - performs a query like
```sql
SELECT `BEObject`.`object_type_id` FROM `objects` AS `BEObject` WHERE `BEObject`.`id` = ### LIMIT 1
```
* `BEObject::getIdFromNickname` - performs a query like
```sql
SELECT `BEObject`.`id` FROM `objects` AS `BEObject` WHERE `nickname` = '###' AND status = ('on') LIMIT 1
``` | 1.0 | cache: add nickname and object_type_id object cache - Add cache (using object cache or similar method) for two intensively used methods in frontends:
* `BEObject::getType / findObjectTypeId` - performs a query like
```sql
SELECT `BEObject`.`object_type_id` FROM `objects` AS `BEObject` WHERE `BEObject`.`id` = ### LIMIT 1
```
* `BEObject::getIdFromNickname` - performs a query like
```sql
SELECT `BEObject`.`id` FROM `objects` AS `BEObject` WHERE `nickname` = '###' AND status = ('on') LIMIT 1
``` | priority | cache add nickname and object type id object cache add cache using object cache or similar method for two intensively used methods in frontends beobject gettype findobjecttypeid performs a query like sql select beobject object type id from objects as beobject where beobject id limit beobject getidfromnickname performs a query like sql select beobject id from objects as beobject where nickname and status on limit | 1 |
757,804 | 26,529,508,954 | IssuesEvent | 2023-01-19 11:20:57 | ballerina-platform/ballerina-standard-library | https://api.github.com/repos/ballerina-platform/ballerina-standard-library | opened | GraalVM Check is failing in FTP | Priority/High Type/Task module/ftp graalvm | **Description:**
> $Subject
> Part of https://github.com/ballerina-platform/ballerina-standard-library/issues/3755
GraalVM Check on ftp is failing due to `org.apache.sshd` dependency which is used in `test-utils`. And replacing this dependency or using a docker image for testing may solve this issue since I could not find a solution wrt to GraalVM native-image. Need to check whether it is possible or not
This separate [`ftp-mock-server`](https://github.com/TharmiganK/ftp-mock-server) project may help to engage docker. | 1.0 | GraalVM Check is failing in FTP - **Description:**
> $Subject
> Part of https://github.com/ballerina-platform/ballerina-standard-library/issues/3755
GraalVM Check on ftp is failing due to `org.apache.sshd` dependency which is used in `test-utils`. And replacing this dependency or using a docker image for testing may solve this issue since I could not find a solution wrt to GraalVM native-image. Need to check whether it is possible or not
This separate [`ftp-mock-server`](https://github.com/TharmiganK/ftp-mock-server) project may help to engage docker. | priority | graalvm check is failing in ftp description subject part of graalvm check on ftp is failing due to org apache sshd dependency which is used in test utils and replacing this dependency or using a docker image for testing may solve this issue since i could not find a solution wrt to graalvm native image need to check whether it is possible or not this separate project may help to engage docker | 1 |
154,082 | 5,908,775,671 | IssuesEvent | 2017-05-19 21:22:53 | Polymer/polymer-build | https://api.github.com/repos/Polymer/polymer-build | closed | [1.2.0] Unable to get document FILENAME.HTML: | Priority: High Status: Accepted Type: Bug | ### Description
My build script simply attempts to build a single-file version of Polymer 2.0. There are no custom files imported.
After upgrading to 1.2.0, my Gulp build script fails to build my Polymer bundle. The message is
> Unable to get document polymer-bundle.html:
1.1.0 works as I would expect.
I tracked it down to `deps-index.js` inside of the `buildDepsIndex()` function. I realize this is a dependency of this project, so if I need to move this report, just let me know.
### Versions & Environment
- polymer-build: v1.2.0
- node: v7.9.0
- Operating System: OSX 10.12.4
#### Steps to Reproduce
My `polymer.json` file:
```
{
"entrypoint": "index.html",
"fragments": [
"polymer-bundle.html"
],
"sources": [],
"extraDependencies": [],
"builds": [
{
"name": "bundled",
"bundle": true,
"js": {
"minify": true
},
"css": {
"minify": true
},
"html": {
"minify": true
}
}
],
"lint": {
"rules": [
"polymer-2-hybrid"
]
}
}
```
My `polymer-bundle.html` file:
```
<!-- polymer library -->
<link rel="import" href="bower_components/polymer/polymer.html">
<!-- extra polymer behaviors -->
<link rel="import" href="bower_components/polymer/lib/mixins/gesture-event-listeners.html">
```
#### Expected Results
A bundled version of Polymer and its imported files.
#### Actual Results
An error is thrown and my build fails.
| 1.0 | [1.2.0] Unable to get document FILENAME.HTML: - ### Description
My build script simply attempts to build a single-file version of Polymer 2.0. There are no custom files imported.
After upgrading to 1.2.0, my Gulp build script fails to build my Polymer bundle. The message is
> Unable to get document polymer-bundle.html:
1.1.0 works as I would expect.
I tracked it down to `deps-index.js` inside of the `buildDepsIndex()` function. I realize this is a dependency of this project, so if I need to move this report, just let me know.
### Versions & Environment
- polymer-build: v1.2.0
- node: v7.9.0
- Operating System: OSX 10.12.4
#### Steps to Reproduce
My `polymer.json` file:
```
{
"entrypoint": "index.html",
"fragments": [
"polymer-bundle.html"
],
"sources": [],
"extraDependencies": [],
"builds": [
{
"name": "bundled",
"bundle": true,
"js": {
"minify": true
},
"css": {
"minify": true
},
"html": {
"minify": true
}
}
],
"lint": {
"rules": [
"polymer-2-hybrid"
]
}
}
```
My `polymer-bundle.html` file:
```
<!-- polymer library -->
<link rel="import" href="bower_components/polymer/polymer.html">
<!-- extra polymer behaviors -->
<link rel="import" href="bower_components/polymer/lib/mixins/gesture-event-listeners.html">
```
#### Expected Results
A bundled version of Polymer and its imported files.
#### Actual Results
An error is thrown and my build fails.
| priority | unable to get document filename html description my build script simply attempts to build a single file version of polymer there are no custom files imported after upgrading to my gulp build script fails to build my polymer bundle the message is unable to get document polymer bundle html works as i would expect i tracked it down to deps index js inside of the builddepsindex function i realize this is a dependency of this project so if i need to move this report just let me know versions environment polymer build node operating system osx steps to reproduce my polymer json file entrypoint index html fragments polymer bundle html sources extradependencies builds name bundled bundle true js minify true css minify true html minify true lint rules polymer hybrid my polymer bundle html file expected results a bundled version of polymer and its imported files actual results an error is thrown and my build fails | 1 |
789,626 | 27,798,749,802 | IssuesEvent | 2023-03-17 14:24:27 | PakizeBozkurt/hugsforbugs | https://api.github.com/repos/PakizeBozkurt/hugsforbugs | closed | Contact trainee via email | enhancement Part2 new high priority | ## Story
As a trainee I want to contact other trainees
## Acceptance Criteria
When I see other trainees availability, I click on email. That goes to contact page where I can email them.
I can see success screen when finished
## Implementation Notes
1. Link to the trainee list page (card with trainee details)
2. onClick => 'send calendar request', pass (user logged in) and tranee data via local storage.
3. Use ender's demo email implementation
4. Extra - (not essential) = display a success banner when done - https://getbootstrap.com/docs/4.0/components/alerts/

## Checklist:
- [ Y ] I have defined the user story
- [ Y ] I have defined some acceptance criteria
- [ Y ] I have included any implementation notes
- [ Y ] I have included all relevant links
| 1.0 | Contact trainee via email - ## Story
As a trainee I want to contact other trainees
## Acceptance Criteria
When I see other trainees availability, I click on email. That goes to contact page where I can email them.
I can see success screen when finished
## Implementation Notes
1. Link to the trainee list page (card with trainee details)
2. onClick => 'send calendar request', pass (user logged in) and tranee data via local storage.
3. Use ender's demo email implementation
4. Extra - (not essential) = display a success banner when done - https://getbootstrap.com/docs/4.0/components/alerts/

## Checklist:
- [ Y ] I have defined the user story
- [ Y ] I have defined some acceptance criteria
- [ Y ] I have included any implementation notes
- [ Y ] I have included all relevant links
| priority | contact trainee via email story as a trainee i want to contact other trainees acceptance criteria when i see other trainees availability i click on email that goes to contact page where i can email them i can see success screen when finished implementation notes link to the trainee list page card with trainee details onclick send calendar request pass user logged in and tranee data via local storage use ender s demo email implementation extra not essential display a success banner when done checklist i have defined the user story i have defined some acceptance criteria i have included any implementation notes i have included all relevant links | 1 |
717,462 | 24,676,275,673 | IssuesEvent | 2022-10-18 17:16:28 | statechannels/go-nitro | https://api.github.com/repos/statechannels/go-nitro | closed | generate-adjudicator-bindings.sh is broken | 🐛 bug :bangbang: high-priority | I think we may have merged some broken code, here.
- The "check bindings" action did not run on this PR because we made no changes to the contracts:
https://github.com/statechannels/go-nitro/blob/d9f17f0d93c09514f6f5cc113dee7f6882c040b8/.github/workflows/bindings-check.yml#L6-L7
(We should run the action when the shell script is modified, too)
- the "check bindings" action does not run on `main`, so we haven't yet detected this problem (but soon will on a pull request)
_Originally posted by @geoknee in https://github.com/statechannels/go-nitro/issues/921#issuecomment-1282345894_
| 1.0 | generate-adjudicator-bindings.sh is broken - I think we may have merged some broken code, here.
- The "check bindings" action did not run on this PR because we made no changes to the contracts:
https://github.com/statechannels/go-nitro/blob/d9f17f0d93c09514f6f5cc113dee7f6882c040b8/.github/workflows/bindings-check.yml#L6-L7
(We should run the action when the shell script is modified, too)
- the "check bindings" action does not run on `main`, so we haven't yet detected this problem (but soon will on a pull request)
_Originally posted by @geoknee in https://github.com/statechannels/go-nitro/issues/921#issuecomment-1282345894_
| priority | generate adjudicator bindings sh is broken i think we may have merged some broken code here the check bindings action did not run on this pr because we made no changes to the contracts we should run the action when the shell script is modified too the check bindings action does not run on main so we haven t yet detected this problem but soon will on a pull request originally posted by geoknee in | 1 |
203,528 | 7,065,311,575 | IssuesEvent | 2018-01-06 18:28:31 | mattbdean/Helium | https://api.github.com/repos/mattbdean/Helium | closed | Inserting zero entries to a part table | bug high priority | If I have a master__part table (e.g. `session__trial`), if I try to insert only to the `session` table with nothing to the `trial` part table I get the following error:
``` bash
{ message: 'generated invalid SQL',
data: { session: [ [Object] ], session__trial: [] } }
{ Error: ER_PARSE_ERROR: You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near '' at line 1
at Query.Sequence._packetToError (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/sequences/Sequence.js:52:14)
at Query.ErrorPacket (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/sequences/Query.js:77:18)
at Protocol._parsePacket (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:279:23)
at Parser.write (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Parser.js:76:12)
at Protocol.write (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:39:16)
at Socket.<anonymous> (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/Connection.js:103:28)
at emitOne (events.js:116:13)
at Socket.emit (events.js:211:7)
at addChunk (_stream_readable.js:263:12)
at readableAddChunk (_stream_readable.js:250:11)
at Socket.Readable.push (_stream_readable.js:208:10)
at TCP.onread (net.js:594:20)
--------------------
at Protocol._enqueue (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:145:48)
at PoolConnection.query (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/Connection.js:208:25)
at /Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/helper.js:16:24
at Promise._execute (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/debuggability.js:303:9)
at Promise._resolveFromExecutor (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/promise.js:483:18)
at new Promise (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/promise.js:79:10)
at PoolConnection.promiseCallback (/Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/helper.js:6:10)
at poolConnection.connection.query (/Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/connection.js:64:28)
at QueryHelper.<anonymous> (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:24:54)
at Generator.next (<anonymous>)
at /Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:7:71
at new Promise (<anonymous>)
at __awaiter (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:3:12)
at QueryHelper.execute (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:22:16)
at SchemaDao.<anonymous> (/Users/moharb/Documents/Repos/Helium/dist/routes/api/schemas/schema.dao.js:177:39)
at Generator.next (<anonymous>)
code: 'ER_PARSE_ERROR',
errno: 1064,
sqlMessage: 'You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near \'\' at line 1',
sqlState: '42000',
index: 0,
sql: 'INSERT INTO `boazmohar_experiment`.`session__trial`' }
PUT /api/v1/schemas/boazmohar_experiment/data 500 104.906 ms - 3155
```
If I add a single entry to the `trial` table the form works, also the same inset to the `session` table works from the dj python API.
| 1.0 | Inserting zero entries to a part table - If I have a master__part table (e.g. `session__trial`), if I try to insert only to the `session` table with nothing to the `trial` part table I get the following error:
``` bash
{ message: 'generated invalid SQL',
data: { session: [ [Object] ], session__trial: [] } }
{ Error: ER_PARSE_ERROR: You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near '' at line 1
at Query.Sequence._packetToError (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/sequences/Sequence.js:52:14)
at Query.ErrorPacket (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/sequences/Query.js:77:18)
at Protocol._parsePacket (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:279:23)
at Parser.write (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Parser.js:76:12)
at Protocol.write (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:39:16)
at Socket.<anonymous> (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/Connection.js:103:28)
at emitOne (events.js:116:13)
at Socket.emit (events.js:211:7)
at addChunk (_stream_readable.js:263:12)
at readableAddChunk (_stream_readable.js:250:11)
at Socket.Readable.push (_stream_readable.js:208:10)
at TCP.onread (net.js:594:20)
--------------------
at Protocol._enqueue (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:145:48)
at PoolConnection.query (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/Connection.js:208:25)
at /Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/helper.js:16:24
at Promise._execute (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/debuggability.js:303:9)
at Promise._resolveFromExecutor (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/promise.js:483:18)
at new Promise (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/promise.js:79:10)
at PoolConnection.promiseCallback (/Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/helper.js:6:10)
at poolConnection.connection.query (/Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/connection.js:64:28)
at QueryHelper.<anonymous> (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:24:54)
at Generator.next (<anonymous>)
at /Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:7:71
at new Promise (<anonymous>)
at __awaiter (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:3:12)
at QueryHelper.execute (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:22:16)
at SchemaDao.<anonymous> (/Users/moharb/Documents/Repos/Helium/dist/routes/api/schemas/schema.dao.js:177:39)
at Generator.next (<anonymous>)
code: 'ER_PARSE_ERROR',
errno: 1064,
sqlMessage: 'You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near \'\' at line 1',
sqlState: '42000',
index: 0,
sql: 'INSERT INTO `boazmohar_experiment`.`session__trial`' }
PUT /api/v1/schemas/boazmohar_experiment/data 500 104.906 ms - 3155
```
If I add a single entry to the `trial` table the form works, also the same inset to the `session` table works from the dj python API.
| priority | inserting zero entries to a part table if i have a master part table e g session trial if i try to insert only to the session table with nothing to the trial part table i get the following error bash message generated invalid sql data session session trial error er parse error you have an error in your sql syntax check the manual that corresponds to your mariadb server version for the right syntax to use near at line at query sequence packettoerror users moharb documents repos helium node modules mysql lib protocol sequences sequence js at query errorpacket users moharb documents repos helium node modules mysql lib protocol sequences query js at protocol parsepacket users moharb documents repos helium node modules mysql lib protocol protocol js at parser write users moharb documents repos helium node modules mysql lib protocol parser js at protocol write users moharb documents repos helium node modules mysql lib protocol protocol js at socket users moharb documents repos helium node modules mysql lib connection js at emitone events js at socket emit events js at addchunk stream readable js at readableaddchunk stream readable js at socket readable push stream readable js at tcp onread net js at protocol enqueue users moharb documents repos helium node modules mysql lib protocol protocol js at poolconnection query users moharb documents repos helium node modules mysql lib connection js at users moharb documents repos helium node modules promise mysql lib helper js at promise execute users moharb documents repos helium node modules bluebird js release debuggability js at promise resolvefromexecutor users moharb documents repos helium node modules bluebird js release promise js at new promise users moharb documents repos helium node modules bluebird js release promise js at poolconnection promisecallback users moharb documents repos helium node modules promise mysql lib helper js at poolconnection connection query users moharb documents repos helium node modules promise mysql lib connection js at queryhelper users moharb documents repos helium dist db query helper js at generator next at users moharb documents repos helium dist db query helper js at new promise at awaiter users moharb documents repos helium dist db query helper js at queryhelper execute users moharb documents repos helium dist db query helper js at schemadao users moharb documents repos helium dist routes api schemas schema dao js at generator next code er parse error errno sqlmessage you have an error in your sql syntax check the manual that corresponds to your mariadb server version for the right syntax to use near at line sqlstate index sql insert into boazmohar experiment session trial put api schemas boazmohar experiment data ms if i add a single entry to the trial table the form works also the same inset to the session table works from the dj python api | 1 |
267,148 | 8,379,544,506 | IssuesEvent | 2018-10-07 03:27:29 | doroxz/test | https://api.github.com/repos/doroxz/test | opened | Add awesome function | Priority/high status/pending type/feature | **Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is.
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | Add awesome function - **Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is.
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
**Additional context**
Add any other context or screenshots about the feature request here.
| priority | add awesome function is your feature request related to a problem please describe a clear and concise description of what the problem is describe the solution you d like a clear and concise description of what you want to happen additional context add any other context or screenshots about the feature request here | 1 |
407,233 | 11,908,488,975 | IssuesEvent | 2020-03-31 01:07:10 | theintern/intern | https://api.github.com/repos/theintern/intern | opened | Call emit() listeners when events are emitted | bug effort-low priority-high | Currently (Intern 4.8 and Intern 5.0-pre), listener calls are chained in `Executor#emit`.
```ts
emit(event, data) {
for (const listener of listeners) {
notifications = notifications
.then(() => Task.resolve(listener(data)))
.then(handleErrorEvent)
.catch(handleListenerError);
}
// ...
return notifications;
}
```
This can lead to handlers receiving events out of order -- a new event may be emitted before the chain for the previous event has finished processing. It's also problematic for situations where timing is important since event handlers may be called at some arbitrary time after an event was actually emitted.
Listeners are not inter-dependent; it shouldn't matter (and doesn't currently) in which order they're called, and there's no need to wait on one before calling the other, so they should all be called at the same time, something more like:
```ts
const notifications = [];
for (const listener of listeners) {
notifications.push(
Task.resolve(listener(data))
.then(handleErrorEvent)
.catch(handleListenerError)
);
}
// ...
return Promise.all(notifications);
``` | 1.0 | Call emit() listeners when events are emitted - Currently (Intern 4.8 and Intern 5.0-pre), listener calls are chained in `Executor#emit`.
```ts
emit(event, data) {
for (const listener of listeners) {
notifications = notifications
.then(() => Task.resolve(listener(data)))
.then(handleErrorEvent)
.catch(handleListenerError);
}
// ...
return notifications;
}
```
This can lead to handlers receiving events out of order -- a new event may be emitted before the chain for the previous event has finished processing. It's also problematic for situations where timing is important since event handlers may be called at some arbitrary time after an event was actually emitted.
Listeners are not inter-dependent; it shouldn't matter (and doesn't currently) in which order they're called, and there's no need to wait on one before calling the other, so they should all be called at the same time, something more like:
```ts
const notifications = [];
for (const listener of listeners) {
notifications.push(
Task.resolve(listener(data))
.then(handleErrorEvent)
.catch(handleListenerError)
);
}
// ...
return Promise.all(notifications);
``` | priority | call emit listeners when events are emitted currently intern and intern pre listener calls are chained in executor emit ts emit event data for const listener of listeners notifications notifications then task resolve listener data then handleerrorevent catch handlelistenererror return notifications this can lead to handlers receiving events out of order a new event may be emitted before the chain for the previous event has finished processing it s also problematic for situations where timing is important since event handlers may be called at some arbitrary time after an event was actually emitted listeners are not inter dependent it shouldn t matter and doesn t currently in which order they re called and there s no need to wait on one before calling the other so they should all be called at the same time something more like ts const notifications for const listener of listeners notifications push task resolve listener data then handleerrorevent catch handlelistenererror return promise all notifications | 1 |
138,376 | 5,332,719,679 | IssuesEvent | 2017-02-15 22:51:28 | influxdata/docs.influxdata.com | https://api.github.com/repos/influxdata/docs.influxdata.com | opened | [Enterprise] Clarify influxd-ctl options | Priority: high | https://docs.influxdata.com/enterprise/v1.2/features/cluster-commands/ and https://docs.influxdata.com/enterprise/v1.2/guides/backup-and-restore/
Link backup and restore to the cluster-commands page to show that every option is applicable to the backup/restore influxd-ctl commands. | 1.0 | [Enterprise] Clarify influxd-ctl options - https://docs.influxdata.com/enterprise/v1.2/features/cluster-commands/ and https://docs.influxdata.com/enterprise/v1.2/guides/backup-and-restore/
Link backup and restore to the cluster-commands page to show that every option is applicable to the backup/restore influxd-ctl commands. | priority | clarify influxd ctl options and link backup and restore to the cluster commands page to show that every option is applicable to the backup restore influxd ctl commands | 1 |
498,023 | 14,399,022,243 | IssuesEvent | 2020-12-03 10:21:45 | wso2/kubernetes-is | https://api.github.com/repos/wso2/kubernetes-is | closed | [5.11.0] Introduce Kubernetes Resources for Simplified Setup of WSO2 IAM | Priority/High Type/Task | **Description:**
Introduce Kubernetes resources for a simplified setup of WSO2 Identity and Access Management version `5.11.0` deployment. | 1.0 | [5.11.0] Introduce Kubernetes Resources for Simplified Setup of WSO2 IAM - **Description:**
Introduce Kubernetes resources for a simplified setup of WSO2 Identity and Access Management version `5.11.0` deployment. | priority | introduce kubernetes resources for simplified setup of iam description introduce kubernetes resources for a simplified setup of identity and access management version deployment | 1 |
742,896 | 25,876,436,118 | IssuesEvent | 2022-12-14 08:15:39 | minio/minio | https://api.github.com/repos/minio/minio | closed | Rebalancing transfers almost all data to other pool | community fixed priority: high | Site-replication between sites with different capacity.
The site with the smallest capacity had one server pool (one node, 4 disks of 1GB each) which reached full capacity.
Expanded this site by adding a new server pool. (same capacity in new pool, 4 disks of 1GB each)
We tried 'mc admin rebalance' to rebalance data over pools. It didn't balance, it just transferred as much possible data to the new pool.
## Expected Behavior
'mc admin rebalance status'
Per-pool usage before rebalancing:
- Pool-0 : 0.97%
- Pool-1 : 0.13%
Per-pool usage after rebalancing:
- Pool-0: 0.55%
- Pool-1: 0.55%
Not exactly 0.55%-0.55%, but at least somewhat close.
## Current Behavior
'mc admin rebalance status'
Per-pool usage before rebalancing:
- Pool-0 : 0.97%
- Pool-1 : 0.13%
Per-pool usage after rebalancing:
- Pool-0: 0.11%
- Pool-1: 0.99%
## Steps to Reproduce (for bugs)
1. Fill one server pool with data
2. Expand by adding a server pool with same capacity
3. mc admin rebalance
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio --version`): RELEASE.2022-11-17T23-20-09Z (commit-id=a22b4adf4c5cc3e4db13fe92da683ef1ce45cd5a)
* Operating System and version (`uname -a`):Linux minio01 5.10.0-15-amd64 SMP Debian 5.10.120-1 (2022-06-09) x86_64 GNU/Linux
| 1.0 | Rebalancing transfers almost all data to other pool - Site-replication between sites with different capacity.
The site with the smallest capacity had one server pool (one node, 4 disks of 1GB each) which reached full capacity.
Expanded this site by adding a new server pool. (same capacity in new pool, 4 disks of 1GB each)
We tried 'mc admin rebalance' to rebalance data over pools. It didn't balance, it just transferred as much possible data to the new pool.
## Expected Behavior
'mc admin rebalance status'
Per-pool usage before rebalancing:
- Pool-0 : 0.97%
- Pool-1 : 0.13%
Per-pool usage after rebalancing:
- Pool-0: 0.55%
- Pool-1: 0.55%
Not exactly 0.55%-0.55%, but at least somewhat close.
## Current Behavior
'mc admin rebalance status'
Per-pool usage before rebalancing:
- Pool-0 : 0.97%
- Pool-1 : 0.13%
Per-pool usage after rebalancing:
- Pool-0: 0.11%
- Pool-1: 0.99%
## Steps to Reproduce (for bugs)
1. Fill one server pool with data
2. Expand by adding a server pool with same capacity
3. mc admin rebalance
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio --version`): RELEASE.2022-11-17T23-20-09Z (commit-id=a22b4adf4c5cc3e4db13fe92da683ef1ce45cd5a)
* Operating System and version (`uname -a`):Linux minio01 5.10.0-15-amd64 SMP Debian 5.10.120-1 (2022-06-09) x86_64 GNU/Linux
| priority | rebalancing transfers almost all data to other pool site replication between sites with different capacity the site with the smallest capacity had one server pool one node disks of each which reached full capacity expanded this site by adding a new server pool same capacity in new pool disks of each we tried mc admin rebalance to rebalance data over pools it didn t balance it just transferred as much possible data to the new pool expected behavior mc admin rebalance status per pool usage before rebalancing pool pool per pool usage after rebalancing pool pool not exactly but at least somewhat close current behavior mc admin rebalance status per pool usage before rebalancing pool pool per pool usage after rebalancing pool pool steps to reproduce for bugs fill one server pool with data expand by adding a server pool with same capacity mc admin rebalance your environment version used minio version release commit id operating system and version uname a linux smp debian gnu linux | 1 |
713,819 | 24,540,990,655 | IssuesEvent | 2022-10-12 03:43:23 | ipfs-force-community/venus-cluster | https://api.github.com/repos/ipfs-force-community/venus-cluster | closed | [venus-cluster] 扇区重建开发规划 | enhancement priority:high | 本 issue 用于规划和跟踪 `扇区重建` 功能的开发
对于以下的检查项,根据必要性,可能进行重组织、继续拆分、转换成独立 issue 等不同处理
- [x] venus-sector-manager
- [x] 重建任务 #373
- [x] 数据结构
- [x] 增删改查
- [x] 重建过程相关接口
- [x] venus-worker
- [x] 热更新 `sealing_thread` 配置
- [x] 改造 planner 的状态转移,使 `sealing_thread` 具备跳出无任务的空循环的能力 #387
- [x] 提供 `sealing_thread` 在每个循环初始重载配置的能力 #376
- [x] 实现 `rebuild` planner | 1.0 | [venus-cluster] 扇区重建开发规划 - 本 issue 用于规划和跟踪 `扇区重建` 功能的开发
对于以下的检查项,根据必要性,可能进行重组织、继续拆分、转换成独立 issue 等不同处理
- [x] venus-sector-manager
- [x] 重建任务 #373
- [x] 数据结构
- [x] 增删改查
- [x] 重建过程相关接口
- [x] venus-worker
- [x] 热更新 `sealing_thread` 配置
- [x] 改造 planner 的状态转移,使 `sealing_thread` 具备跳出无任务的空循环的能力 #387
- [x] 提供 `sealing_thread` 在每个循环初始重载配置的能力 #376
- [x] 实现 `rebuild` planner | priority | 扇区重建开发规划 本 issue 用于规划和跟踪 扇区重建 功能的开发 对于以下的检查项,根据必要性,可能进行重组织、继续拆分、转换成独立 issue 等不同处理 venus sector manager 重建任务 数据结构 增删改查 重建过程相关接口 venus worker 热更新 sealing thread 配置 改造 planner 的状态转移,使 sealing thread 具备跳出无任务的空循环的能力 提供 sealing thread 在每个循环初始重载配置的能力 实现 rebuild planner | 1 |
90,092 | 3,810,570,615 | IssuesEvent | 2016-03-26 06:17:27 | osmosacademy/vr-campaign | https://api.github.com/repos/osmosacademy/vr-campaign | opened | Arrange coral reefs into a pretty scene | help wanted priority:high status:blocked | Once the coral reefs are created:
https://github.com/osmosacademy/vr-campaign/issues/11
https://github.com/osmosacademy/vr-campaign/issues/6
Someone has to arrange them into a pretty looking underwater scene in Unity. We may end up just using that scene as our one scene. Or, it will be used to inspire a programmer to randomly generate new levels, based on how you arrange the coral reefs and other things. It will also greatly help us understand and visualize the flow and mechanics of the game.
To set up the scene, please learn about nesting game objects in unity to organize your scene. Look it up or ask someone for help. Like these pictures:



| 1.0 | Arrange coral reefs into a pretty scene - Once the coral reefs are created:
https://github.com/osmosacademy/vr-campaign/issues/11
https://github.com/osmosacademy/vr-campaign/issues/6
Someone has to arrange them into a pretty looking underwater scene in Unity. We may end up just using that scene as our one scene. Or, it will be used to inspire a programmer to randomly generate new levels, based on how you arrange the coral reefs and other things. It will also greatly help us understand and visualize the flow and mechanics of the game.
To set up the scene, please learn about nesting game objects in unity to organize your scene. Look it up or ask someone for help. Like these pictures:



| priority | arrange coral reefs into a pretty scene once the coral reefs are created someone has to arrange them into a pretty looking underwater scene in unity we may end up just using that scene as our one scene or it will be used to inspire a programmer to randomly generate new levels based on how you arrange the coral reefs and other things it will also greatly help us understand and visualize the flow and mechanics of the game to set up the scene please learn about nesting game objects in unity to organize your scene look it up or ask someone for help like these pictures | 1 |
47,856 | 2,986,639,654 | IssuesEvent | 2015-07-20 05:56:37 | UnifiedViews/Plugins | https://api.github.com/repos/UnifiedViews/Plugins | closed | e-rdfFromSparql has to be reworked | priority: High resolution: fixed severity: enhancement | So far, e-rdfFromSparql was not update to helpers 2.0 because updating the DPU would take more effort than rewriting it. So we should rewrite it using helpers 2.0.
TODO: Add reference to the @skodape DPU in mff-uk/DPUs which may be used as a base. | 1.0 | e-rdfFromSparql has to be reworked - So far, e-rdfFromSparql was not update to helpers 2.0 because updating the DPU would take more effort than rewriting it. So we should rewrite it using helpers 2.0.
TODO: Add reference to the @skodape DPU in mff-uk/DPUs which may be used as a base. | priority | e rdffromsparql has to be reworked so far e rdffromsparql was not update to helpers because updating the dpu would take more effort than rewriting it so we should rewrite it using helpers todo add reference to the skodape dpu in mff uk dpus which may be used as a base | 1 |
41,120 | 2,868,977,171 | IssuesEvent | 2015-06-05 22:19:49 | SCIInstitute/shapeworks | https://api.github.com/repos/SCIInstitute/shapeworks | closed | Allow View2 to open without command line. | enhancement High Priority IBBM | The first thing to happen with view2 is a dialog to pop up asking for the parameter file. If it is a good parameter file, then everything should work as normal. If not, program will keep asking for a good one, or to quit. | 1.0 | Allow View2 to open without command line. - The first thing to happen with view2 is a dialog to pop up asking for the parameter file. If it is a good parameter file, then everything should work as normal. If not, program will keep asking for a good one, or to quit. | priority | allow to open without command line the first thing to happen with is a dialog to pop up asking for the parameter file if it is a good parameter file then everything should work as normal if not program will keep asking for a good one or to quit | 1 |
822,475 | 30,874,155,311 | IssuesEvent | 2023-08-03 13:18:54 | unlock-protocol/unlock | https://api.github.com/repos/unlock-protocol/unlock | closed | Can't find a way to disconnect from the Events dapp using Brave | 🚨 High Priority | **Describe the bug**
Using Brave on laptop, cannot see any way to disconnect from the app.
**To Reproduce**
Using Brave, go to https://events.unlock-protocol.com
For some reason (not sure why) the Events app was showing me as connected.
Tried to disconnect and could not find a way to do so.
(On the other hand, using the Chrome browser and proper wallet, things work as expected.)
**Expected behavior**
I would expect to see some way to "disconnect."
**Screenshots**
https://www.loom.com/share/fd220211250049778398413225fe83af?sid=fef511c6-865b-4a16-802c-a4154d50583f | 1.0 | Can't find a way to disconnect from the Events dapp using Brave - **Describe the bug**
Using Brave on laptop, cannot see any way to disconnect from the app.
**To Reproduce**
Using Brave, go to https://events.unlock-protocol.com
For some reason (not sure why) the Events app was showing me as connected.
Tried to disconnect and could not find a way to do so.
(On the other hand, using the Chrome browser and proper wallet, things work as expected.)
**Expected behavior**
I would expect to see some way to "disconnect."
**Screenshots**
https://www.loom.com/share/fd220211250049778398413225fe83af?sid=fef511c6-865b-4a16-802c-a4154d50583f | priority | can t find a way to disconnect from the events dapp using brave describe the bug using brave on laptop cannot see any way to disconnect from the app to reproduce using brave go to for some reason not sure why the events app was showing me as connected tried to disconnect and could not find a way to do so on the other hand using the chrome browser and proper wallet things work as expected expected behavior i would expect to see some way to disconnect screenshots | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.