Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
609,100
18,853,997,419
IssuesEvent
2021-11-12 02:10:37
lokka30/Treasury
https://api.github.com/repos/lokka30/Treasury
closed
Address the difficulty of use with CompletableFutures, etc.
help wanted priority: high approved
Please see #37 which is an in-progress resolution to this issue.
1.0
Address the difficulty of use with CompletableFutures, etc. - Please see #37 which is an in-progress resolution to this issue.
priority
address the difficulty of use with completablefutures etc please see which is an in progress resolution to this issue
1
73,830
3,421,786,014
IssuesEvent
2015-12-08 20:12:42
tipsybear/actors-simulation
https://api.github.com/repos/tipsybear/actors-simulation
closed
Baseline Simulation
priority: high type: feature
Here is the baseline simulation that we need to show that our simulation model is realistic and serves as a proof of concept about how our cluster compares to the actor simulators. The simulation is composed of programs (1 per processor) that do the following: 1. wait for a message 2. sleep for a random amount of time (work) 3. send message(s) of varying sizes to other nodes The input (parameters of the simulation are as follows: 1. # of nodes that initially receive messages (to get things started) 2. (min, max) sleep time (amount of work() 3. (min, max) message size (This is the key as we're measuring latency in a fully utilized cluster.) Output: a time series of the following aggregate information: 1. # of messages in flight 2. total (average?) message size in flight (e.g. the bandwidth capacity) 3. current average latency (which means we have to measure how long it takes a message to arrive) 4. available bandwidth though this might be the same as 2. The visualization of this time series will be reported in the presentation.
1.0
Baseline Simulation - Here is the baseline simulation that we need to show that our simulation model is realistic and serves as a proof of concept about how our cluster compares to the actor simulators. The simulation is composed of programs (1 per processor) that do the following: 1. wait for a message 2. sleep for a random amount of time (work) 3. send message(s) of varying sizes to other nodes The input (parameters of the simulation are as follows: 1. # of nodes that initially receive messages (to get things started) 2. (min, max) sleep time (amount of work() 3. (min, max) message size (This is the key as we're measuring latency in a fully utilized cluster.) Output: a time series of the following aggregate information: 1. # of messages in flight 2. total (average?) message size in flight (e.g. the bandwidth capacity) 3. current average latency (which means we have to measure how long it takes a message to arrive) 4. available bandwidth though this might be the same as 2. The visualization of this time series will be reported in the presentation.
priority
baseline simulation here is the baseline simulation that we need to show that our simulation model is realistic and serves as a proof of concept about how our cluster compares to the actor simulators the simulation is composed of programs per processor that do the following wait for a message sleep for a random amount of time work send message s of varying sizes to other nodes the input parameters of the simulation are as follows of nodes that initially receive messages to get things started min max sleep time amount of work min max message size this is the key as we re measuring latency in a fully utilized cluster output a time series of the following aggregate information of messages in flight total average message size in flight e g the bandwidth capacity current average latency which means we have to measure how long it takes a message to arrive available bandwidth though this might be the same as the visualization of this time series will be reported in the presentation
1
378,719
11,206,730,695
IssuesEvent
2020-01-05 23:34:39
akiraux/Akira
https://api.github.com/repos/akiraux/Akira
closed
Split the Canvas
Not an Issue Priority: High
The `Canvas.vala` file is growing exponentially at every PR. Since that's the main core of our application, we should define how to divide it into multiple classes in order to keep our core library as scalable and modular as possible. Let's throw a bunch of ideas on how to approach this, and open some draft PRs to prototype solutions.
1.0
Split the Canvas - The `Canvas.vala` file is growing exponentially at every PR. Since that's the main core of our application, we should define how to divide it into multiple classes in order to keep our core library as scalable and modular as possible. Let's throw a bunch of ideas on how to approach this, and open some draft PRs to prototype solutions.
priority
split the canvas the canvas vala file is growing exponentially at every pr since that s the main core of our application we should define how to divide it into multiple classes in order to keep our core library as scalable and modular as possible let s throw a bunch of ideas on how to approach this and open some draft prs to prototype solutions
1
414,168
12,100,100,661
IssuesEvent
2020-04-20 13:18:13
cilium/cilium
https://api.github.com/repos/cilium/cilium
closed
clustermesh issue: io.cilium/global-service=true is not working for an existing service
area/clustermesh kind/community-report priority/high
<!-- If you have usage questions, please try the [slack channel](http://cilium.io/slack) and see the [FAQ](https://goo.gl/qG2YmU) first. Choose either "Proposal" or "Bug report" --> ## Bug report Here's what I did 1. I took your clustermesh/rebel-base example and downloaded your cluster1.yml and cluster2.yml. 2. Removed the "io.cilium/global-service=true" from the rebel-base service. 3. Renamed rebel-base to bipin-base as I was having your original service running to test the environment. alias kl='kubectl config get-contexts' alias ka='kubectl config use-context a.xxxxxxxxxxxx' alias kb='kubectl config use-context b.xxxxxxxxxxx' alias ks='kubectl -n kube-system' alias k='kubectl' ka k apply -f k8s-yaml/nginx/cluster1.yml kb k apply -f k8s-yaml/nginx/cluster2.yml ka kubectl annotate svc bipin-base io.cilium/global-service=true kb kubectl annotate svc bipin-base io.cilium/global-service=true k get svc NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE bipin-base ClusterIP 100.64.9.182 <none> 80/TCP 10m kubernetes ClusterIP 100.64.0.1 <none> 443/TCP 4h ks exec cilium-cq9vw -- cilium service list ID Frontend Backend 1 100.64.0.1:443 1 => 10.0.1.206:443 3 100.64.0.10:53 1 => 100.64.129.212:53 2 => 100.64.129.245:53 4 100.64.26.37:2379 1 => 100.64.129.101:2379 2 => 100.64.129.193:2379 3 => 100.64.129.27:2379 5 100.64.18.136:2379 1 => 100.64.129.101:2379 2 => 100.64.129.193:2379 3 => 100.64.129.27:2379 8 100.64.9.182:80 1 => 100.64.129.102:80 2 => 100.64.129.128:80 Please look at the service-id=8 It show only the local ips. The cluster1 and cluster2 YAML files which came with your examples works file. That proves that this global-service annotation is not being picked for an existing service. As a workaround if you can tell me how to update the service definition manually for time being for demo'ing this to team...it would help greatly <!-- Important: For security related issues: We strongly encourage you to report security vulnerabilities to our private security mailing list: security@cilium.io - first, before disclosing them in any public forums. --> **General Information** - Cilium version (run `cilium version`) ks exec cilium-cq9vw -- cilium version Client: 1.4.90 3adbd7f 2019-02-12T19:27:16-08:00 go version go1.11.1 linux/amd64 Daemon: 1.4.90 3adbd7f 2019-02-12T19:27:16-08:00 go version go1.11.1 linux/amd64 - Kernel version (run `uname -a`) 4.14.96-coreos - Orchestration system version in use (e.g. `kubectl version`, Mesos, ...) kubectl version Client Version: version.Info{Major:"1", Minor:"12", GitVersion:"v1.12.2", GitCommit:"17c77c7898218073f14c8d573582e8d2313dc740", GitTreeState:"clean", BuildDate:"2018-10-24T06:54:59Z", GoVersion:"go1.10.4", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.6", GitCommit:"b1d75deca493a24a2f87eb1efde1a569e52fc8d9", GitTreeState:"clean", BuildDate:"2018-12-16T04:30:10Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"} - Link to relevant artifacts (policies, deployments scripts, ...) [nginx.zip](https://github.com/cilium/cilium/files/2860741/nginx.zip) - Upload a system dump (run `curl -sLO [cluster-diagnosis.zip](https://github.com/cilium/cilium/files/2860756/cluster-diagnosis.zip) releases.cilium.io/tools/cluster-diagnosis.zip && python cluster-diagnosis.zip sysdump` and then attach the generated zip file) [cluster-diagnosis.zip](https://github.com/cilium/cilium/files/2860757/cluster-diagnosis.zip) **How to reproduce the issue** 1. instruction 1 2. instruction 2
1.0
clustermesh issue: io.cilium/global-service=true is not working for an existing service - <!-- If you have usage questions, please try the [slack channel](http://cilium.io/slack) and see the [FAQ](https://goo.gl/qG2YmU) first. Choose either "Proposal" or "Bug report" --> ## Bug report Here's what I did 1. I took your clustermesh/rebel-base example and downloaded your cluster1.yml and cluster2.yml. 2. Removed the "io.cilium/global-service=true" from the rebel-base service. 3. Renamed rebel-base to bipin-base as I was having your original service running to test the environment. alias kl='kubectl config get-contexts' alias ka='kubectl config use-context a.xxxxxxxxxxxx' alias kb='kubectl config use-context b.xxxxxxxxxxx' alias ks='kubectl -n kube-system' alias k='kubectl' ka k apply -f k8s-yaml/nginx/cluster1.yml kb k apply -f k8s-yaml/nginx/cluster2.yml ka kubectl annotate svc bipin-base io.cilium/global-service=true kb kubectl annotate svc bipin-base io.cilium/global-service=true k get svc NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE bipin-base ClusterIP 100.64.9.182 <none> 80/TCP 10m kubernetes ClusterIP 100.64.0.1 <none> 443/TCP 4h ks exec cilium-cq9vw -- cilium service list ID Frontend Backend 1 100.64.0.1:443 1 => 10.0.1.206:443 3 100.64.0.10:53 1 => 100.64.129.212:53 2 => 100.64.129.245:53 4 100.64.26.37:2379 1 => 100.64.129.101:2379 2 => 100.64.129.193:2379 3 => 100.64.129.27:2379 5 100.64.18.136:2379 1 => 100.64.129.101:2379 2 => 100.64.129.193:2379 3 => 100.64.129.27:2379 8 100.64.9.182:80 1 => 100.64.129.102:80 2 => 100.64.129.128:80 Please look at the service-id=8 It show only the local ips. The cluster1 and cluster2 YAML files which came with your examples works file. That proves that this global-service annotation is not being picked for an existing service. As a workaround if you can tell me how to update the service definition manually for time being for demo'ing this to team...it would help greatly <!-- Important: For security related issues: We strongly encourage you to report security vulnerabilities to our private security mailing list: security@cilium.io - first, before disclosing them in any public forums. --> **General Information** - Cilium version (run `cilium version`) ks exec cilium-cq9vw -- cilium version Client: 1.4.90 3adbd7f 2019-02-12T19:27:16-08:00 go version go1.11.1 linux/amd64 Daemon: 1.4.90 3adbd7f 2019-02-12T19:27:16-08:00 go version go1.11.1 linux/amd64 - Kernel version (run `uname -a`) 4.14.96-coreos - Orchestration system version in use (e.g. `kubectl version`, Mesos, ...) kubectl version Client Version: version.Info{Major:"1", Minor:"12", GitVersion:"v1.12.2", GitCommit:"17c77c7898218073f14c8d573582e8d2313dc740", GitTreeState:"clean", BuildDate:"2018-10-24T06:54:59Z", GoVersion:"go1.10.4", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.6", GitCommit:"b1d75deca493a24a2f87eb1efde1a569e52fc8d9", GitTreeState:"clean", BuildDate:"2018-12-16T04:30:10Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"} - Link to relevant artifacts (policies, deployments scripts, ...) [nginx.zip](https://github.com/cilium/cilium/files/2860741/nginx.zip) - Upload a system dump (run `curl -sLO [cluster-diagnosis.zip](https://github.com/cilium/cilium/files/2860756/cluster-diagnosis.zip) releases.cilium.io/tools/cluster-diagnosis.zip && python cluster-diagnosis.zip sysdump` and then attach the generated zip file) [cluster-diagnosis.zip](https://github.com/cilium/cilium/files/2860757/cluster-diagnosis.zip) **How to reproduce the issue** 1. instruction 1 2. instruction 2
priority
clustermesh issue io cilium global service true is not working for an existing service if you have usage questions please try the slack channel and see the first choose either proposal or bug report bug report here s what i did i took your clustermesh rebel base example and downloaded your yml and yml removed the io cilium global service true from the rebel base service renamed rebel base to bipin base as i was having your original service running to test the environment alias kl kubectl config get contexts alias ka kubectl config use context a xxxxxxxxxxxx alias kb kubectl config use context b xxxxxxxxxxx alias ks kubectl n kube system alias k kubectl ka k apply f yaml nginx yml kb k apply f yaml nginx yml ka kubectl annotate svc bipin base io cilium global service true kb kubectl annotate svc bipin base io cilium global service true k get svc name type cluster ip external ip port s age bipin base clusterip tcp kubernetes clusterip tcp ks exec cilium cilium service list id frontend backend please look at the service id it show only the local ips the and yaml files which came with your examples works file that proves that this global service annotation is not being picked for an existing service as a workaround if you can tell me how to update the service definition manually for time being for demo ing this to team it would help greatly important for security related issues we strongly encourage you to report security vulnerabilities to our private security mailing list security cilium io first before disclosing them in any public forums general information cilium version run cilium version ks exec cilium cilium version client go version linux daemon go version linux kernel version run uname a coreos orchestration system version in use e g kubectl version mesos kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux link to relevant artifacts policies deployments scripts upload a system dump run curl slo releases cilium io tools cluster diagnosis zip python cluster diagnosis zip sysdump and then attach the generated zip file how to reproduce the issue instruction instruction
1
643,398
20,956,223,631
IssuesEvent
2022-03-27 05:56:35
AY2122S2-CS2103T-T09-2/tp
https://api.github.com/repos/AY2122S2-CS2103T-T09-2/tp
closed
As a Recruiter, I want to be able to search applicants by name
type.Story priority.High
so that I can view their job applications if they have applied for more than 1 job
1.0
As a Recruiter, I want to be able to search applicants by name - so that I can view their job applications if they have applied for more than 1 job
priority
as a recruiter i want to be able to search applicants by name so that i can view their job applications if they have applied for more than job
1
306,515
9,395,935,249
IssuesEvent
2019-04-08 05:16:07
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Outbound provisioning to Salesforce does not work as expected
Affected/5.8.0-Alpha2 Complexity/Medium Component/Extensions Priority/High Resolution/Cannot Reproduce Severity/Blocker Type/Bug
**Suggested Labels** Affected: 5.8.0-alpha2 Severity: Major Priority: High Component: Salesforce Outbound Provisioning Type-Bug **Steps to reproduce** - Follow the instructions in the [documentation](https://docs.wso2.com/display/IS580/Outbound+Provisioning+with+Salesforce) **Actual Behavior** BadRequestException is thrown from the backend and the user is not created at Salesforce.
1.0
Outbound provisioning to Salesforce does not work as expected - **Suggested Labels** Affected: 5.8.0-alpha2 Severity: Major Priority: High Component: Salesforce Outbound Provisioning Type-Bug **Steps to reproduce** - Follow the instructions in the [documentation](https://docs.wso2.com/display/IS580/Outbound+Provisioning+with+Salesforce) **Actual Behavior** BadRequestException is thrown from the backend and the user is not created at Salesforce.
priority
outbound provisioning to salesforce does not work as expected suggested labels affected severity major priority high component salesforce outbound provisioning type bug steps to reproduce follow the instructions in the actual behavior badrequestexception is thrown from the backend and the user is not created at salesforce
1
439,462
12,683,080,874
IssuesEvent
2020-06-19 18:53:02
Helioviewer-Project/helioviewer.org
https://api.github.com/repos/Helioviewer-Project/helioviewer.org
closed
Many of the Images Are Darker and Darker in Comparison with Previous Releases and with the SDO Page
3.3 movies priority: high reported by user screenshots
_From @live4right on July 30, 2016 1:28_ Hi, I really enjoy using the jhelioviewer for monitoring the sun. I have actually determined there is a connection between the solar flares and severe weather events... Anyway I noticed that with each progressive release the images are becoming darker and darker. While on the SDO website they are as bright as ever. I have tried working with the various gamma settings and such but to no avail. What has happened to the coloring? I will include 2 screen shots same settings with jhelioviewer just 6 years apart. Maybe you can shed some light on this as I would like the coloring of 2010 back in the current version Thanks, Carl ![sdo 304 2010 image](https://cloud.githubusercontent.com/assets/13044191/17267257/35fb9f2c-55ca-11e6-975e-64c0736c7ecd.png) ![sdo 304 2016 image](https://cloud.githubusercontent.com/assets/13044191/17267256/35c80856-55ca-11e6-928a-ccdb66bf7ae1.png) _Copied from original issue: Helioviewer-Project/JHelioviewer#290_
1.0
Many of the Images Are Darker and Darker in Comparison with Previous Releases and with the SDO Page - _From @live4right on July 30, 2016 1:28_ Hi, I really enjoy using the jhelioviewer for monitoring the sun. I have actually determined there is a connection between the solar flares and severe weather events... Anyway I noticed that with each progressive release the images are becoming darker and darker. While on the SDO website they are as bright as ever. I have tried working with the various gamma settings and such but to no avail. What has happened to the coloring? I will include 2 screen shots same settings with jhelioviewer just 6 years apart. Maybe you can shed some light on this as I would like the coloring of 2010 back in the current version Thanks, Carl ![sdo 304 2010 image](https://cloud.githubusercontent.com/assets/13044191/17267257/35fb9f2c-55ca-11e6-975e-64c0736c7ecd.png) ![sdo 304 2016 image](https://cloud.githubusercontent.com/assets/13044191/17267256/35c80856-55ca-11e6-928a-ccdb66bf7ae1.png) _Copied from original issue: Helioviewer-Project/JHelioviewer#290_
priority
many of the images are darker and darker in comparison with previous releases and with the sdo page from on july hi i really enjoy using the jhelioviewer for monitoring the sun i have actually determined there is a connection between the solar flares and severe weather events anyway i noticed that with each progressive release the images are becoming darker and darker while on the sdo website they are as bright as ever i have tried working with the various gamma settings and such but to no avail what has happened to the coloring i will include screen shots same settings with jhelioviewer just years apart maybe you can shed some light on this as i would like the coloring of back in the current version thanks carl copied from original issue helioviewer project jhelioviewer
1
311,498
9,534,133,545
IssuesEvent
2019-04-29 23:57:57
hackla-engage/engage-backend
https://api.github.com/repos/hackla-engage/engage-backend
opened
PDF Report enhancement - clickable ToC
Priority: High Status: Help Wanted Type: Enhancement
Per hackla-engage/Product-Documentation#3, city would like to be able to hyper to the feedback for each agenda item. A clickable Table of Contents on the 2nd page of the report should address this request.
1.0
PDF Report enhancement - clickable ToC - Per hackla-engage/Product-Documentation#3, city would like to be able to hyper to the feedback for each agenda item. A clickable Table of Contents on the 2nd page of the report should address this request.
priority
pdf report enhancement clickable toc per hackla engage product documentation city would like to be able to hyper to the feedback for each agenda item a clickable table of contents on the page of the report should address this request
1
663,264
22,171,631,093
IssuesEvent
2022-06-06 01:50:01
ctm/mb2-doc
https://api.github.com/repos/ctm/mb2-doc
closed
UI hang after final cards shown on final hand
bug high priority easy
Fix whatever caused the UI to not update at the very end of yesterday's evening PRP tournament. > kidZee: [&hellip;] the UI did hang once the final hole cards were shown. This may be related to #947, but I haven't even looked at the timestamps yet since I need to get out on the trail at daylight.
1.0
UI hang after final cards shown on final hand - Fix whatever caused the UI to not update at the very end of yesterday's evening PRP tournament. > kidZee: [&hellip;] the UI did hang once the final hole cards were shown. This may be related to #947, but I haven't even looked at the timestamps yet since I need to get out on the trail at daylight.
priority
ui hang after final cards shown on final hand fix whatever caused the ui to not update at the very end of yesterday s evening prp tournament kidzee the ui did hang once the final hole cards were shown this may be related to but i haven t even looked at the timestamps yet since i need to get out on the trail at daylight
1
122,105
4,827,520,196
IssuesEvent
2016-11-07 13:53:33
CS2103AUG2016-F10-C2/main
https://api.github.com/repos/CS2103AUG2016-F10-C2/main
closed
Things to complete for V0.5 submission
priority.high
- [x] Create a xml file with sample data at \src\test\data\ManualTesting\SampleData.xml File should contain at least 50+ entries - [x] Create a .md file outlining the steps to test the program at src\test\data\ManualTesting\TestScript.md The test should contain expected results (screenshots not needed) and the whole test should not exceed 20 minutes. The test should cover all functionality of the app. - [x] Submit source code by tagging v0.5 on github. Make sure all collated codes are up to date because they cannot be updated after the deadline - [x] Upload the fat jar file to IVLE. File name: [team][project name].jar e.g. [T09-C1][ToDoo Pro].jar - [x] Update documentation on Github. - [x] Copy and paste contents of documentation into a .docx file following the format. [[TeamID][ProjectName].docx](https://github.com/CS2103AUG2016-F10-C2/main/files/573190/TeamID.ProjectName.docx) - [x] Convert the said .docx into pdf with name [TEAM_ID][Project Name].pdf e.g. [T09-C1][ToDoo Pro].pdf and upload it to IVLE #### Deadline: Monday 2359
1.0
Things to complete for V0.5 submission - - [x] Create a xml file with sample data at \src\test\data\ManualTesting\SampleData.xml File should contain at least 50+ entries - [x] Create a .md file outlining the steps to test the program at src\test\data\ManualTesting\TestScript.md The test should contain expected results (screenshots not needed) and the whole test should not exceed 20 minutes. The test should cover all functionality of the app. - [x] Submit source code by tagging v0.5 on github. Make sure all collated codes are up to date because they cannot be updated after the deadline - [x] Upload the fat jar file to IVLE. File name: [team][project name].jar e.g. [T09-C1][ToDoo Pro].jar - [x] Update documentation on Github. - [x] Copy and paste contents of documentation into a .docx file following the format. [[TeamID][ProjectName].docx](https://github.com/CS2103AUG2016-F10-C2/main/files/573190/TeamID.ProjectName.docx) - [x] Convert the said .docx into pdf with name [TEAM_ID][Project Name].pdf e.g. [T09-C1][ToDoo Pro].pdf and upload it to IVLE #### Deadline: Monday 2359
priority
things to complete for submission create a xml file with sample data at src test data manualtesting sampledata xml file should contain at least entries create a md file outlining the steps to test the program at src test data manualtesting testscript md the test should contain expected results screenshots not needed and the whole test should not exceed minutes the test should cover all functionality of the app submit source code by tagging on github make sure all collated codes are up to date because they cannot be updated after the deadline upload the fat jar file to ivle file name jar e g jar update documentation on github copy and paste contents of documentation into a docx file following the format docx convert the said docx into pdf with name pdf e g pdf and upload it to ivle deadline monday
1
166,497
6,305,814,709
IssuesEvent
2017-07-21 19:19:01
DashboardHub/PipelineDashboard
https://api.github.com/repos/DashboardHub/PipelineDashboard
closed
Auto deploy to dashboardhub.io (travis -> release branches)
BLOCKED help wanted priority: high
Using custom deploy on travis get this error when adding `pem` file: ``` $ ssh-add .travis/deploy.pem Could not open a connection to your authentication agent. ``` Posted in Gitter Travis chat room for support.
1.0
Auto deploy to dashboardhub.io (travis -> release branches) - Using custom deploy on travis get this error when adding `pem` file: ``` $ ssh-add .travis/deploy.pem Could not open a connection to your authentication agent. ``` Posted in Gitter Travis chat room for support.
priority
auto deploy to dashboardhub io travis release branches using custom deploy on travis get this error when adding pem file ssh add travis deploy pem could not open a connection to your authentication agent posted in gitter travis chat room for support
1
242,765
7,846,620,978
IssuesEvent
2018-06-19 15:57:42
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
opened
[studio] dependency resolution error causes create site to fail
CI bug priority: highest!
### Expected behavior Malformed paths should generate warnings/errors in the log and warnings to the user but should not fail the action. ### Actual behavior Encountering bad content such as `/static-assets/documents/someproduct/*` bubbles up an exception and terminates the user's actions. In this case, a create site. ### Steps to reproduce the problem working on a simple test case ### Log/stack trace (use https://gist.github.com) https://gist.github.com/russdanner/382c88b03650d19d7710cdb075b7e0da ### Specs #### Version Studio Version Number: 3.0.14-SNAPSHOT-4838ba Build Number: 4838bae0b3586ea0edabb4e8d57e3b1019f4e04e Build Date/Time: 06-14-2018 14:17:47 -0400 #### OS Any #### Browser N/A
1.0
[studio] dependency resolution error causes create site to fail - ### Expected behavior Malformed paths should generate warnings/errors in the log and warnings to the user but should not fail the action. ### Actual behavior Encountering bad content such as `/static-assets/documents/someproduct/*` bubbles up an exception and terminates the user's actions. In this case, a create site. ### Steps to reproduce the problem working on a simple test case ### Log/stack trace (use https://gist.github.com) https://gist.github.com/russdanner/382c88b03650d19d7710cdb075b7e0da ### Specs #### Version Studio Version Number: 3.0.14-SNAPSHOT-4838ba Build Number: 4838bae0b3586ea0edabb4e8d57e3b1019f4e04e Build Date/Time: 06-14-2018 14:17:47 -0400 #### OS Any #### Browser N/A
priority
dependency resolution error causes create site to fail expected behavior malformed paths should generate warnings errors in the log and warnings to the user but should not fail the action actual behavior encountering bad content such as static assets documents someproduct bubbles up an exception and terminates the user s actions in this case a create site steps to reproduce the problem working on a simple test case log stack trace use specs version studio version number snapshot build number build date time os any browser n a
1
820,446
30,772,403,235
IssuesEvent
2023-07-31 01:44:12
steedos/steedos-platform
https://api.github.com/repos/steedos/steedos-platform
closed
[Bug]: 首页刷新后报错
bug done priority: High
### Description ![image](https://github.com/steedos/steedos-platform/assets/15027092/9e4c457f-0bf5-4e28-9dab-917b8c2269f6) ### Steps To Reproduce 重现步骤 更新到2.5最新的正式版,进首页后报错 ### Version 版本 2.5.9
1.0
[Bug]: 首页刷新后报错 - ### Description ![image](https://github.com/steedos/steedos-platform/assets/15027092/9e4c457f-0bf5-4e28-9dab-917b8c2269f6) ### Steps To Reproduce 重现步骤 更新到2.5最新的正式版,进首页后报错 ### Version 版本 2.5.9
priority
首页刷新后报错 description steps to reproduce 重现步骤 ,进首页后报错 version 版本
1
782,591
27,500,569,922
IssuesEvent
2023-03-05 16:50:28
Ellivers/WorldTool
https://api.github.com/repos/Ellivers/WorldTool
closed
Undoing is inconsistent
bug priority: high
Undoing changes seems to be very inconsistent in whether it works or not, and which changes it undoes. This will probably require a lot of testing...
1.0
Undoing is inconsistent - Undoing changes seems to be very inconsistent in whether it works or not, and which changes it undoes. This will probably require a lot of testing...
priority
undoing is inconsistent undoing changes seems to be very inconsistent in whether it works or not and which changes it undoes this will probably require a lot of testing
1
45,438
2,931,442,601
IssuesEvent
2015-06-29 12:34:41
ufal/lindat-dspace
https://api.github.com/repos/ufal/lindat-dspace
opened
Download of zip archive of data larger than 1073741824B is forbidden.
high priority
Its a strange error, I tried to access all the items during the timeframe of the exception, and I am unable to reproduce it. We need to investigate more about this. org.apache.cocoon.ProcessingException: Download of zip archive of data larger than 1073741824B is forbidden. at <map:read type="AllBitstreamZipArchiveReader"> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:334:77 at <map:select type="AuthenticatedSelector"> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:332:46 at <map:match> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:331:64 at cz.cuni.mff.ufal.AllBitstreamZipArchiveReader.setup(AllBitstreamZipArchiveReader.java:187) at sun.reflect.GeneratedMethodAccessor69.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.apache.cocoon.core.container.spring.avalon.PoolableProxyHandler.invoke(PoolableProxyHandler.java:71) at com.sun.proxy.$Proxy51.setup(Unknown Source) at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.setupReader(AbstractProcessingPipeline.java:560) at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.preparePipeline(AbstractProcessingPipeline.java:464) at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.process(AbstractProcessingPipeline.java:411) at sun.reflect.GeneratedMethodAccessor195.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.apache.cocoon.core.container.spring.avalon.PoolableProxyHandler.invoke(PoolableProxyHandler.java:71) at com.sun.proxy.$Proxy29.process(Unknown Source) at org.apache.cocoon.components.treeprocessor.sitemap.ReadNode.invoke(ReadNode.java:94) at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78) at org.apache.cocoon.components.treeprocessor.sitemap.SelectNode.invoke(SelectNode.java:82) at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:55) at org.apache.cocoon.components.treeprocessor.sitemap.MatchNode.invoke(MatchNode.java:87) at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78) at org.apache.cocoon.components.treeprocessor.sitemap.PipelineNode.invoke(PipelineNode.java:143) at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78) at org.apache.cocoon.components.treeprocessor.sitemap.PipelinesNode.invoke(PipelinesNode.java:81) at org.apache.cocoon.components.treeprocessor.ConcreteTreeProcessor.process(ConcreteTreeProcessor.java:239) at org.apache.cocoon.components.treeprocessor.ConcreteTreeProcessor.process(ConcreteTreeProcessor.java:171) at org.apache.cocoon.components.treeprocessor.TreeProcessor.process(TreeProcessor.java:247) at org.apache.cocoon.servlet.RequestProcessor.process(RequestProcessor.java:351) at org.apache.cocoon.servlet.RequestProcessor.service(RequestProcessor.java:169) at org.apache.cocoon.sitemap.SitemapServlet.service(SitemapServlet.java:84) at javax.servlet.http.HttpServlet.service(HttpServlet.java:729) at org.apache.cocoon.servletservice.ServletServiceContext$PathDispatcher.forward(ServletServiceContext.java:468) at org.apache.cocoon.servletservice.ServletServiceContext$PathDispatcher.forward(ServletServiceContext.java:443) at org.apache.cocoon.servletservice.spring.ServletFactoryBean$ServiceInterceptor.invoke(ServletFactoryBean.java:264) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172) at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:202) at com.sun.proxy.$Proxy20.service(Unknown Source) at org.dspace.springmvc.CocoonView.render(CocoonView.java:117) at org.springframework.web.servlet.DispatcherServlet.render(DispatcherServlet.java:1180) at org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:950) at org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:852) at org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:882) at org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:778) at javax.servlet.http.HttpServlet.service(HttpServlet.java:622) at javax.servlet.http.HttpServlet.service(HttpServlet.java:729) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:291) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.app.xmlui.cocoon.SetCharacterEncodingFilter.doFilter(SetCharacterEncodingFilter.java:111) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.app.xmlui.cocoon.DSpaceCocoonServletFilter.doFilter(DSpaceCocoonServletFilter.java:276) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.app.xmlui.cocoon.servlet.multipart.DSpaceMultipartFilter.doFilter(DSpaceMultipartFilter.java:119) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.rdf.negotiation.NegotiationFilter.doFilter(NegotiationFilter.java:50) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.utils.servlet.DSpaceWebappServletFilter.doFilter(DSpaceWebappServletFilter.java:78) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:219) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:142) at com.googlecode.psiprobe.Tomcat80AgentValve.invoke(Tomcat80AgentValve.java:36) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:610) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) at org.apache.coyote.ajp.AbstractAjpProcessor.process(AbstractAjpProcessor.java:844) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1517) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1474) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:745)
1.0
Download of zip archive of data larger than 1073741824B is forbidden. - Its a strange error, I tried to access all the items during the timeframe of the exception, and I am unable to reproduce it. We need to investigate more about this. org.apache.cocoon.ProcessingException: Download of zip archive of data larger than 1073741824B is forbidden. at <map:read type="AllBitstreamZipArchiveReader"> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:334:77 at <map:select type="AuthenticatedSelector"> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:332:46 at <map:match> - file:///mnt/DATA/home/dspace/installations/dspace/lindat-dspace/webapps/xmlui/sitemap.xmap:331:64 at cz.cuni.mff.ufal.AllBitstreamZipArchiveReader.setup(AllBitstreamZipArchiveReader.java:187) at sun.reflect.GeneratedMethodAccessor69.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.apache.cocoon.core.container.spring.avalon.PoolableProxyHandler.invoke(PoolableProxyHandler.java:71) at com.sun.proxy.$Proxy51.setup(Unknown Source) at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.setupReader(AbstractProcessingPipeline.java:560) at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.preparePipeline(AbstractProcessingPipeline.java:464) at org.apache.cocoon.components.pipeline.AbstractProcessingPipeline.process(AbstractProcessingPipeline.java:411) at sun.reflect.GeneratedMethodAccessor195.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.apache.cocoon.core.container.spring.avalon.PoolableProxyHandler.invoke(PoolableProxyHandler.java:71) at com.sun.proxy.$Proxy29.process(Unknown Source) at org.apache.cocoon.components.treeprocessor.sitemap.ReadNode.invoke(ReadNode.java:94) at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78) at org.apache.cocoon.components.treeprocessor.sitemap.SelectNode.invoke(SelectNode.java:82) at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:55) at org.apache.cocoon.components.treeprocessor.sitemap.MatchNode.invoke(MatchNode.java:87) at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78) at org.apache.cocoon.components.treeprocessor.sitemap.PipelineNode.invoke(PipelineNode.java:143) at org.apache.cocoon.components.treeprocessor.AbstractParentProcessingNode.invokeNodes(AbstractParentProcessingNode.java:78) at org.apache.cocoon.components.treeprocessor.sitemap.PipelinesNode.invoke(PipelinesNode.java:81) at org.apache.cocoon.components.treeprocessor.ConcreteTreeProcessor.process(ConcreteTreeProcessor.java:239) at org.apache.cocoon.components.treeprocessor.ConcreteTreeProcessor.process(ConcreteTreeProcessor.java:171) at org.apache.cocoon.components.treeprocessor.TreeProcessor.process(TreeProcessor.java:247) at org.apache.cocoon.servlet.RequestProcessor.process(RequestProcessor.java:351) at org.apache.cocoon.servlet.RequestProcessor.service(RequestProcessor.java:169) at org.apache.cocoon.sitemap.SitemapServlet.service(SitemapServlet.java:84) at javax.servlet.http.HttpServlet.service(HttpServlet.java:729) at org.apache.cocoon.servletservice.ServletServiceContext$PathDispatcher.forward(ServletServiceContext.java:468) at org.apache.cocoon.servletservice.ServletServiceContext$PathDispatcher.forward(ServletServiceContext.java:443) at org.apache.cocoon.servletservice.spring.ServletFactoryBean$ServiceInterceptor.invoke(ServletFactoryBean.java:264) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172) at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:202) at com.sun.proxy.$Proxy20.service(Unknown Source) at org.dspace.springmvc.CocoonView.render(CocoonView.java:117) at org.springframework.web.servlet.DispatcherServlet.render(DispatcherServlet.java:1180) at org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:950) at org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:852) at org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:882) at org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:778) at javax.servlet.http.HttpServlet.service(HttpServlet.java:622) at javax.servlet.http.HttpServlet.service(HttpServlet.java:729) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:291) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.app.xmlui.cocoon.SetCharacterEncodingFilter.doFilter(SetCharacterEncodingFilter.java:111) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.app.xmlui.cocoon.DSpaceCocoonServletFilter.doFilter(DSpaceCocoonServletFilter.java:276) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.app.xmlui.cocoon.servlet.multipart.DSpaceMultipartFilter.doFilter(DSpaceMultipartFilter.java:119) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.rdf.negotiation.NegotiationFilter.doFilter(NegotiationFilter.java:50) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.dspace.utils.servlet.DSpaceWebappServletFilter.doFilter(DSpaceWebappServletFilter.java:78) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:239) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:219) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:106) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:142) at com.googlecode.psiprobe.Tomcat80AgentValve.invoke(Tomcat80AgentValve.java:36) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:610) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:518) at org.apache.coyote.ajp.AbstractAjpProcessor.process(AbstractAjpProcessor.java:844) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:668) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1517) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1474) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:745)
priority
download of zip archive of data larger than is forbidden its a strange error i tried to access all the items during the timeframe of the exception and i am unable to reproduce it we need to investigate more about this org apache cocoon processingexception download of zip archive of data larger than is forbidden at file mnt data home dspace installations dspace lindat dspace webapps xmlui sitemap xmap at file mnt data home dspace installations dspace lindat dspace webapps xmlui sitemap xmap at file mnt data home dspace installations dspace lindat dspace webapps xmlui sitemap xmap at cz cuni mff ufal allbitstreamziparchivereader setup allbitstreamziparchivereader java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache cocoon core container spring avalon poolableproxyhandler invoke poolableproxyhandler java at com sun proxy setup unknown source at org apache cocoon components pipeline abstractprocessingpipeline setupreader abstractprocessingpipeline java at org apache cocoon components pipeline abstractprocessingpipeline preparepipeline abstractprocessingpipeline java at org apache cocoon components pipeline abstractprocessingpipeline process abstractprocessingpipeline java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache cocoon core container spring avalon poolableproxyhandler invoke poolableproxyhandler java at com sun proxy process unknown source at org apache cocoon components treeprocessor sitemap readnode invoke readnode java at org apache cocoon components treeprocessor abstractparentprocessingnode invokenodes abstractparentprocessingnode java at org apache cocoon components treeprocessor sitemap selectnode invoke selectnode java at org apache cocoon components treeprocessor abstractparentprocessingnode invokenodes abstractparentprocessingnode java at org apache cocoon components treeprocessor sitemap matchnode invoke matchnode java at org apache cocoon components treeprocessor abstractparentprocessingnode invokenodes abstractparentprocessingnode java at org apache cocoon components treeprocessor sitemap pipelinenode invoke pipelinenode java at org apache cocoon components treeprocessor abstractparentprocessingnode invokenodes abstractparentprocessingnode java at org apache cocoon components treeprocessor sitemap pipelinesnode invoke pipelinesnode java at org apache cocoon components treeprocessor concretetreeprocessor process concretetreeprocessor java at org apache cocoon components treeprocessor concretetreeprocessor process concretetreeprocessor java at org apache cocoon components treeprocessor treeprocessor process treeprocessor java at org apache cocoon servlet requestprocessor process requestprocessor java at org apache cocoon servlet requestprocessor service requestprocessor java at org apache cocoon sitemap sitemapservlet service sitemapservlet java at javax servlet http httpservlet service httpservlet java at org apache cocoon servletservice servletservicecontext pathdispatcher forward servletservicecontext java at org apache cocoon servletservice servletservicecontext pathdispatcher forward servletservicecontext java at org apache cocoon servletservice spring servletfactorybean serviceinterceptor invoke servletfactorybean java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework aop framework jdkdynamicaopproxy invoke jdkdynamicaopproxy java at com sun proxy service unknown source at org dspace springmvc cocoonview render cocoonview java at org springframework web servlet dispatcherservlet render dispatcherservlet java at org springframework web servlet dispatcherservlet dodispatch dispatcherservlet java at org springframework web servlet dispatcherservlet doservice dispatcherservlet java at org springframework web servlet frameworkservlet processrequest frameworkservlet java at org springframework web servlet frameworkservlet doget frameworkservlet java at javax servlet http httpservlet service httpservlet java at javax servlet http httpservlet service httpservlet java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace app xmlui cocoon setcharacterencodingfilter dofilter setcharacterencodingfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace app xmlui cocoon dspacecocoonservletfilter dofilter dspacecocoonservletfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace app xmlui cocoon servlet multipart dspacemultipartfilter dofilter dspacemultipartfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace rdf negotiation negotiationfilter dofilter negotiationfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org dspace utils servlet dspacewebappservletfilter dofilter dspacewebappservletfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at com googlecode psiprobe invoke java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote ajp abstractajpprocessor process abstractajpprocessor java at org apache coyote abstractprotocol abstractconnectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net nioendpoint socketprocessor run nioendpoint java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java
1
137,746
5,315,858,848
IssuesEvent
2017-02-13 18:23:53
Valhalla-Gaming/Tracker
https://api.github.com/repos/Valhalla-Gaming/Tracker
closed
Druid talent Warrior of Elune.
Class-Druid Priority-High Type-Spell
**Describe the issue you're having**: After using Warrior of Elune (talent) it gives you 2 Lunar Strikes without casting it, but instead you can cast it unlimited times. **Explain how you expect it work**: You should get only 2 Lunar Strikes without casting This also works with mage's talent which let's you use Arcane Blast 2 times without casting it. **Steps to reproduce the problem**: 1. Learn the talent 2. Use it 3. Spam Lunar Strike **Links to Wowhead, YouTube, etc**: http://www.wowhead.com/spell=202425/warrior-of-elune
1.0
Druid talent Warrior of Elune. - **Describe the issue you're having**: After using Warrior of Elune (talent) it gives you 2 Lunar Strikes without casting it, but instead you can cast it unlimited times. **Explain how you expect it work**: You should get only 2 Lunar Strikes without casting This also works with mage's talent which let's you use Arcane Blast 2 times without casting it. **Steps to reproduce the problem**: 1. Learn the talent 2. Use it 3. Spam Lunar Strike **Links to Wowhead, YouTube, etc**: http://www.wowhead.com/spell=202425/warrior-of-elune
priority
druid talent warrior of elune describe the issue you re having after using warrior of elune talent it gives you lunar strikes without casting it but instead you can cast it unlimited times explain how you expect it work you should get only lunar strikes without casting this also works with mage s talent which let s you use arcane blast times without casting it steps to reproduce the problem learn the talent use it spam lunar strike links to wowhead youtube etc
1
152,391
5,846,278,406
IssuesEvent
2017-05-10 15:50:56
radcampaign/pulitzer2016
https://api.github.com/repos/radcampaign/pulitzer2016
closed
event - image caption missing
needs work priority HIGH
Not a huge rush, but something to be sure to fix. You can see that captions on old events are still appearing, but the field on the backend to enter/edit is missing: http://pulitzercenter.org/event/gallery-opening-god-government-lgbt-rights-russia-india-uganda
1.0
event - image caption missing - Not a huge rush, but something to be sure to fix. You can see that captions on old events are still appearing, but the field on the backend to enter/edit is missing: http://pulitzercenter.org/event/gallery-opening-god-government-lgbt-rights-russia-india-uganda
priority
event image caption missing not a huge rush but something to be sure to fix you can see that captions on old events are still appearing but the field on the backend to enter edit is missing
1
579,016
17,170,446,623
IssuesEvent
2021-07-15 03:02:49
TeamDooRiBon/DooRi-iOS
https://api.github.com/repos/TeamDooRiBon/DooRi-iOS
closed
[FEAT] 여행 보드 조회 API 연결
Feat P1 / Priority High Taehyeon 🐮
# 👀 이슈 (issue) 여행 보드 조회하는 API를 연결합니다. # 🚀 to-do <!-- 진행할 작업에 대해 적어주세요 --> <img width="250" alt="스크린샷 2021-07-15 오전 3 37 25" src="https://user-images.githubusercontent.com/61109660/125675169-6efbc15d-d20a-42b7-9cd5-afe45f25c4a3.png"> - [ ] Model 생성 - [ ] Service 구현 - [ ] API 연결 및 테스트
1.0
[FEAT] 여행 보드 조회 API 연결 - # 👀 이슈 (issue) 여행 보드 조회하는 API를 연결합니다. # 🚀 to-do <!-- 진행할 작업에 대해 적어주세요 --> <img width="250" alt="스크린샷 2021-07-15 오전 3 37 25" src="https://user-images.githubusercontent.com/61109660/125675169-6efbc15d-d20a-42b7-9cd5-afe45f25c4a3.png"> - [ ] Model 생성 - [ ] Service 구현 - [ ] API 연결 및 테스트
priority
여행 보드 조회 api 연결 👀 이슈 issue 여행 보드 조회하는 api를 연결합니다 🚀 to do img width alt 스크린샷 오전 src model 생성 service 구현 api 연결 및 테스트
1
316,548
9,651,538,211
IssuesEvent
2019-05-18 08:49:20
usemoslinux/aprelendo
https://api.github.com/repos/usemoslinux/aprelendo
closed
Optimize static images to reduce their size
Priority: high Type: improvement
Losslessly compressing https://www.aprelendo.com/img/backgrounds/welcome-page-background.jpeg could saves 12.4KiB (3% reduction).
1.0
Optimize static images to reduce their size - Losslessly compressing https://www.aprelendo.com/img/backgrounds/welcome-page-background.jpeg could saves 12.4KiB (3% reduction).
priority
optimize static images to reduce their size losslessly compressing could saves reduction
1
210,792
7,194,737,769
IssuesEvent
2018-02-04 09:14:04
hpi-swt2/sport-portal
https://api.github.com/repos/hpi-swt2/sport-portal
closed
Notifications: events/matches
epic notifications po-review priority high team kickerz
As User I want to be informed about changes in my events to plan my matches accordingly. **Acceptance criteria** - general requirements see #301 - emails are automatically generated when triggered - the text of emails is predefined and includes placeholders for customization (see texts below) Triggers are: - a match where user participates is scheduled > Hallo _user name_, > dein Spiel gegen _oponent team_ in _event name_ findet am _datum_ statt. > Viel Erfolg beim Training! > _Link to the event_ - a match where user participates is cancelled or postponed > Hallo _user name_, > dein Spiel gegen _oponent team_ in _event name_ wurde (auf _datum_ verschoben / abgesagt). > Viel Erfolg beim Training! > _Link to the event_ - an event where user is registered is started/cancelled/finished > Hallo _user name_, > das Event _event name_ hat gestartet (ist abgesagt/ ist abgeschlossen). > Viel Erfolg beim Training! > _Link to the event_ **Important** Check issues #57 and #149 and communicate with other dev teams
1.0
Notifications: events/matches - As User I want to be informed about changes in my events to plan my matches accordingly. **Acceptance criteria** - general requirements see #301 - emails are automatically generated when triggered - the text of emails is predefined and includes placeholders for customization (see texts below) Triggers are: - a match where user participates is scheduled > Hallo _user name_, > dein Spiel gegen _oponent team_ in _event name_ findet am _datum_ statt. > Viel Erfolg beim Training! > _Link to the event_ - a match where user participates is cancelled or postponed > Hallo _user name_, > dein Spiel gegen _oponent team_ in _event name_ wurde (auf _datum_ verschoben / abgesagt). > Viel Erfolg beim Training! > _Link to the event_ - an event where user is registered is started/cancelled/finished > Hallo _user name_, > das Event _event name_ hat gestartet (ist abgesagt/ ist abgeschlossen). > Viel Erfolg beim Training! > _Link to the event_ **Important** Check issues #57 and #149 and communicate with other dev teams
priority
notifications events matches as user i want to be informed about changes in my events to plan my matches accordingly acceptance criteria general requirements see emails are automatically generated when triggered the text of emails is predefined and includes placeholders for customization see texts below triggers are a match where user participates is scheduled hallo user name dein spiel gegen oponent team in event name findet am datum statt viel erfolg beim training link to the event a match where user participates is cancelled or postponed hallo user name dein spiel gegen oponent team in event name wurde auf datum verschoben abgesagt viel erfolg beim training link to the event an event where user is registered is started cancelled finished hallo user name das event event name hat gestartet ist abgesagt ist abgeschlossen viel erfolg beim training link to the event important check issues and and communicate with other dev teams
1
343,001
10,324,312,113
IssuesEvent
2019-09-01 07:58:53
OpenSRP/opensrp-client-chw-anc
https://api.github.com/repos/OpenSRP/opensrp-client-chw-anc
closed
Accent characters are not rendering anywhere in the app UI
High Priority bug
Inside the forms, the accent characters are rendering, but anywhere else in the app UI (register page, family profile page, etc.) they are not. ![Screenshot_20190807-124324](https://user-images.githubusercontent.com/3488986/62513524-03e2b280-b825-11e9-89ad-0721ef1c4493.png)
1.0
Accent characters are not rendering anywhere in the app UI - Inside the forms, the accent characters are rendering, but anywhere else in the app UI (register page, family profile page, etc.) they are not. ![Screenshot_20190807-124324](https://user-images.githubusercontent.com/3488986/62513524-03e2b280-b825-11e9-89ad-0721ef1c4493.png)
priority
accent characters are not rendering anywhere in the app ui inside the forms the accent characters are rendering but anywhere else in the app ui register page family profile page etc they are not
1
178,072
6,598,684,518
IssuesEvent
2017-09-16 09:15:33
oSoc17/rideaway-frontend
https://api.github.com/repos/oSoc17/rideaway-frontend
opened
Lock top controls when calculating a route
bug high priority
Because the loading screen is on the map (feels more intuitive), we should lock the geocoder inputs because else we get weird behaviour, aka 2 routes being calculated at the same time.
1.0
Lock top controls when calculating a route - Because the loading screen is on the map (feels more intuitive), we should lock the geocoder inputs because else we get weird behaviour, aka 2 routes being calculated at the same time.
priority
lock top controls when calculating a route because the loading screen is on the map feels more intuitive we should lock the geocoder inputs because else we get weird behaviour aka routes being calculated at the same time
1
197,797
6,964,071,990
IssuesEvent
2017-12-08 20:01:56
uva-slp/meals
https://api.github.com/repos/uva-slp/meals
closed
Improve alerts page
customer-requests priority-high story-points-5 volunteer-database
Currently, the alerts page only shows what routes and jobs are always missing volunteers. It needs to be modified to show which routes and jobs need volunteers because they have a substitution request submitted.
1.0
Improve alerts page - Currently, the alerts page only shows what routes and jobs are always missing volunteers. It needs to be modified to show which routes and jobs need volunteers because they have a substitution request submitted.
priority
improve alerts page currently the alerts page only shows what routes and jobs are always missing volunteers it needs to be modified to show which routes and jobs need volunteers because they have a substitution request submitted
1
328,930
10,001,617,697
IssuesEvent
2019-07-12 16:03:25
ampproject/amp-wp
https://api.github.com/repos/ampproject/amp-wp
closed
Choosing pre-defined font size causing error
AMP Stories Sprint 10 [Priority] High [Type] Bug
Steps to reproduce: - Create a text block. - Add some text. - Turn off the setting for amp-fit-text ("Automatically fit text to container"). - Switch the font size to one of the pre-defined sizes, e.g. to “Large” <img width="1374" alt="Screenshot 2019-07-11 at 13 26 47" src="https://user-images.githubusercontent.com/3294597/61049084-df411980-a3e3-11e9-8f53-b4d2e0c46989.png"> AC: User should be able to set a custom font size and choose between different font sizes.
1.0
Choosing pre-defined font size causing error - Steps to reproduce: - Create a text block. - Add some text. - Turn off the setting for amp-fit-text ("Automatically fit text to container"). - Switch the font size to one of the pre-defined sizes, e.g. to “Large” <img width="1374" alt="Screenshot 2019-07-11 at 13 26 47" src="https://user-images.githubusercontent.com/3294597/61049084-df411980-a3e3-11e9-8f53-b4d2e0c46989.png"> AC: User should be able to set a custom font size and choose between different font sizes.
priority
choosing pre defined font size causing error steps to reproduce create a text block add some text turn off the setting for amp fit text automatically fit text to container switch the font size to one of the pre defined sizes e g to “large” img width alt screenshot at src ac user should be able to set a custom font size and choose between different font sizes
1
689,470
23,621,834,462
IssuesEvent
2022-08-24 21:26:29
choderalab/perses
https://api.github.com/repos/choderalab/perses
opened
`RelativeFEPSetup` cannot deal with list of ligand input files
bug :bug: priority: high effort: low
In PR #1073 we included a way to deal with files from different cloud providers using `cloudpathlib`. But when doing this we generated a bug where if we pass a list of filenames for the `RelativeFEPSetup` object as the `ligand_input` parameter, we end up with an error like the following: ```python Traceback (most recent call last): File "/home/user/workdir/debugging/perses/perses-new-issue/test_RelativeFEPSetup_API.py", line 19, in <module> fe_setup = RelativeFEPSetup( File "/home/user/miniconda3/envs/perses-1092-omm-nightly-test/lib/python3.9/site-packages/perses/app/relative_setup.py", line 165, in __init__ self._ligand_input = AnyPath(ligand_input) File "/home/user/miniconda3/envs/perses-1092-omm-nightly-test/lib/python3.9/site-packages/cloudpathlib/anypath.py", line 38, in __new__ raise AnyPathTypeError( cloudpathlib.exceptions.AnyPathTypeError: Invalid input for both CloudPath and Path. CloudPath exception: InvalidPrefixError("Path ['./a1.sybyl.mol2', './a2.sybyl.mol2'] does not begin with a known prefix ['az://', 's3://', 'gs://'].") Path exception: TypeError('expected str, bytes or os.PathLike object, not list') (perses-1092-omm-nightly-test) ``` This is caused by trying to convert the list to an `cloudpathlib.AnyPath` object [here](https://github.com/choderalab/perses/blob/c96e65572b26cfdbce2d3ecaec7dae708e01568c/perses/app/relative_setup.py#L164-L165). A quick fix would be to just convert the given `input_ligand` (or even `self._input_ligand`) to an `AnyPath` object after we check it isn't a list in [line 242](https://github.com/choderalab/perses/blob/c96e65572b26cfdbce2d3ecaec7dae708e01568c/perses/app/relative_setup.py#L242). A longer term solution would be to refactor how we are dealing with this and have special methods for each case (list or str path) instead of having the huge if-else statements we currently have.
1.0
`RelativeFEPSetup` cannot deal with list of ligand input files - In PR #1073 we included a way to deal with files from different cloud providers using `cloudpathlib`. But when doing this we generated a bug where if we pass a list of filenames for the `RelativeFEPSetup` object as the `ligand_input` parameter, we end up with an error like the following: ```python Traceback (most recent call last): File "/home/user/workdir/debugging/perses/perses-new-issue/test_RelativeFEPSetup_API.py", line 19, in <module> fe_setup = RelativeFEPSetup( File "/home/user/miniconda3/envs/perses-1092-omm-nightly-test/lib/python3.9/site-packages/perses/app/relative_setup.py", line 165, in __init__ self._ligand_input = AnyPath(ligand_input) File "/home/user/miniconda3/envs/perses-1092-omm-nightly-test/lib/python3.9/site-packages/cloudpathlib/anypath.py", line 38, in __new__ raise AnyPathTypeError( cloudpathlib.exceptions.AnyPathTypeError: Invalid input for both CloudPath and Path. CloudPath exception: InvalidPrefixError("Path ['./a1.sybyl.mol2', './a2.sybyl.mol2'] does not begin with a known prefix ['az://', 's3://', 'gs://'].") Path exception: TypeError('expected str, bytes or os.PathLike object, not list') (perses-1092-omm-nightly-test) ``` This is caused by trying to convert the list to an `cloudpathlib.AnyPath` object [here](https://github.com/choderalab/perses/blob/c96e65572b26cfdbce2d3ecaec7dae708e01568c/perses/app/relative_setup.py#L164-L165). A quick fix would be to just convert the given `input_ligand` (or even `self._input_ligand`) to an `AnyPath` object after we check it isn't a list in [line 242](https://github.com/choderalab/perses/blob/c96e65572b26cfdbce2d3ecaec7dae708e01568c/perses/app/relative_setup.py#L242). A longer term solution would be to refactor how we are dealing with this and have special methods for each case (list or str path) instead of having the huge if-else statements we currently have.
priority
relativefepsetup cannot deal with list of ligand input files in pr we included a way to deal with files from different cloud providers using cloudpathlib but when doing this we generated a bug where if we pass a list of filenames for the relativefepsetup object as the ligand input parameter we end up with an error like the following python traceback most recent call last file home user workdir debugging perses perses new issue test relativefepsetup api py line in fe setup relativefepsetup file home user envs perses omm nightly test lib site packages perses app relative setup py line in init self ligand input anypath ligand input file home user envs perses omm nightly test lib site packages cloudpathlib anypath py line in new raise anypathtypeerror cloudpathlib exceptions anypathtypeerror invalid input for both cloudpath and path cloudpath exception invalidprefixerror path does not begin with a known prefix path exception typeerror expected str bytes or os pathlike object not list perses omm nightly test this is caused by trying to convert the list to an cloudpathlib anypath object a quick fix would be to just convert the given input ligand or even self input ligand to an anypath object after we check it isn t a list in a longer term solution would be to refactor how we are dealing with this and have special methods for each case list or str path instead of having the huge if else statements we currently have
1
822,586
30,878,991,082
IssuesEvent
2023-08-03 16:06:43
dhowe/AdNauseam
https://api.github.com/repos/dhowe/AdNauseam
closed
YouTube AdBlock ban bypass?
PRIORITY: High Ads Visible
Hello, could you guys please see if this can be replicated on your end? ![image](https://github.com/dhowe/AdNauseam/assets/32787189/edc41bf5-c64c-4f09-87eb-f46a38551784) Reddit Page: https://www.reddit.com/r/youtube/comments/13cfdbi/apparently_ad_blockers_are_not_allowed_on_youtube/ Thought i'd point this out to you guys to prepare Adnauseum when this more widely rolls out. Hopefully we'll continue blocking youtube ads.
1.0
YouTube AdBlock ban bypass? - Hello, could you guys please see if this can be replicated on your end? ![image](https://github.com/dhowe/AdNauseam/assets/32787189/edc41bf5-c64c-4f09-87eb-f46a38551784) Reddit Page: https://www.reddit.com/r/youtube/comments/13cfdbi/apparently_ad_blockers_are_not_allowed_on_youtube/ Thought i'd point this out to you guys to prepare Adnauseum when this more widely rolls out. Hopefully we'll continue blocking youtube ads.
priority
youtube adblock ban bypass hello could you guys please see if this can be replicated on your end reddit page thought i d point this out to you guys to prepare adnauseum when this more widely rolls out hopefully we ll continue blocking youtube ads
1
783,568
27,537,077,991
IssuesEvent
2023-03-07 04:54:05
AY2223S2-CS2113-W15-1/tp
https://api.github.com/repos/AY2223S2-CS2113-W15-1/tp
opened
Calculate Calories
priority.High type.Story
As a user, I can calculate my caloric needs based on my height and weight so that I will know how much calories I can take in, in a day
1.0
Calculate Calories - As a user, I can calculate my caloric needs based on my height and weight so that I will know how much calories I can take in, in a day
priority
calculate calories as a user i can calculate my caloric needs based on my height and weight so that i will know how much calories i can take in in a day
1
524,527
15,215,882,267
IssuesEvent
2021-02-17 14:54:33
bexis/Module_ResourceManagement
https://api.github.com/repos/bexis/Module_ResourceManagement
opened
Edit resource time : end-date changes unexpected
Priority: High bug
If you change the "from-date" the "to-date" by itself to "to-date"+1 day Should stay as it is.
1.0
Edit resource time : end-date changes unexpected - If you change the "from-date" the "to-date" by itself to "to-date"+1 day Should stay as it is.
priority
edit resource time end date changes unexpected if you change the from date the to date by itself to to date day should stay as it is
1
198,040
6,968,970,438
IssuesEvent
2017-12-11 01:49:30
madeline-bauer/backend-web-dev
https://api.github.com/repos/madeline-bauer/backend-web-dev
closed
Make authorization system
Priority: High Status: Accepted Type: Enhancement
@ccourtney121, I've made a file `authorization.js` that gets passed the request every time an API call is made. It should return `true` or `false` for if the request is authorized. Right now I've just hard-coded `return true`. When you make security happen, this is where it should happen. If it returns `false` then the request will get an `Error 401 (Unauthorized)` instead of the data they were after. I think you're planning on using session tokens? So it would be something like `if (req.query.token == /*is in authorized tokens db*/){return true;}`
1.0
Make authorization system - @ccourtney121, I've made a file `authorization.js` that gets passed the request every time an API call is made. It should return `true` or `false` for if the request is authorized. Right now I've just hard-coded `return true`. When you make security happen, this is where it should happen. If it returns `false` then the request will get an `Error 401 (Unauthorized)` instead of the data they were after. I think you're planning on using session tokens? So it would be something like `if (req.query.token == /*is in authorized tokens db*/){return true;}`
priority
make authorization system i ve made a file authorization js that gets passed the request every time an api call is made it should return true or false for if the request is authorized right now i ve just hard coded return true when you make security happen this is where it should happen if it returns false then the request will get an error unauthorized instead of the data they were after i think you re planning on using session tokens so it would be something like if req query token is in authorized tokens db return true
1
655,786
21,709,209,033
IssuesEvent
2022-05-10 12:31:30
wazuh/wazuh-documentation
https://api.github.com/repos/wazuh/wazuh-documentation
closed
Create Release Notes for Release 4.3.0
priority: high type: refactor
Hi team, This issue aims to create the Release Notes for 4.3.0 that will be published on the GA date. The commits of this issue can be seen here: https://github.com/wazuh/wazuh-documentation/commits/4.3-rc ### Stage 1: - [x] Create branch and template - [x] Check Changelogs and add issues from `wazuh/wazuh rc1`: - Highlights - [x] Manager - [x] Agent - [x] Wazuh Kibana plugin - [x] RESTful API - What’s new - [x] Manager - [x] Agent - [x] RESTful API - [x] Ruleset - [x] Wazuh Kibana plugin - [x] Others - Resolved issues - [x] Manager - [x] Agent - [x] RESTful API - [x] Ruleset - [x] Wazuh Kibana plugin - [x] Others - [x] Changelogs - [x] Edit the content and check that the new added issues are consistent - [x] Check correspondence between issue number and number in link - [x] Make necessary linguistic adjustments - [x] Check Changelogs and add issues from `wazuh/wazuh rc2` - [x] Check Changelogs and add issues from `wazuh-kibana-app/CHANGELOG.md` - [x] Check Changelogs and add issues from `wazuh-splunk/CHANGELOG.md` - [x] Check Changelogs and add issues from `wazuh/wazuh rc3` - [x] Check Changelogs and add issues from `wazuh/wazuh rc4` - [x] Check Changelogs and add issues from `wazuh/wazuh rc5` - [x] Check Changelogs and add issues from `wazuh/wazuh rc6` - [x] Check Changelogs and add issues from `wazuh/wazuh rc7` - [x] Check Changelogs and add issues from `wazuh-packages/CHANGELOG.md` - [x] Edit the content and check that the new added issues are consistent - [x] Check correspondence between issue number and number in link - [x] Make necessary linguistic adjustments - [x] Review from the content team ### Stage 2: - [x] Work on the branch `4.3.0-release-notes-update` - [x] Make a review, add updates and make fixes - [x] Review from the content team Further steps will be added if additional revisions are needed for new release candidates. Regards, Mariel and Damián
1.0
Create Release Notes for Release 4.3.0 - Hi team, This issue aims to create the Release Notes for 4.3.0 that will be published on the GA date. The commits of this issue can be seen here: https://github.com/wazuh/wazuh-documentation/commits/4.3-rc ### Stage 1: - [x] Create branch and template - [x] Check Changelogs and add issues from `wazuh/wazuh rc1`: - Highlights - [x] Manager - [x] Agent - [x] Wazuh Kibana plugin - [x] RESTful API - What’s new - [x] Manager - [x] Agent - [x] RESTful API - [x] Ruleset - [x] Wazuh Kibana plugin - [x] Others - Resolved issues - [x] Manager - [x] Agent - [x] RESTful API - [x] Ruleset - [x] Wazuh Kibana plugin - [x] Others - [x] Changelogs - [x] Edit the content and check that the new added issues are consistent - [x] Check correspondence between issue number and number in link - [x] Make necessary linguistic adjustments - [x] Check Changelogs and add issues from `wazuh/wazuh rc2` - [x] Check Changelogs and add issues from `wazuh-kibana-app/CHANGELOG.md` - [x] Check Changelogs and add issues from `wazuh-splunk/CHANGELOG.md` - [x] Check Changelogs and add issues from `wazuh/wazuh rc3` - [x] Check Changelogs and add issues from `wazuh/wazuh rc4` - [x] Check Changelogs and add issues from `wazuh/wazuh rc5` - [x] Check Changelogs and add issues from `wazuh/wazuh rc6` - [x] Check Changelogs and add issues from `wazuh/wazuh rc7` - [x] Check Changelogs and add issues from `wazuh-packages/CHANGELOG.md` - [x] Edit the content and check that the new added issues are consistent - [x] Check correspondence between issue number and number in link - [x] Make necessary linguistic adjustments - [x] Review from the content team ### Stage 2: - [x] Work on the branch `4.3.0-release-notes-update` - [x] Make a review, add updates and make fixes - [x] Review from the content team Further steps will be added if additional revisions are needed for new release candidates. Regards, Mariel and Damián
priority
create release notes for release hi team this issue aims to create the release notes for that will be published on the ga date the commits of this issue can be seen here stage create branch and template check changelogs and add issues from wazuh wazuh highlights manager agent wazuh kibana plugin restful api what’s new manager agent restful api ruleset wazuh kibana plugin others resolved issues manager agent restful api ruleset wazuh kibana plugin others changelogs edit the content and check that the new added issues are consistent check correspondence between issue number and number in link make necessary linguistic adjustments check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh kibana app changelog md check changelogs and add issues from wazuh splunk changelog md check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh wazuh check changelogs and add issues from wazuh packages changelog md edit the content and check that the new added issues are consistent check correspondence between issue number and number in link make necessary linguistic adjustments review from the content team stage work on the branch release notes update make a review add updates and make fixes review from the content team further steps will be added if additional revisions are needed for new release candidates regards mariel and damián
1
463,323
13,263,311,977
IssuesEvent
2020-08-21 00:08:53
radcampaign/allaboveall2020
https://api.github.com/repos/radcampaign/allaboveall2020
opened
[OG Image] Get working
Priority: High
I think this is still unsolved? Or I missed what's up on slack. Putting an image out there, and prioritizing it high, as they need to go through all their pages and make sure the right images are loading.
1.0
[OG Image] Get working - I think this is still unsolved? Or I missed what's up on slack. Putting an image out there, and prioritizing it high, as they need to go through all their pages and make sure the right images are loading.
priority
get working i think this is still unsolved or i missed what s up on slack putting an image out there and prioritizing it high as they need to go through all their pages and make sure the right images are loading
1
233,217
7,695,554,208
IssuesEvent
2018-05-18 12:44:19
bitshares/bitshares-ui
https://api.github.com/repos/bitshares/bitshares-ui
closed
[3][happyconcepts] Order Expiration Date bug
bug high priority
OSX, latest, Chrome Browser ## Problem It appears that the order expiration is off by the timezone difference. In This case, it is 7:18 pm. I chose a 12 hour expiration. The confirmation modal displays 12:18 AM CDT. ![houra](https://user-images.githubusercontent.com/1254810/38282202-8372d0c6-3774-11e8-93f4-1e306864a4f9.gif) ## Expected Behavior My current time + offset chosen = Time displayed on modal.
1.0
[3][happyconcepts] Order Expiration Date bug - OSX, latest, Chrome Browser ## Problem It appears that the order expiration is off by the timezone difference. In This case, it is 7:18 pm. I chose a 12 hour expiration. The confirmation modal displays 12:18 AM CDT. ![houra](https://user-images.githubusercontent.com/1254810/38282202-8372d0c6-3774-11e8-93f4-1e306864a4f9.gif) ## Expected Behavior My current time + offset chosen = Time displayed on modal.
priority
order expiration date bug osx latest chrome browser problem it appears that the order expiration is off by the timezone difference in this case it is pm i chose a hour expiration the confirmation modal displays am cdt expected behavior my current time offset chosen time displayed on modal
1
713,277
24,523,456,429
IssuesEvent
2022-10-11 11:21:18
opencrvs/opencrvs-core
https://api.github.com/repos/opencrvs/opencrvs-core
opened
OpenHIM Channel security improvements
Priority: high
**CORS is not setup in OpenHIM** It seems to be also possible to access OpenHIM and specifically the wildcard DNS exposes unused traefik routes "api." The workflow passthrough channel regexp essentially forwards any URL suffix to it. We need to strictly set the workflow passthrough regexp and update the backend where appropriate when the workflow passthrough is called. See screenshots ... ![Screenshot 2022-10-11 at 11.46.12.png](https://images.zenhubusercontent.com/5e96a2d4befb239e4f2bce95/b737a4ab-0c2c-4aa1-baf8-ac9386ceee6e) ![Screenshot 2022-10-11 at 11.45.56.png](https://images.zenhubusercontent.com/5e96a2d4befb239e4f2bce95/c5d61420-d852-4e56-bb8d-5063fc8d907e) ![Screenshot 2022-10-11 at 11.45.34.png](https://images.zenhubusercontent.com/5e96a2d4befb239e4f2bce95/5516cb8d-efcf-4d43-a8e9-4f3afbd4dada) Run full cypress suite when done
1.0
OpenHIM Channel security improvements - **CORS is not setup in OpenHIM** It seems to be also possible to access OpenHIM and specifically the wildcard DNS exposes unused traefik routes "api." The workflow passthrough channel regexp essentially forwards any URL suffix to it. We need to strictly set the workflow passthrough regexp and update the backend where appropriate when the workflow passthrough is called. See screenshots ... ![Screenshot 2022-10-11 at 11.46.12.png](https://images.zenhubusercontent.com/5e96a2d4befb239e4f2bce95/b737a4ab-0c2c-4aa1-baf8-ac9386ceee6e) ![Screenshot 2022-10-11 at 11.45.56.png](https://images.zenhubusercontent.com/5e96a2d4befb239e4f2bce95/c5d61420-d852-4e56-bb8d-5063fc8d907e) ![Screenshot 2022-10-11 at 11.45.34.png](https://images.zenhubusercontent.com/5e96a2d4befb239e4f2bce95/5516cb8d-efcf-4d43-a8e9-4f3afbd4dada) Run full cypress suite when done
priority
openhim channel security improvements cors is not setup in openhim it seems to be also possible to access openhim and specifically the wildcard dns exposes unused traefik routes api the workflow passthrough channel regexp essentially forwards any url suffix to it we need to strictly set the workflow passthrough regexp and update the backend where appropriate when the workflow passthrough is called see screenshots run full cypress suite when done
1
567,405
16,857,863,513
IssuesEvent
2021-06-21 09:09:36
PlaceOS/user-interfaces
https://api.github.com/repos/PlaceOS/user-interfaces
closed
Workplace template errors when booking a desk from explore view
Priority: High Type: Bug focus: front end
When attempting to book a desk from explore view, the following error is hit. https://github.com/PlaceOS/user-interfaces/blob/81f167ff1753f94903168b07c4e5bf0c0594cca5/libs/bookings/src/lib/desk.service.ts#L44 At no point during this flow is the ability to select a host provided to the user.
1.0
Workplace template errors when booking a desk from explore view - When attempting to book a desk from explore view, the following error is hit. https://github.com/PlaceOS/user-interfaces/blob/81f167ff1753f94903168b07c4e5bf0c0594cca5/libs/bookings/src/lib/desk.service.ts#L44 At no point during this flow is the ability to select a host provided to the user.
priority
workplace template errors when booking a desk from explore view when attempting to book a desk from explore view the following error is hit at no point during this flow is the ability to select a host provided to the user
1
584,510
17,456,798,026
IssuesEvent
2021-08-06 03:18:35
opsramp/openapi-specs
https://api.github.com/repos/opsramp/openapi-specs
closed
Confirm use of query string object for patching endpoint
bug high priority
The endpoint: `/api/v2/tenants/{tenantId}/patches` Starts here: https://github.com/opsramp/openapi-specs/blob/5ca943366fe3e717050f9ce0ea9df001339c5e76/v2/opsramp-patching.v2.yaml#L1861 Parameter in question: https://github.com/opsramp/openapi-specs/blob/5ca943366fe3e717050f9ce0ea9df001339c5e76/v2/opsramp-patching.v2.yaml#L1936 This seems to imply a request would look similar to: `/api/v2/tenants/{tenantId}/patches?pageNo=1&pageSize=100&queryString[name]=TheName&querystring[uid]=00101&queryString[startReleaseDate]=2018-08-19T08%3A30%3A20%2B0000` Can you confirm what a properly formatted request for this endpoint would look like?
1.0
Confirm use of query string object for patching endpoint - The endpoint: `/api/v2/tenants/{tenantId}/patches` Starts here: https://github.com/opsramp/openapi-specs/blob/5ca943366fe3e717050f9ce0ea9df001339c5e76/v2/opsramp-patching.v2.yaml#L1861 Parameter in question: https://github.com/opsramp/openapi-specs/blob/5ca943366fe3e717050f9ce0ea9df001339c5e76/v2/opsramp-patching.v2.yaml#L1936 This seems to imply a request would look similar to: `/api/v2/tenants/{tenantId}/patches?pageNo=1&pageSize=100&queryString[name]=TheName&querystring[uid]=00101&queryString[startReleaseDate]=2018-08-19T08%3A30%3A20%2B0000` Can you confirm what a properly formatted request for this endpoint would look like?
priority
confirm use of query string object for patching endpoint the endpoint api tenants tenantid patches starts here parameter in question this seems to imply a request would look similar to api tenants tenantid patches pageno pagesize querystring thename querystring querystring can you confirm what a properly formatted request for this endpoint would look like
1
4,009
2,544,669,644
IssuesEvent
2015-01-29 11:53:33
IntellectualCrafters/PlotSquared
https://api.github.com/repos/IntellectualCrafters/PlotSquared
closed
Plot protection system
enhancement high priority
``` Enhancement Another well known and quite similar plot management plugin offers in this configuration to define which blocks or items like buttons or levers are protected on plots and which are public accessible. I really liked this feature because for example you could make buttons made from wood usable for everyone meanwhile stone buttons can only be used from authorized players. So there could be public and private areas. Another example would be chests and trapped chests. If players want to have a shared chest with everyone on the server without having added each and every player to a plot there could be trapped chests unprotected and common chests protected. Maybe you wanna give this feature a chance. I would love to see that :). ``` ``` Enhancement Users which have no helper or trust status on a plot can not use quickshop chests to buy or sell items. Pex debug tells me that the plugin asks for plots.admin.interact.other which is obviously not present on common players without any special rights. Therefore QuickShops cannot be used in a Plotworld seriously. It would be very nice if this could be fixed. Maybe this could be combined with #94 ```
1.0
Plot protection system - ``` Enhancement Another well known and quite similar plot management plugin offers in this configuration to define which blocks or items like buttons or levers are protected on plots and which are public accessible. I really liked this feature because for example you could make buttons made from wood usable for everyone meanwhile stone buttons can only be used from authorized players. So there could be public and private areas. Another example would be chests and trapped chests. If players want to have a shared chest with everyone on the server without having added each and every player to a plot there could be trapped chests unprotected and common chests protected. Maybe you wanna give this feature a chance. I would love to see that :). ``` ``` Enhancement Users which have no helper or trust status on a plot can not use quickshop chests to buy or sell items. Pex debug tells me that the plugin asks for plots.admin.interact.other which is obviously not present on common players without any special rights. Therefore QuickShops cannot be used in a Plotworld seriously. It would be very nice if this could be fixed. Maybe this could be combined with #94 ```
priority
plot protection system enhancement another well known and quite similar plot management plugin offers in this configuration to define which blocks or items like buttons or levers are protected on plots and which are public accessible i really liked this feature because for example you could make buttons made from wood usable for everyone meanwhile stone buttons can only be used from authorized players so there could be public and private areas another example would be chests and trapped chests if players want to have a shared chest with everyone on the server without having added each and every player to a plot there could be trapped chests unprotected and common chests protected maybe you wanna give this feature a chance i would love to see that enhancement users which have no helper or trust status on a plot can not use quickshop chests to buy or sell items pex debug tells me that the plugin asks for plots admin interact other which is obviously not present on common players without any special rights therefore quickshops cannot be used in a plotworld seriously it would be very nice if this could be fixed maybe this could be combined with
1
648,744
21,193,020,857
IssuesEvent
2022-04-08 19:50:05
NOAA-GSL/MATS
https://api.github.com/repos/NOAA-GSL/MATS
closed
Fix Deploy action failure
Type: Bug Priority: High CI/CD
The deploy action associated with the v4.4.1 release failed due to a syntax error with `GITHUB_REF`. This was believed to have been fixed in #716. The failing job is here: https://github.com/NOAA-GSL/MATS/runs/5875625209?check_suite_focus=true According to the GHA docs, a release event should populate the `GITHUB_REF` with the tag: https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#release. Current thoughts are that the variable is quoted incorrectly and/or we need to trim the result from the variable as it returns like so: `refs/tags/<tag_name>`. This info also exists in the `github.ref` context so we could look at using that instead if it'd be advantageous. https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#release
1.0
Fix Deploy action failure - The deploy action associated with the v4.4.1 release failed due to a syntax error with `GITHUB_REF`. This was believed to have been fixed in #716. The failing job is here: https://github.com/NOAA-GSL/MATS/runs/5875625209?check_suite_focus=true According to the GHA docs, a release event should populate the `GITHUB_REF` with the tag: https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#release. Current thoughts are that the variable is quoted incorrectly and/or we need to trim the result from the variable as it returns like so: `refs/tags/<tag_name>`. This info also exists in the `github.ref` context so we could look at using that instead if it'd be advantageous. https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#release
priority
fix deploy action failure the deploy action associated with the release failed due to a syntax error with github ref this was believed to have been fixed in the failing job is here according to the gha docs a release event should populate the github ref with the tag current thoughts are that the variable is quoted incorrectly and or we need to trim the result from the variable as it returns like so refs tags this info also exists in the github ref context so we could look at using that instead if it d be advantageous
1
388,100
11,474,539,802
IssuesEvent
2020-02-10 04:44:25
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
Supplier Credits straight from a Supplier Invoice
Docs: not needed Effort: small Feature Module: dispensary Priority: high
## Is your feature request related to a problem? Please describe. Can create a supplier credit from the current stock page. Should be able to create one straight from an SI ## Describe the solution you'd like As above ## Implementation N/A ## Describe alternatives you've considered N/A ## Additional context N/A
1.0
Supplier Credits straight from a Supplier Invoice - ## Is your feature request related to a problem? Please describe. Can create a supplier credit from the current stock page. Should be able to create one straight from an SI ## Describe the solution you'd like As above ## Implementation N/A ## Describe alternatives you've considered N/A ## Additional context N/A
priority
supplier credits straight from a supplier invoice is your feature request related to a problem please describe can create a supplier credit from the current stock page should be able to create one straight from an si describe the solution you d like as above implementation n a describe alternatives you ve considered n a additional context n a
1
622,501
19,640,366,255
IssuesEvent
2022-01-09 00:02:24
SourceWriters/SmoothTimber
https://api.github.com/repos/SourceWriters/SmoothTimber
closed
[REVISION] Remove the permission handle of the SmoothTimber command from the plugin.yml
Done Priority: High Revision
**Is your revision request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Currently bukkit is handling the first permission check which resolves in the "You are not permitted to execute this command" permission message. However SmoothTimber itself does also a permission check before redirecting to any subcommand and if the permission isn't set it would send a custom permission message which it currently doesn't as it is set in the plugin.yml **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> Update the plugin.yml to not contain the permission node for the "smoothtimber" command **Additional context** Only the plugin.yml is changed Related to: #32
1.0
[REVISION] Remove the permission handle of the SmoothTimber command from the plugin.yml - **Is your revision request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Currently bukkit is handling the first permission check which resolves in the "You are not permitted to execute this command" permission message. However SmoothTimber itself does also a permission check before redirecting to any subcommand and if the permission isn't set it would send a custom permission message which it currently doesn't as it is set in the plugin.yml **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> Update the plugin.yml to not contain the permission node for the "smoothtimber" command **Additional context** Only the plugin.yml is changed Related to: #32
priority
remove the permission handle of the smoothtimber command from the plugin yml is your revision request related to a problem please describe currently bukkit is handling the first permission check which resolves in the you are not permitted to execute this command permission message however smoothtimber itself does also a permission check before redirecting to any subcommand and if the permission isn t set it would send a custom permission message which it currently doesn t as it is set in the plugin yml describe the solution you d like update the plugin yml to not contain the permission node for the smoothtimber command additional context only the plugin yml is changed related to
1
219,687
7,344,969,801
IssuesEvent
2018-03-07 16:06:23
prettier/prettier
https://api.github.com/repos/prettier/prettier
closed
In Markdown Files, Nested Code Block/Raw HTML Indent For Each Format on Save
lang:markdown priority:high status:has pr type:bug
I apologize if this is already an open ticket, I didn't see anything for this in particular. The issue is appearing in markdown files. When attempting to nest code blocks or raw HTML the format on save feature will add more spacing each time the file is saved which breaks the indentation. When using code blocks the issue is intermittent where once the code is nested correctly it will not break again but with raw HTML the only way to prevent the issues is to turn off format on save or ignore markdown files, neither of which is a solution. Any thoughts on a setting that may help are appreciated. p.s. I'm using Prettier in VS Code. ![formatonsave](https://user-images.githubusercontent.com/25302911/37053023-4a36a348-2138-11e8-9f2a-af057a9998e7.gif) **Input:** 1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync. <table class="table table-striped"> <tr> <th>Test</th> <th>Table</th> </tr> <tbody> <tr> <td>will</td> <td>be</td> </tr> <tr> <td>pushed</td> <td>When</td> </tr> <tr> <td>Format on</td> <td>Save</td> </tr> </tbody> </table> **Output on first save:** 1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync. <table class="table table-striped"> <tr> <th>Test</th> <th>Table</th> </tr> <tbody> <tr> <td>will</td> <td>be</td> </tr> <tr> <td>pushed</td> <td>When</td> </tr> <tr> <td>Format on</td> <td>Save</td> </tr> </tbody> </table> **Output on fourth save:** 1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync. <table class="table table-striped"> <tr> <th>Test</th> <th>Table</th> </tr> <tbody> <tr> <td>will</td> <td>be</td> </tr> <tr> <td>pushed</td> <td>When</td> </tr> <tr> <td>Format on</td> <td>Save</td> </tr> </tbody> </table> **Expected behavior:** 1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync. <table class="table table-striped"> <tr> <th>Test</th> <th>Table</th> </tr> <tbody> <tr> <td>will</td> <td>be</td> </tr> <tr> <td>pushed</td> <td>When</td> </tr> <tr> <td>Format on</td> <td>Save</td> </tr> </tbody> </table>
1.0
In Markdown Files, Nested Code Block/Raw HTML Indent For Each Format on Save - I apologize if this is already an open ticket, I didn't see anything for this in particular. The issue is appearing in markdown files. When attempting to nest code blocks or raw HTML the format on save feature will add more spacing each time the file is saved which breaks the indentation. When using code blocks the issue is intermittent where once the code is nested correctly it will not break again but with raw HTML the only way to prevent the issues is to turn off format on save or ignore markdown files, neither of which is a solution. Any thoughts on a setting that may help are appreciated. p.s. I'm using Prettier in VS Code. ![formatonsave](https://user-images.githubusercontent.com/25302911/37053023-4a36a348-2138-11e8-9f2a-af057a9998e7.gif) **Input:** 1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync. <table class="table table-striped"> <tr> <th>Test</th> <th>Table</th> </tr> <tbody> <tr> <td>will</td> <td>be</td> </tr> <tr> <td>pushed</td> <td>When</td> </tr> <tr> <td>Format on</td> <td>Save</td> </tr> </tbody> </table> **Output on first save:** 1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync. <table class="table table-striped"> <tr> <th>Test</th> <th>Table</th> </tr> <tbody> <tr> <td>will</td> <td>be</td> </tr> <tr> <td>pushed</td> <td>When</td> </tr> <tr> <td>Format on</td> <td>Save</td> </tr> </tbody> </table> **Output on fourth save:** 1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync. <table class="table table-striped"> <tr> <th>Test</th> <th>Table</th> </tr> <tbody> <tr> <td>will</td> <td>be</td> </tr> <tr> <td>pushed</td> <td>When</td> </tr> <tr> <td>Format on</td> <td>Save</td> </tr> </tbody> </table> **Expected behavior:** 1. Some test text, the goal is to have the HTML table below nested within this number. When formatting on save Prettier will continue to add an indent each time pushing the table further and further out of sync. <table class="table table-striped"> <tr> <th>Test</th> <th>Table</th> </tr> <tbody> <tr> <td>will</td> <td>be</td> </tr> <tr> <td>pushed</td> <td>When</td> </tr> <tr> <td>Format on</td> <td>Save</td> </tr> </tbody> </table>
priority
in markdown files nested code block raw html indent for each format on save i apologize if this is already an open ticket i didn t see anything for this in particular the issue is appearing in markdown files when attempting to nest code blocks or raw html the format on save feature will add more spacing each time the file is saved which breaks the indentation when using code blocks the issue is intermittent where once the code is nested correctly it will not break again but with raw html the only way to prevent the issues is to turn off format on save or ignore markdown files neither of which is a solution any thoughts on a setting that may help are appreciated p s i m using prettier in vs code input some test text the goal is to have the html table below nested within this number when formatting on save prettier will continue to add an indent each time pushing the table further and further out of sync test table will be pushed when format on save output on first save some test text the goal is to have the html table below nested within this number when formatting on save prettier will continue to add an indent each time pushing the table further and further out of sync test table will be pushed when format on save output on fourth save some test text the goal is to have the html table below nested within this number when formatting on save prettier will continue to add an indent each time pushing the table further and further out of sync test table will be pushed when format on save expected behavior some test text the goal is to have the html table below nested within this number when formatting on save prettier will continue to add an indent each time pushing the table further and further out of sync test table will be pushed when format on save
1
310,696
9,523,024,288
IssuesEvent
2019-04-27 13:57:38
NOVACProject/NOVACProgram
https://api.github.com/repos/NOVACProject/NOVACProgram
closed
Add Support for SFTP
enhancement high priority
The new upload server at Chalmers will use SFTP protocol once implemented. This change is required before we can switch to the new server.
1.0
Add Support for SFTP - The new upload server at Chalmers will use SFTP protocol once implemented. This change is required before we can switch to the new server.
priority
add support for sftp the new upload server at chalmers will use sftp protocol once implemented this change is required before we can switch to the new server
1
1,313
2,511,569,656
IssuesEvent
2015-01-14 09:19:04
olga-jane/prizm
https://api.github.com/repos/olga-jane/prizm
opened
Export/Import set flag "to export" in proper places
Coding Coding - POOL COMMON HIGH priority
Google document PRIZMA Design contains 2 diagrams in Data export/import section. These diagrams show, when and at what conditions should be set the flag "to export". DB column exists already.
1.0
Export/Import set flag "to export" in proper places - Google document PRIZMA Design contains 2 diagrams in Data export/import section. These diagrams show, when and at what conditions should be set the flag "to export". DB column exists already.
priority
export import set flag to export in proper places google document prizma design contains diagrams in data export import section these diagrams show when and at what conditions should be set the flag to export db column exists already
1
582,212
17,356,087,333
IssuesEvent
2021-07-29 14:33:19
staynomad/Nomad-Back
https://api.github.com/repos/staynomad/Nomad-Back
closed
Cannot Create a Listing
dev:bug difficulty:medium priority:high
# Background <!--- Put any relevant background information here. ---> This error pops up when I try to save the listing I'm creating as a draft ![image](https://user-images.githubusercontent.com/44332326/126838439-a3994e79-0715-4228-9fc4-b036f86eda86.png) Same error when I submit: ![image](https://user-images.githubusercontent.com/44332326/126838619-84c95c86-79d6-4a03-a8ba-7ef330690d39.png) # Task <!--- Put the task here (ideally bullet points). ---> * Fix the route and make it correctly save the draft * Fix the route and make it correctly submit the listing # Done When <!--- Put the completion criteria for the issue here. ---> * Loading spinner doesn't spin indefinitely
1.0
Cannot Create a Listing - # Background <!--- Put any relevant background information here. ---> This error pops up when I try to save the listing I'm creating as a draft ![image](https://user-images.githubusercontent.com/44332326/126838439-a3994e79-0715-4228-9fc4-b036f86eda86.png) Same error when I submit: ![image](https://user-images.githubusercontent.com/44332326/126838619-84c95c86-79d6-4a03-a8ba-7ef330690d39.png) # Task <!--- Put the task here (ideally bullet points). ---> * Fix the route and make it correctly save the draft * Fix the route and make it correctly submit the listing # Done When <!--- Put the completion criteria for the issue here. ---> * Loading spinner doesn't spin indefinitely
priority
cannot create a listing background this error pops up when i try to save the listing i m creating as a draft same error when i submit task fix the route and make it correctly save the draft fix the route and make it correctly submit the listing done when loading spinner doesn t spin indefinitely
1
533,634
15,595,677,285
IssuesEvent
2021-03-18 15:06:55
HDFGroup/hermes
https://api.github.com/repos/HDFGroup/hermes
closed
Go to neighborhood Targets if a Put fails on local Targets
priority-high
Add tests using the cluster simulator.
1.0
Go to neighborhood Targets if a Put fails on local Targets - Add tests using the cluster simulator.
priority
go to neighborhood targets if a put fails on local targets add tests using the cluster simulator
1
307,263
9,415,131,507
IssuesEvent
2019-04-10 11:55:11
meumobi/sitebuilder
https://api.github.com/repos/meumobi/sitebuilder
closed
Enable apache log rotation on it.c-se
high priority
### Expected behaviour On it.c-se env #### Enable apache log rotation; on */etc/apache2/sites-available/{site}* ``` CustomLog "/var/log/apache2/{site}/access.%Y-%m-%d" combined ``` ```$ a2ensite {site}``` #### Track log on scalyr on */etc/scalyr-agent-2/agent.json* ``` { path: "/var/log/apache2/{site}/access.*", attributes: {parser: "accessLog"}, rename_logfile: { match: "/var/log/apache2/{site}/access.([0-9]{4}-[0-9]{2}-[0-9]{2})", replacement: "/scalyr/access.log" } } ``` ```$ sudo scalyr-agent-2 restart``` ### Actual behaviour - There is one single log file ### How to test - open it app - perform some requests - certify if the requests were sent to scalyr
1.0
Enable apache log rotation on it.c-se - ### Expected behaviour On it.c-se env #### Enable apache log rotation; on */etc/apache2/sites-available/{site}* ``` CustomLog "/var/log/apache2/{site}/access.%Y-%m-%d" combined ``` ```$ a2ensite {site}``` #### Track log on scalyr on */etc/scalyr-agent-2/agent.json* ``` { path: "/var/log/apache2/{site}/access.*", attributes: {parser: "accessLog"}, rename_logfile: { match: "/var/log/apache2/{site}/access.([0-9]{4}-[0-9]{2}-[0-9]{2})", replacement: "/scalyr/access.log" } } ``` ```$ sudo scalyr-agent-2 restart``` ### Actual behaviour - There is one single log file ### How to test - open it app - perform some requests - certify if the requests were sent to scalyr
priority
enable apache log rotation on it c se expected behaviour on it c se env enable apache log rotation on etc sites available site customlog var log site access y m d combined site track log on scalyr on etc scalyr agent agent json path var log site access attributes parser accesslog rename logfile match var log site access replacement scalyr access log sudo scalyr agent restart actual behaviour there is one single log file how to test open it app perform some requests certify if the requests were sent to scalyr
1
803,315
29,171,645,475
IssuesEvent
2023-05-19 02:44:25
mlcommons/medperf
https://api.github.com/repos/mlcommons/medperf
opened
[FEATURE] Integration tests need to consider every possible command
enhancement High Priority
We should add integration tests for all possible CLI commands.
1.0
[FEATURE] Integration tests need to consider every possible command - We should add integration tests for all possible CLI commands.
priority
integration tests need to consider every possible command we should add integration tests for all possible cli commands
1
220,929
7,372,243,961
IssuesEvent
2018-03-13 14:16:44
SANBIBiodiversityforLife/species
https://api.github.com/repos/SANBIBiodiversityforLife/species
closed
List of taxa without images and maps within each group
high-priority
Dewidine wants a list of taxa which don't have images, and a list of taxa which don't have any distribution info/maps.
1.0
List of taxa without images and maps within each group - Dewidine wants a list of taxa which don't have images, and a list of taxa which don't have any distribution info/maps.
priority
list of taxa without images and maps within each group dewidine wants a list of taxa which don t have images and a list of taxa which don t have any distribution info maps
1
349,884
10,475,226,755
IssuesEvent
2019-09-23 15:54:26
electrala/documentation
https://api.github.com/repos/electrala/documentation
opened
AAD - I want to create the official db for electra.
Database backend high priority
We need to go into postgres and make the following tables to use in our final project. Critiques table User table comments table Gallery(optional not high priority) Be extra careful and double check everything when creating this, it will be our final db.
1.0
AAD - I want to create the official db for electra. - We need to go into postgres and make the following tables to use in our final project. Critiques table User table comments table Gallery(optional not high priority) Be extra careful and double check everything when creating this, it will be our final db.
priority
aad i want to create the official db for electra we need to go into postgres and make the following tables to use in our final project critiques table user table comments table gallery optional not high priority be extra careful and double check everything when creating this it will be our final db
1
509,328
14,728,460,751
IssuesEvent
2021-01-06 09:59:25
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
web.facebook.com - desktop site instead of mobile site
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox 85.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:85.0) Gecko/20100101 Firefox/85.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/65044 --> **URL**: https://web.facebook.com/ **Browser / Version**: Firefox 85.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/207c5eaf-995d-4723-9933-b73d1e38d86a.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210105185604</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/1/e967b02c-d117-45df-95d9-a3c9125e0413) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
web.facebook.com - desktop site instead of mobile site - <!-- @browser: Firefox 85.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:85.0) Gecko/20100101 Firefox/85.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/65044 --> **URL**: https://web.facebook.com/ **Browser / Version**: Firefox 85.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/207c5eaf-995d-4723-9933-b73d1e38d86a.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210105185604</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/1/e967b02c-d117-45df-95d9-a3c9125e0413) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
web facebook com desktop site instead of mobile site url browser version firefox operating system windows tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
788,410
27,752,094,174
IssuesEvent
2023-03-15 21:41:25
nasa/prog_models
https://api.github.com/repos/nasa/prog_models
opened
Divide by Zero in calc_error
bug Priority: High
Identified by @aqitya PrognosticsModel.calc_error() throws a divide by zero exception when the model is configured in a way that is unstable (i.e., a state becomes NaN) in the first step (e.g., BatteryElectroChemEOD with a qMax of 4000). See https://github.com/nasa/prog_models/blob/f0891d59e4afd726f12812d273f6d97e7abc7a9f/src/prog_models/prognostics_model.py#L1176-L1178 This is of course undesired behavior. I see three options here: 1. if the model goes unstable at any point return NaN error (or Inf if NaN isn't treated correctly by estimate_params) 2. Of the model goes unstable before some cutoff (e.g., 90% of the points have passed), return NaN error. If it goes unstable at any other point, warn and return an estimate of error from the points to that point. 3. Only if the model goes unstable at the first timestep, return NaN error. If it goes unstable at any other point, warn and return an estimate of error from the points to that point. We will want to make sure consider the following cases in our decision: 1. The case where this was first identified, the model is never stable so we get NaN on the first timestep, then we definitely want to return Nan (or Inf, see message above) 2. Case where we're running estimate_params and the model goes unstable midway through. Not sure what we want to do here 3. Case where we're running estimate_params and the model goes unstable at the very last timestep. This is frequently the case when training on run-to-failure data. In this case the parameters may be very close to the "correct" or best values. 4. Case where we're running estimate_params and the model never goes unstable- this one we obviously want to work We also should keep in mind that calc_error can be used outside of estimate_params
1.0
Divide by Zero in calc_error - Identified by @aqitya PrognosticsModel.calc_error() throws a divide by zero exception when the model is configured in a way that is unstable (i.e., a state becomes NaN) in the first step (e.g., BatteryElectroChemEOD with a qMax of 4000). See https://github.com/nasa/prog_models/blob/f0891d59e4afd726f12812d273f6d97e7abc7a9f/src/prog_models/prognostics_model.py#L1176-L1178 This is of course undesired behavior. I see three options here: 1. if the model goes unstable at any point return NaN error (or Inf if NaN isn't treated correctly by estimate_params) 2. Of the model goes unstable before some cutoff (e.g., 90% of the points have passed), return NaN error. If it goes unstable at any other point, warn and return an estimate of error from the points to that point. 3. Only if the model goes unstable at the first timestep, return NaN error. If it goes unstable at any other point, warn and return an estimate of error from the points to that point. We will want to make sure consider the following cases in our decision: 1. The case where this was first identified, the model is never stable so we get NaN on the first timestep, then we definitely want to return Nan (or Inf, see message above) 2. Case where we're running estimate_params and the model goes unstable midway through. Not sure what we want to do here 3. Case where we're running estimate_params and the model goes unstable at the very last timestep. This is frequently the case when training on run-to-failure data. In this case the parameters may be very close to the "correct" or best values. 4. Case where we're running estimate_params and the model never goes unstable- this one we obviously want to work We also should keep in mind that calc_error can be used outside of estimate_params
priority
divide by zero in calc error identified by aqitya prognosticsmodel calc error throws a divide by zero exception when the model is configured in a way that is unstable i e a state becomes nan in the first step e g batteryelectrochemeod with a qmax of see this is of course undesired behavior i see three options here if the model goes unstable at any point return nan error or inf if nan isn t treated correctly by estimate params of the model goes unstable before some cutoff e g of the points have passed return nan error if it goes unstable at any other point warn and return an estimate of error from the points to that point only if the model goes unstable at the first timestep return nan error if it goes unstable at any other point warn and return an estimate of error from the points to that point we will want to make sure consider the following cases in our decision the case where this was first identified the model is never stable so we get nan on the first timestep then we definitely want to return nan or inf see message above case where we re running estimate params and the model goes unstable midway through not sure what we want to do here case where we re running estimate params and the model goes unstable at the very last timestep this is frequently the case when training on run to failure data in this case the parameters may be very close to the correct or best values case where we re running estimate params and the model never goes unstable this one we obviously want to work we also should keep in mind that calc error can be used outside of estimate params
1
265,125
8,337,590,821
IssuesEvent
2018-09-28 11:41:58
ThmmyNoLife/mTHMMY
https://api.github.com/repos/ThmmyNoLife/mTHMMY
closed
crash on orientation change in topic
bug high priority
### Steps to reproduce the problem 1. Open a topic 2. Try to edit a post or write a reply 3. Change the orientation to landscape 4. Change the orientation back to portrait 5 App crashes ### What causes the crash [AutoFitGridLayout](https://github.com/ThmmyNoLife/mTHMMY/blob/develop/app/src/main/java/gr/thmmy/mthmmy/editorview/AutoFitGridLayout.java#L44) auto-changes its columnCount to fit as many items as it can in each row. However, when the app goes from portrait to landscape, this value gets smaller, while some of the layout's children have the old layout params which may have larger column index, causeing the crash ### How to fix this Replace [AutoFitGridLayout](https://github.com/ThmmyNoLife/mTHMMY/blob/develop/app/src/main/java/gr/thmmy/mthmmy/editorview/AutoFitGridLayout.java) with a recyclerview that uses a GridLayoutManager
1.0
crash on orientation change in topic - ### Steps to reproduce the problem 1. Open a topic 2. Try to edit a post or write a reply 3. Change the orientation to landscape 4. Change the orientation back to portrait 5 App crashes ### What causes the crash [AutoFitGridLayout](https://github.com/ThmmyNoLife/mTHMMY/blob/develop/app/src/main/java/gr/thmmy/mthmmy/editorview/AutoFitGridLayout.java#L44) auto-changes its columnCount to fit as many items as it can in each row. However, when the app goes from portrait to landscape, this value gets smaller, while some of the layout's children have the old layout params which may have larger column index, causeing the crash ### How to fix this Replace [AutoFitGridLayout](https://github.com/ThmmyNoLife/mTHMMY/blob/develop/app/src/main/java/gr/thmmy/mthmmy/editorview/AutoFitGridLayout.java) with a recyclerview that uses a GridLayoutManager
priority
crash on orientation change in topic steps to reproduce the problem open a topic try to edit a post or write a reply change the orientation to landscape change the orientation back to portrait app crashes what causes the crash auto changes its columncount to fit as many items as it can in each row however when the app goes from portrait to landscape this value gets smaller while some of the layout s children have the old layout params which may have larger column index causeing the crash how to fix this replace with a recyclerview that uses a gridlayoutmanager
1
113,385
4,553,199,043
IssuesEvent
2016-09-13 03:12:48
FelipeRosa/clojure-gsgp
https://api.github.com/repos/FelipeRosa/clojure-gsgp
closed
Implement data exporters
Priority: High Type: Feature Request
We need data exporters (.txt, .csv, etc) in order to plot prediction results.
1.0
Implement data exporters - We need data exporters (.txt, .csv, etc) in order to plot prediction results.
priority
implement data exporters we need data exporters txt csv etc in order to plot prediction results
1
83,688
3,640,755,389
IssuesEvent
2016-02-13 03:55:17
RestComm/RestComm-Core
https://api.github.com/repos/RestComm/RestComm-Core
opened
Add support for Japanese language in RVD Say verb
High-Priority Task
Add Japanese option to list of languages in RVD Say verb.
1.0
Add support for Japanese language in RVD Say verb - Add Japanese option to list of languages in RVD Say verb.
priority
add support for japanese language in rvd say verb add japanese option to list of languages in rvd say verb
1
392,098
11,583,522,021
IssuesEvent
2020-02-22 11:37:40
WoWManiaUK/Blackwing-Lair
https://api.github.com/repos/WoWManiaUK/Blackwing-Lair
closed
[Boss] Omnotron Defense System - Blackwing Descent
Dungeon/Raid Fixed in Dev Priority-High
**Links:** https://www.wowhead.com/npc=42180/toxitron#comments **What is happening:** Nothing worked with this fight 1) the mobs healed faster than we could do damage 2) only two were killable with GM commands 3) they never moved 4)tank and spank didn't do any ability's even when we tried to kill them before using GM commands 5) **it stops the progression of the raid** because the gates wouldn't open without gm on **What should happen:** When the fight starts, only one mob will be active, and three stone golems will be frozen near it. Like most council fights, there is a shared HP pool. you start by attacking the active one and when it reaches 50% energy a second one becomes active with 100% energy when this one hits 50% it will activate a shield and the next one become targetable shutting the first one down. and so on until the end of the fight the activation is random, but it will always be one of the two that is is inactive Golems have 100 Energy when first activated, and the way the Energy works is the same for all of them: 100 Energy= Golem just activated 50 Energy= Next Golem is activated 45ish Energy= Golem puts up his Shield 25ish Energy = Golem's shield falls 0 Energy = Golem is shut down, becomes stone and untargetable. (Zero energy on your current target means 50 energy on the other active golem, so a new golem will be active). **Magmatron: abilities** - Incineration Security Measure: Basic fire AOE. Acquiring Target: Picks a target, with a giant red laser to let you know who it is. After a few seconds, Magmatron Turns and blasts a cone of Fire AOE at the target. Barrier: This is his Shield, absorbs damage then nukes the raid if it takes too much. https://www.wowhead.com/spell=78740 -activated https://www.wowhead.com/spell=78726 -inactive https://www.wowhead.com/spell=79023 -security measures https://www.wowhead.com/spell=78746 shutting down **Arcanotron: Abilities** - Power Generator: Puts this down on the ground. Anyone in it gets a damage buff and mana regen buff. Don't let the boss stand in it. Think Steelbreaker from Ulduar. The fight being long, these are essential to healers getting mana back. Arcane Annihlator: Fast basic cast on anyone. Can be interrupted but is fast enough to be difficult to do so. Power Conversion: This is his Shield. Each attack gives him a buff increasing his damage. Spell Stealabl https://www.wowhead.com/spell=78740 - activated https://www.wowhead.com/spell=79710 -arcane anhillitaor https://www.wowhead.com/spell=78726 - inactive https://www.wowhead.com/spell=78746 -shutting down **Toxitron abilities:** Poison Cloud: anyone in it takes 50 percent damage, including the boss Poison Protocol: summons three slimes. These seek out random targets, and if they hit them, explode for massive damage. This can include Melee and the offtank, If it hits a target, it leaves a poison pool on the ground. The adds don't have much HP and are pretty slow. Poison Soaked Shell: This is his Shield.Each time you attack you get a stacking dot. However each stack causes you to inflict extra damage. https://www.wowhead.com/spell=78740 -activates https://www.wowhead.com/spell=80157 -chemical boss https://www.wowhead.com/spell=78726 -inactive https://www.wowhead.com/spell=80053 -posion protocol https://www.wowhead.com/spell=78746 -shutting down **Electron abilties** Single Target Ability - (Electrical Discharge) - Randomly targets a raid member and zaps them with a chain lightning for ~30k damage and jumps to up to three targets for +20% damage each jump. AoE Ability - (Lightning Conductor) - This is an ability that looks like an Ele Shaman's Thunderstorm. It will be cast on a random raid member... Including both tanks and any DPS. . It hits everyone within range for ~20k damage per second. Shield Ability - (Unstable Shield) - This is his shield ability, and if too much damage is done to him while this is active it will do an empowered Electrical Discharge type thing. It's a raid wiper if this hurts anyone. It's ~60k damage per second to anyone in range. https://www.wowhead.com/spell=78740 -activate https://www.wowhead.com/spell=79879 -electrical discharge https://www.wowhead.com/spell=78726 -inactive https://www.wowhead.com/spell=79888 -lighnigh conductor https://www.wowhead.com/spell=78746 - shutting down
1.0
[Boss] Omnotron Defense System - Blackwing Descent - **Links:** https://www.wowhead.com/npc=42180/toxitron#comments **What is happening:** Nothing worked with this fight 1) the mobs healed faster than we could do damage 2) only two were killable with GM commands 3) they never moved 4)tank and spank didn't do any ability's even when we tried to kill them before using GM commands 5) **it stops the progression of the raid** because the gates wouldn't open without gm on **What should happen:** When the fight starts, only one mob will be active, and three stone golems will be frozen near it. Like most council fights, there is a shared HP pool. you start by attacking the active one and when it reaches 50% energy a second one becomes active with 100% energy when this one hits 50% it will activate a shield and the next one become targetable shutting the first one down. and so on until the end of the fight the activation is random, but it will always be one of the two that is is inactive Golems have 100 Energy when first activated, and the way the Energy works is the same for all of them: 100 Energy= Golem just activated 50 Energy= Next Golem is activated 45ish Energy= Golem puts up his Shield 25ish Energy = Golem's shield falls 0 Energy = Golem is shut down, becomes stone and untargetable. (Zero energy on your current target means 50 energy on the other active golem, so a new golem will be active). **Magmatron: abilities** - Incineration Security Measure: Basic fire AOE. Acquiring Target: Picks a target, with a giant red laser to let you know who it is. After a few seconds, Magmatron Turns and blasts a cone of Fire AOE at the target. Barrier: This is his Shield, absorbs damage then nukes the raid if it takes too much. https://www.wowhead.com/spell=78740 -activated https://www.wowhead.com/spell=78726 -inactive https://www.wowhead.com/spell=79023 -security measures https://www.wowhead.com/spell=78746 shutting down **Arcanotron: Abilities** - Power Generator: Puts this down on the ground. Anyone in it gets a damage buff and mana regen buff. Don't let the boss stand in it. Think Steelbreaker from Ulduar. The fight being long, these are essential to healers getting mana back. Arcane Annihlator: Fast basic cast on anyone. Can be interrupted but is fast enough to be difficult to do so. Power Conversion: This is his Shield. Each attack gives him a buff increasing his damage. Spell Stealabl https://www.wowhead.com/spell=78740 - activated https://www.wowhead.com/spell=79710 -arcane anhillitaor https://www.wowhead.com/spell=78726 - inactive https://www.wowhead.com/spell=78746 -shutting down **Toxitron abilities:** Poison Cloud: anyone in it takes 50 percent damage, including the boss Poison Protocol: summons three slimes. These seek out random targets, and if they hit them, explode for massive damage. This can include Melee and the offtank, If it hits a target, it leaves a poison pool on the ground. The adds don't have much HP and are pretty slow. Poison Soaked Shell: This is his Shield.Each time you attack you get a stacking dot. However each stack causes you to inflict extra damage. https://www.wowhead.com/spell=78740 -activates https://www.wowhead.com/spell=80157 -chemical boss https://www.wowhead.com/spell=78726 -inactive https://www.wowhead.com/spell=80053 -posion protocol https://www.wowhead.com/spell=78746 -shutting down **Electron abilties** Single Target Ability - (Electrical Discharge) - Randomly targets a raid member and zaps them with a chain lightning for ~30k damage and jumps to up to three targets for +20% damage each jump. AoE Ability - (Lightning Conductor) - This is an ability that looks like an Ele Shaman's Thunderstorm. It will be cast on a random raid member... Including both tanks and any DPS. . It hits everyone within range for ~20k damage per second. Shield Ability - (Unstable Shield) - This is his shield ability, and if too much damage is done to him while this is active it will do an empowered Electrical Discharge type thing. It's a raid wiper if this hurts anyone. It's ~60k damage per second to anyone in range. https://www.wowhead.com/spell=78740 -activate https://www.wowhead.com/spell=79879 -electrical discharge https://www.wowhead.com/spell=78726 -inactive https://www.wowhead.com/spell=79888 -lighnigh conductor https://www.wowhead.com/spell=78746 - shutting down
priority
omnotron defense system blackwing descent links what is happening nothing worked with this fight the mobs healed faster than we could do damage only two were killable with gm commands they never moved tank and spank didn t do any ability s even when we tried to kill them before using gm commands it stops the progression of the raid because the gates wouldn t open without gm on what should happen when the fight starts only one mob will be active and three stone golems will be frozen near it like most council fights there is a shared hp pool you start by attacking the active one and when it reaches energy a second one becomes active with energy when this one hits it will activate a shield and the next one become targetable shutting the first one down and so on until the end of the fight the activation is random but it will always be one of the two that is is inactive golems have energy when first activated and the way the energy works is the same for all of them energy golem just activated energy next golem is activated energy golem puts up his shield energy golem s shield falls energy golem is shut down becomes stone and untargetable zero energy on your current target means energy on the other active golem so a new golem will be active magmatron abilities incineration security measure basic fire aoe acquiring target picks a target with a giant red laser to let you know who it is after a few seconds magmatron turns and blasts a cone of fire aoe at the target barrier this is his shield absorbs damage then nukes the raid if it takes too much activated inactive security measures shutting down arcanotron abilities power generator puts this down on the ground anyone in it gets a damage buff and mana regen buff don t let the boss stand in it think steelbreaker from ulduar the fight being long these are essential to healers getting mana back arcane annihlator fast basic cast on anyone can be interrupted but is fast enough to be difficult to do so power conversion this is his shield each attack gives him a buff increasing his damage spell stealabl activated arcane anhillitaor inactive shutting down toxitron abilities poison cloud anyone in it takes percent damage including the boss poison protocol summons three slimes these seek out random targets and if they hit them explode for massive damage this can include melee and the offtank if it hits a target it leaves a poison pool on the ground the adds don t have much hp and are pretty slow poison soaked shell this is his shield each time you attack you get a stacking dot however each stack causes you to inflict extra damage activates chemical boss inactive posion protocol shutting down electron abilties single target ability electrical discharge randomly targets a raid member and zaps them with a chain lightning for damage and jumps to up to three targets for damage each jump aoe ability lightning conductor this is an ability that looks like an ele shaman s thunderstorm it will be cast on a random raid member including both tanks and any dps it hits everyone within range for damage per second shield ability unstable shield this is his shield ability and if too much damage is done to him while this is active it will do an empowered electrical discharge type thing it s a raid wiper if this hurts anyone it s damage per second to anyone in range activate electrical discharge inactive lighnigh conductor shutting down
1
396,680
11,712,175,692
IssuesEvent
2020-03-09 07:40:35
AY1920S2-CS2103T-T10-2/main
https://api.github.com/repos/AY1920S2-CS2103T-T10-2/main
closed
Add ability to save user answers to corresponding flashcards
priority.High type.Task
For use later, when user needs to view their answer history/statistics
1.0
Add ability to save user answers to corresponding flashcards - For use later, when user needs to view their answer history/statistics
priority
add ability to save user answers to corresponding flashcards for use later when user needs to view their answer history statistics
1
209,473
7,176,521,177
IssuesEvent
2018-01-31 10:18:56
VALIDproject/netflower
https://api.github.com/repos/VALIDproject/netflower
closed
BUG with Filters and update time
bug high priority
- Filters update not appropriately - Filters are not applying to all and some are to restrictive - The data isn't fully correct - The UI should update as well appropiately
1.0
BUG with Filters and update time - - Filters update not appropriately - Filters are not applying to all and some are to restrictive - The data isn't fully correct - The UI should update as well appropiately
priority
bug with filters and update time filters update not appropriately filters are not applying to all and some are to restrictive the data isn t fully correct the ui should update as well appropiately
1
524,070
15,195,403,564
IssuesEvent
2021-02-16 06:16:10
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.3.0 beta staging-1930]Industrial Elevator missing properties in .csv
Category: Balance Priority: High Regression Status: Fixed Type: Bug
Industrial elevator does not have a recipe accompanied as its blank in .csv but populated in techtree.xlsm There is an inconsistency here if someone didnt properly pull an up to date techtree.csv when updating?
1.0
[0.9.3.0 beta staging-1930]Industrial Elevator missing properties in .csv - Industrial elevator does not have a recipe accompanied as its blank in .csv but populated in techtree.xlsm There is an inconsistency here if someone didnt properly pull an up to date techtree.csv when updating?
priority
industrial elevator missing properties in csv industrial elevator does not have a recipe accompanied as its blank in csv but populated in techtree xlsm there is an inconsistency here if someone didnt properly pull an up to date techtree csv when updating
1
494,675
14,263,003,799
IssuesEvent
2020-11-20 13:50:25
reichlab/covidHubUtils
https://api.github.com/repos/reichlab/covidHubUtils
opened
create facet-ready plot_forecasts() function
High Priority
specs via google doc design. #' @param forecast_data data frame with truth and forecasts from load_forecasts() #' @param truth_data optional data frame with forecasts in the format returned #' by load_truth(). #' @param models model_abbr specifying model to plot. Optional if there is only #' one model available in forecast data. #' @param target_variable string specifying target type. It should be one of #' "cum death", "inc case", "inc death" #' @param locations string for fips code or 'US'. Optional if there is only one #' location available in forecast data. #' @param facet interpretable facet option for ggplot #’ @param facet_scales argument for scales in ggplot2::facet_wrap #’ @param forecast_dates forecast dates to plot, defaults to forecast_dates present #’ in the data #' @param intervals values indicating which central prediction interval levels #' to plot, defaults to c(.5, .8, .95). NULL means only plotting point forecasts. #' If not provided, it will default to all available intervals in forecast data. #' @param horizon forecasts are plotted for the horizon time steps after the #' forecast date. Default to all available horizons in forecast data. #' @param truth_source character specifying where the truth data will #' be loaded from if truth_data is not provided. #' Otherwise, this character specifies the data source to plot. #' Currently support "JHU","USAFacts" and "NYTimes". #' @param truth_as_of the plot includes the truth data that would have been #' in real time as of the truth_as_of date (not using this parameter when truth data #' is from github repo) #' @param title optional text for the title of the plot. if left as "default", #' the title will be automatically generated. if "none", no title will be plotted. #' @param show.caption logical, if TRUE, caption will be included showing data sources #’ @param plot logical; if TRUE, the plot is printed #' #' @return invisible ggplot object #' #' @export plot_forecasts <- function(forecast_data, truth_data = NULL, models, target_variable, locations, forecast_dates, facet, intervals = c(.5, .8, .95), horizon, truth_source = "JHU", truth_as_of = NULL, title = "default", show.caption = TRUE, plot = TRUE)
1.0
create facet-ready plot_forecasts() function - specs via google doc design. #' @param forecast_data data frame with truth and forecasts from load_forecasts() #' @param truth_data optional data frame with forecasts in the format returned #' by load_truth(). #' @param models model_abbr specifying model to plot. Optional if there is only #' one model available in forecast data. #' @param target_variable string specifying target type. It should be one of #' "cum death", "inc case", "inc death" #' @param locations string for fips code or 'US'. Optional if there is only one #' location available in forecast data. #' @param facet interpretable facet option for ggplot #’ @param facet_scales argument for scales in ggplot2::facet_wrap #’ @param forecast_dates forecast dates to plot, defaults to forecast_dates present #’ in the data #' @param intervals values indicating which central prediction interval levels #' to plot, defaults to c(.5, .8, .95). NULL means only plotting point forecasts. #' If not provided, it will default to all available intervals in forecast data. #' @param horizon forecasts are plotted for the horizon time steps after the #' forecast date. Default to all available horizons in forecast data. #' @param truth_source character specifying where the truth data will #' be loaded from if truth_data is not provided. #' Otherwise, this character specifies the data source to plot. #' Currently support "JHU","USAFacts" and "NYTimes". #' @param truth_as_of the plot includes the truth data that would have been #' in real time as of the truth_as_of date (not using this parameter when truth data #' is from github repo) #' @param title optional text for the title of the plot. if left as "default", #' the title will be automatically generated. if "none", no title will be plotted. #' @param show.caption logical, if TRUE, caption will be included showing data sources #’ @param plot logical; if TRUE, the plot is printed #' #' @return invisible ggplot object #' #' @export plot_forecasts <- function(forecast_data, truth_data = NULL, models, target_variable, locations, forecast_dates, facet, intervals = c(.5, .8, .95), horizon, truth_source = "JHU", truth_as_of = NULL, title = "default", show.caption = TRUE, plot = TRUE)
priority
create facet ready plot forecasts function specs via google doc design param forecast data data frame with truth and forecasts from load forecasts param truth data optional data frame with forecasts in the format returned by load truth param models model abbr specifying model to plot optional if there is only one model available in forecast data param target variable string specifying target type it should be one of cum death inc case inc death param locations string for fips code or us optional if there is only one location available in forecast data param facet interpretable facet option for ggplot ’ param facet scales argument for scales in facet wrap ’ param forecast dates forecast dates to plot defaults to forecast dates present ’ in the data param intervals values indicating which central prediction interval levels to plot defaults to c null means only plotting point forecasts if not provided it will default to all available intervals in forecast data param horizon forecasts are plotted for the horizon time steps after the forecast date default to all available horizons in forecast data param truth source character specifying where the truth data will be loaded from if truth data is not provided otherwise this character specifies the data source to plot currently support jhu usafacts and nytimes param truth as of the plot includes the truth data that would have been in real time as of the truth as of date not using this parameter when truth data is from github repo param title optional text for the title of the plot if left as default the title will be automatically generated if none no title will be plotted param show caption logical if true caption will be included showing data sources ’ param plot logical if true the plot is printed return invisible ggplot object export plot forecasts function forecast data truth data null models target variable locations forecast dates facet intervals c horizon truth source jhu truth as of null title default show caption true plot true
1
613,561
19,093,797,039
IssuesEvent
2021-11-29 14:47:39
yukiHaga/regex-hunting
https://api.github.com/repos/yukiHaga/regex-hunting
closed
seed_fuを使ってDBにデータを追加する。
Priority: high
## 概要 seed_fuを使ってDBにデータを追加します。 ## やること - [x] `gem 'seed-fu'`をGemfileに記入してインストールする。 - [x] `mkdir -p db/fixtures/{development, production}`を実行して、ディレクトリを作成する。 - [x] `touch db/fixtures/development/ファイル名.rb`で開発DB用のシードファイルを作成する。 - [x] テーブルの数だけシードファイルを作成する。 - [x] `rails db:seed_fu`を実行して、DBにデータを追加する。 ## 受け入れ条件 - [x] seedデータがちゃんとDBに存在する。 - [x] `rails db:seed_fu`を2回実行しても、重複したデータがDBに保存されない。 ## 参考記事 - [【Rails】seed_fuまとめ](https://qiita.com/d0ne1s/items/11d4e51fa11e65653e12) - [railsで初期データを入れる(seed-fuの使い方)](https://qiita.com/ko2ic/items/be96e450a33d631e0059)
1.0
seed_fuを使ってDBにデータを追加する。 - ## 概要 seed_fuを使ってDBにデータを追加します。 ## やること - [x] `gem 'seed-fu'`をGemfileに記入してインストールする。 - [x] `mkdir -p db/fixtures/{development, production}`を実行して、ディレクトリを作成する。 - [x] `touch db/fixtures/development/ファイル名.rb`で開発DB用のシードファイルを作成する。 - [x] テーブルの数だけシードファイルを作成する。 - [x] `rails db:seed_fu`を実行して、DBにデータを追加する。 ## 受け入れ条件 - [x] seedデータがちゃんとDBに存在する。 - [x] `rails db:seed_fu`を2回実行しても、重複したデータがDBに保存されない。 ## 参考記事 - [【Rails】seed_fuまとめ](https://qiita.com/d0ne1s/items/11d4e51fa11e65653e12) - [railsで初期データを入れる(seed-fuの使い方)](https://qiita.com/ko2ic/items/be96e450a33d631e0059)
priority
seed fuを使ってdbにデータを追加する。 概要 seed fuを使ってdbにデータを追加します。 やること gem seed fu をgemfileに記入してインストールする。 mkdir p db fixtures development production を実行して、ディレクトリを作成する。 touch db fixtures development ファイル名 rb で開発db用のシードファイルを作成する。 テーブルの数だけシードファイルを作成する。 rails db seed fu を実行して、dbにデータを追加する。 受け入れ条件 seedデータがちゃんとdbに存在する。 rails db seed fu 、重複したデータがdbに保存されない。 参考記事
1
303,306
9,305,140,688
IssuesEvent
2019-03-25 05:02:32
richmondrcmp/mobileapp
https://api.github.com/repos/richmondrcmp/mobileapp
closed
Weather high and lows reversed with negative temperatures
Issue Type - Bug Priority - High
If the highs and/or lows have a negative value the high and lows are reversed (i.e., the low (-4oC) is the high and the high (1oC) is the low): ![image](https://user-images.githubusercontent.com/47233113/54873393-9ca2f500-4d92-11e9-8a6f-371e147a0642.png) and there seems to be a disconnect between the current, highs and lows (and maybe the weather icon) between the main menu and the weather page: ![image](https://user-images.githubusercontent.com/47233113/54873394-a298d600-4d92-11e9-8c3b-150da7d85048.png)
1.0
Weather high and lows reversed with negative temperatures - If the highs and/or lows have a negative value the high and lows are reversed (i.e., the low (-4oC) is the high and the high (1oC) is the low): ![image](https://user-images.githubusercontent.com/47233113/54873393-9ca2f500-4d92-11e9-8a6f-371e147a0642.png) and there seems to be a disconnect between the current, highs and lows (and maybe the weather icon) between the main menu and the weather page: ![image](https://user-images.githubusercontent.com/47233113/54873394-a298d600-4d92-11e9-8c3b-150da7d85048.png)
priority
weather high and lows reversed with negative temperatures if the highs and or lows have a negative value the high and lows are reversed i e the low is the high and the high is the low and there seems to be a disconnect between the current highs and lows and maybe the weather icon between the main menu and the weather page
1
244,397
7,874,877,070
IssuesEvent
2018-06-25 18:30:24
socialappslab/appcivist-platform
https://api.github.com/repos/socialappslab/appcivist-platform
closed
Logout from profile page
Priority: High duplicate
Logging out from the profile page goes to a non-LDAP window because you are at a page that has no information about any particular assembly or campaign and so AppCivist has no way of knowing which specific sign-in to show. The problem is that logout redirects to the same page, and when it redirects to the profile page without a logged in user, you have all the data empty. Change logout from profile page to redirect to the last page visited.
1.0
Logout from profile page - Logging out from the profile page goes to a non-LDAP window because you are at a page that has no information about any particular assembly or campaign and so AppCivist has no way of knowing which specific sign-in to show. The problem is that logout redirects to the same page, and when it redirects to the profile page without a logged in user, you have all the data empty. Change logout from profile page to redirect to the last page visited.
priority
logout from profile page logging out from the profile page goes to a non ldap window because you are at a page that has no information about any particular assembly or campaign and so appcivist has no way of knowing which specific sign in to show the problem is that logout redirects to the same page and when it redirects to the profile page without a logged in user you have all the data empty change logout from profile page to redirect to the last page visited
1
206,758
7,121,060,314
IssuesEvent
2018-01-19 05:23:12
wso2/product-iots
https://api.github.com/repos/wso2/product-iots
closed
[iOS] Device vitals are not getting updated
Component/IOS Priority/High Type/Bug
**Description:** iOS device vitals are not getting updated after the initial enrolment. At the time of the enrolment, device publishes its initial vitals and then after, it will not get changed. **Affected Product Version:** IoT-3.1.0 **OS, DB, other environment details and versions:** Mac OSX 10.13.2 MySQL 5.7 **Steps to reproduce:** Enrol an iOS device and wait for few minutes until vital are changed from the device. Then check the device details page.
1.0
[iOS] Device vitals are not getting updated - **Description:** iOS device vitals are not getting updated after the initial enrolment. At the time of the enrolment, device publishes its initial vitals and then after, it will not get changed. **Affected Product Version:** IoT-3.1.0 **OS, DB, other environment details and versions:** Mac OSX 10.13.2 MySQL 5.7 **Steps to reproduce:** Enrol an iOS device and wait for few minutes until vital are changed from the device. Then check the device details page.
priority
device vitals are not getting updated description ios device vitals are not getting updated after the initial enrolment at the time of the enrolment device publishes its initial vitals and then after it will not get changed affected product version iot os db other environment details and versions mac osx mysql steps to reproduce enrol an ios device and wait for few minutes until vital are changed from the device then check the device details page
1
651,192
21,468,973,302
IssuesEvent
2022-04-26 07:46:37
labsquare/cutevariant
https://api.github.com/repos/labsquare/cutevariant
closed
[Validation] Not refresh when new project/DB
bug high-priority CRITICAL
Please edit this template: **Describe the bug** When a new project/DB is created or loaded, list of samples in Validation widget is not refreshed. **To Reproduce** Steps to reproduce the behavior, for example: 1. Create/Open a project/DB 2. Select Samples in Validation widget 3. Create/Open a NEW project/DB 4. See that sample list is still the same **Expected behavior** List of samples cleared in Validation widget when create/open a new project/DB. **Screenshots** ![new_project_same_sample_in_validation_widget](https://user-images.githubusercontent.com/40463532/165132076-84ae8db5-fa15-45e2-8e9a-5c77391ca5b6.gif) **Desktop (please complete the following information):** - OS: MacOs - Version of Cutevariant: devel
1.0
[Validation] Not refresh when new project/DB - Please edit this template: **Describe the bug** When a new project/DB is created or loaded, list of samples in Validation widget is not refreshed. **To Reproduce** Steps to reproduce the behavior, for example: 1. Create/Open a project/DB 2. Select Samples in Validation widget 3. Create/Open a NEW project/DB 4. See that sample list is still the same **Expected behavior** List of samples cleared in Validation widget when create/open a new project/DB. **Screenshots** ![new_project_same_sample_in_validation_widget](https://user-images.githubusercontent.com/40463532/165132076-84ae8db5-fa15-45e2-8e9a-5c77391ca5b6.gif) **Desktop (please complete the following information):** - OS: MacOs - Version of Cutevariant: devel
priority
not refresh when new project db please edit this template describe the bug when a new project db is created or loaded list of samples in validation widget is not refreshed to reproduce steps to reproduce the behavior for example create open a project db select samples in validation widget create open a new project db see that sample list is still the same expected behavior list of samples cleared in validation widget when create open a new project db screenshots desktop please complete the following information os macos version of cutevariant devel
1
611,080
18,944,377,345
IssuesEvent
2021-11-18 08:33:44
finos/legend
https://api.github.com/repos/finos/legend
opened
Legend Studio fails with 401 (Unauthorized)
bug high priority
[Opening issue on behalf of Shane Morris (RBC)] Hi! We are attempting to install Legend Studio locally so that we can begin to contribute some code. We were thinking that a good starting point would be to try some of the good first issues as recommended in the documentation. When we were attempting to install using the yarn commands, we ran into some issues. We have SDLC, Engine, Gitlab, and mongodb running in a Linux box and were planning on using this for development. We wanted to run yarn on our local machine and connect to the Linux box for the Engine and SDLC backbends. To do this we change the config.json file located at `legend-studio/packages/legend-studio-deployment/dev/config.json` to the following: ``` { "appName": "studio", "env": "local", "sdlc": { "url": "http://<linux-box-ip>:7070/api" }, "engine": { "url": "http://<linux-box-ip>:6060/api" }, "depot": { "url": "http://<linux-box-ip>:9090/depot/api" }, "documentation": { "url": "https://legend.finos.org" } } ``` After doing this we received the following error: (I have blocked out the Linux box IP in the console logs) ![image002](https://user-images.githubusercontent.com/327285/142379369-561f4549-5695-46ad-80d7-449ccaeada98.jpg) We were wondering if this type of setup is possible, and why this error might be occurring. We understand that this issue is known to be caused by SDLC and Engine not being authenticated, but we already attempted to: 1. Grab the authorization tokens from SDLC and Engine 2. Hard refresh Chrome window and delete cookies 3. Use Chrome Incognito 4. Check the api/currentUser endpoint on Engine (and it returns user data as expected) Before doing more experiments (ie, running SDLC and Engine locally, and see if that solves the issue), could someone from the Legend team confirm that it is possible to run a Legend Studio instance in a local environment, but having SDLC and Engine on a remote server? Was this configuration ever tested? Thanks in advance!
1.0
Legend Studio fails with 401 (Unauthorized) - [Opening issue on behalf of Shane Morris (RBC)] Hi! We are attempting to install Legend Studio locally so that we can begin to contribute some code. We were thinking that a good starting point would be to try some of the good first issues as recommended in the documentation. When we were attempting to install using the yarn commands, we ran into some issues. We have SDLC, Engine, Gitlab, and mongodb running in a Linux box and were planning on using this for development. We wanted to run yarn on our local machine and connect to the Linux box for the Engine and SDLC backbends. To do this we change the config.json file located at `legend-studio/packages/legend-studio-deployment/dev/config.json` to the following: ``` { "appName": "studio", "env": "local", "sdlc": { "url": "http://<linux-box-ip>:7070/api" }, "engine": { "url": "http://<linux-box-ip>:6060/api" }, "depot": { "url": "http://<linux-box-ip>:9090/depot/api" }, "documentation": { "url": "https://legend.finos.org" } } ``` After doing this we received the following error: (I have blocked out the Linux box IP in the console logs) ![image002](https://user-images.githubusercontent.com/327285/142379369-561f4549-5695-46ad-80d7-449ccaeada98.jpg) We were wondering if this type of setup is possible, and why this error might be occurring. We understand that this issue is known to be caused by SDLC and Engine not being authenticated, but we already attempted to: 1. Grab the authorization tokens from SDLC and Engine 2. Hard refresh Chrome window and delete cookies 3. Use Chrome Incognito 4. Check the api/currentUser endpoint on Engine (and it returns user data as expected) Before doing more experiments (ie, running SDLC and Engine locally, and see if that solves the issue), could someone from the Legend team confirm that it is possible to run a Legend Studio instance in a local environment, but having SDLC and Engine on a remote server? Was this configuration ever tested? Thanks in advance!
priority
legend studio fails with unauthorized hi we are attempting to install legend studio locally so that we can begin to contribute some code we were thinking that a good starting point would be to try some of the good first issues as recommended in the documentation when we were attempting to install using the yarn commands we ran into some issues we have sdlc engine gitlab and mongodb running in a linux box and were planning on using this for development we wanted to run yarn on our local machine and connect to the linux box for the engine and sdlc backbends to do this we change the config json file located at legend studio packages legend studio deployment dev config json to the following appname studio env local sdlc url engine url depot url documentation url after doing this we received the following error i have blocked out the linux box ip in the console logs we were wondering if this type of setup is possible and why this error might be occurring we understand that this issue is known to be caused by sdlc and engine not being authenticated but we already attempted to grab the authorization tokens from sdlc and engine hard refresh chrome window and delete cookies use chrome incognito check the api currentuser endpoint on engine and it returns user data as expected before doing more experiments ie running sdlc and engine locally and see if that solves the issue could someone from the legend team confirm that it is possible to run a legend studio instance in a local environment but having sdlc and engine on a remote server was this configuration ever tested thanks in advance
1
57,857
3,086,120,241
IssuesEvent
2015-08-25 00:08:51
google/paco
https://api.github.com/repos/google/paco
closed
web ui: Participation Stats for Joined should append who query param
Component-Server Component-UI Priority-High
when clicking on 'Stats' under the joined list it should return just that user's data by appending the "who" query param. For example, /participantStats?experimentId=1234&who=user1@gmail.com
1.0
web ui: Participation Stats for Joined should append who query param - when clicking on 'Stats' under the joined list it should return just that user's data by appending the "who" query param. For example, /participantStats?experimentId=1234&who=user1@gmail.com
priority
web ui participation stats for joined should append who query param when clicking on stats under the joined list it should return just that user s data by appending the who query param for example participantstats experimentid who gmail com
1
783,159
27,520,739,603
IssuesEvent
2023-03-06 14:53:41
AY2223S2-CS2103-W16-3/tp
https://api.github.com/repos/AY2223S2-CS2103-W16-3/tp
closed
As an event planner, I can view all my events
type.Story priority.high
... so that I can track all the upcoming events I have.
1.0
As an event planner, I can view all my events - ... so that I can track all the upcoming events I have.
priority
as an event planner i can view all my events so that i can track all the upcoming events i have
1
138,655
5,345,255,790
IssuesEvent
2017-02-17 16:31:13
CoderDojo/community-platform
https://api.github.com/repos/CoderDojo/community-platform
closed
Add timezones to Events
bug events high priority top priority
All event times are being inputted in GMT by default and there are no options to change this. This is problematic as the event gets removed at the date & GMT time on the event rather than local time, particularly in American timezones which are behind GMT meaning the event gets removed too early. The quick fix for this would be for events to only remove +10 hours after the event but in the long term we will want a drop down or ability to add custom timezones by the user when creating the event.
2.0
Add timezones to Events - All event times are being inputted in GMT by default and there are no options to change this. This is problematic as the event gets removed at the date & GMT time on the event rather than local time, particularly in American timezones which are behind GMT meaning the event gets removed too early. The quick fix for this would be for events to only remove +10 hours after the event but in the long term we will want a drop down or ability to add custom timezones by the user when creating the event.
priority
add timezones to events all event times are being inputted in gmt by default and there are no options to change this this is problematic as the event gets removed at the date gmt time on the event rather than local time particularly in american timezones which are behind gmt meaning the event gets removed too early the quick fix for this would be for events to only remove hours after the event but in the long term we will want a drop down or ability to add custom timezones by the user when creating the event
1
59,468
3,113,786,047
IssuesEvent
2015-09-03 02:25:56
onyxfish/agate
https://api.github.com/repos/onyxfish/agate
closed
In tutorial Formula generates DecimalColumn even though text_type is specified?
bug priority-high
See traceback on #201
1.0
In tutorial Formula generates DecimalColumn even though text_type is specified? - See traceback on #201
priority
in tutorial formula generates decimalcolumn even though text type is specified see traceback on
1
88,619
3,783,250,645
IssuesEvent
2016-03-19 01:30:01
durhamatletico/durhamatletico-cms
https://api.github.com/repos/durhamatletico/durhamatletico-cms
closed
registration instructions in Spanish
content high priority
This text should appear early in the registration instructions. It says, "if you want to register an entire team, the cost is $600. Call or send a text to xxx... for more information." The number is Alfonso's Google Voice number. Si quieren registrar al equipo entero, el coste es de $600. Llame o escribe un texto al 919-937-2109 para mas información"
1.0
registration instructions in Spanish - This text should appear early in the registration instructions. It says, "if you want to register an entire team, the cost is $600. Call or send a text to xxx... for more information." The number is Alfonso's Google Voice number. Si quieren registrar al equipo entero, el coste es de $600. Llame o escribe un texto al 919-937-2109 para mas información"
priority
registration instructions in spanish this text should appear early in the registration instructions it says if you want to register an entire team the cost is call or send a text to xxx for more information the number is alfonso s google voice number si quieren registrar al equipo entero el coste es de llame o escribe un texto al para mas información
1
165,588
6,278,762,520
IssuesEvent
2017-07-18 14:58:05
qutebrowser/qutebrowser
https://api.github.com/repos/qutebrowser/qutebrowser
closed
FileNotFoundError in ipc.update_atime
bug needs triage priority: 0 - high
I can reproduce this somewhat reliably by running the tests in a loop on the Debian Jessie and Ubuntu Trusty buildbots: ``` ============================= ERRORS ============================= ________ ERROR at teardown of TestListen.test_atime_update _______ Qt exceptions in virtual methods: ________________________________________________________________________________ File "/home/buildbotx/qutebrowser/qutebrowser/misc/ipc.py", line 368, in update_atime os.utime(path) FileNotFoundError: [Errno 2] No such file or directory ________________________________________________________________________________ ============================ FAILURES ============================ __________________ TestListen.test_atime_update __________________ self = <test_ipc.TestListen object at 0x7f08de08cac8>, qtbot = <pytestqt.qtbot.QtBot object at 0x7f08ddf01a90>, ipc_server = <qutebrowser.misc.ipc.IPCServer object at 0x7f08dec18288> @pytest.mark.posix def test_atime_update(self, qtbot, ipc_server): ipc_server._atime_timer.setInterval(500) # We don't want to wait 6h ipc_server.listen() old_atime = os.stat(ipc_server._server.fullServerName()).st_atime_ns with qtbot.waitSignal(ipc_server._atime_timer.timeout, timeout=2000, raising=True): pass # Make sure the timer is not singleShot with qtbot.waitSignal(ipc_server._atime_timer.timeout, timeout=2000, raising=True): pass > new_atime = os.stat(ipc_server._server.fullServerName()).st_atime_ns E FileNotFoundError: [Errno 2] No such file or directory: '/tmp/qute-test' tests/unit/misc/test_ipc.py:357: FileNotFoundError -------------------------- Captured log -------------------------- ipc.py 193 DEBUG Not calling setSocketOptions ipc.py 204 DEBUG Listening as qute-test ipc.py 367 DEBUG Touching /tmp/qute-test ipc.py 367 DEBUG Touching /tmp/qute-test ```
1.0
FileNotFoundError in ipc.update_atime - I can reproduce this somewhat reliably by running the tests in a loop on the Debian Jessie and Ubuntu Trusty buildbots: ``` ============================= ERRORS ============================= ________ ERROR at teardown of TestListen.test_atime_update _______ Qt exceptions in virtual methods: ________________________________________________________________________________ File "/home/buildbotx/qutebrowser/qutebrowser/misc/ipc.py", line 368, in update_atime os.utime(path) FileNotFoundError: [Errno 2] No such file or directory ________________________________________________________________________________ ============================ FAILURES ============================ __________________ TestListen.test_atime_update __________________ self = <test_ipc.TestListen object at 0x7f08de08cac8>, qtbot = <pytestqt.qtbot.QtBot object at 0x7f08ddf01a90>, ipc_server = <qutebrowser.misc.ipc.IPCServer object at 0x7f08dec18288> @pytest.mark.posix def test_atime_update(self, qtbot, ipc_server): ipc_server._atime_timer.setInterval(500) # We don't want to wait 6h ipc_server.listen() old_atime = os.stat(ipc_server._server.fullServerName()).st_atime_ns with qtbot.waitSignal(ipc_server._atime_timer.timeout, timeout=2000, raising=True): pass # Make sure the timer is not singleShot with qtbot.waitSignal(ipc_server._atime_timer.timeout, timeout=2000, raising=True): pass > new_atime = os.stat(ipc_server._server.fullServerName()).st_atime_ns E FileNotFoundError: [Errno 2] No such file or directory: '/tmp/qute-test' tests/unit/misc/test_ipc.py:357: FileNotFoundError -------------------------- Captured log -------------------------- ipc.py 193 DEBUG Not calling setSocketOptions ipc.py 204 DEBUG Listening as qute-test ipc.py 367 DEBUG Touching /tmp/qute-test ipc.py 367 DEBUG Touching /tmp/qute-test ```
priority
filenotfounderror in ipc update atime i can reproduce this somewhat reliably by running the tests in a loop on the debian jessie and ubuntu trusty buildbots errors error at teardown of testlisten test atime update qt exceptions in virtual methods file home buildbotx qutebrowser qutebrowser misc ipc py line in update atime os utime path filenotfounderror no such file or directory failures testlisten test atime update self qtbot ipc server pytest mark posix def test atime update self qtbot ipc server ipc server atime timer setinterval we don t want to wait ipc server listen old atime os stat ipc server server fullservername st atime ns with qtbot waitsignal ipc server atime timer timeout timeout raising true pass make sure the timer is not singleshot with qtbot waitsignal ipc server atime timer timeout timeout raising true pass new atime os stat ipc server server fullservername st atime ns e filenotfounderror no such file or directory tmp qute test tests unit misc test ipc py filenotfounderror captured log ipc py debug not calling setsocketoptions ipc py debug listening as qute test ipc py debug touching tmp qute test ipc py debug touching tmp qute test
1
626,973
19,848,795,619
IssuesEvent
2022-01-21 09:56:11
TeamSparker/Spark-iOS
https://api.github.com/repos/TeamSparker/Spark-iOS
opened
[FEAT] 스파크 보내기 토스트메세지
Feat 🦹t없e맑은水빈 P1 / Priority High
## 📌 Issue <!-- 이슈에 대해 간략하게 설명해주세요 --> 스파크 보내기 토스트메세지 만들겠습니다 ## 📝 To-do <!-- 진행할 작업에 대해 적어주세요 --> - [ ] 토스트메세지 extension 빼기 - [ ] 토스트메세지 띄우기
1.0
[FEAT] 스파크 보내기 토스트메세지 - ## 📌 Issue <!-- 이슈에 대해 간략하게 설명해주세요 --> 스파크 보내기 토스트메세지 만들겠습니다 ## 📝 To-do <!-- 진행할 작업에 대해 적어주세요 --> - [ ] 토스트메세지 extension 빼기 - [ ] 토스트메세지 띄우기
priority
스파크 보내기 토스트메세지 📌 issue 스파크 보내기 토스트메세지 만들겠습니다 📝 to do 토스트메세지 extension 빼기 토스트메세지 띄우기
1
449,781
12,974,972,083
IssuesEvent
2020-07-21 16:13:50
MatthewSpofford/Multiscale-Statistical-Analysis
https://api.github.com/repos/MatthewSpofford/Multiscale-Statistical-Analysis
opened
PlotData open_file2 does not correctly handle scale discrepencies
bug priority high
If there are scales that may not be contained within all of the opened datasets, then these scales will continue to be appended to the `results_scale` list. Since the order of the scales is based on the files being imported, the `results_scale` list is completely out of order. The `open_file2` function then continues to append on to the `relative_area` and `complexity` lists, causing there to be more scale values than actual area/complexity values, completely disrupting the data table. Blocking: #86
1.0
PlotData open_file2 does not correctly handle scale discrepencies - If there are scales that may not be contained within all of the opened datasets, then these scales will continue to be appended to the `results_scale` list. Since the order of the scales is based on the files being imported, the `results_scale` list is completely out of order. The `open_file2` function then continues to append on to the `relative_area` and `complexity` lists, causing there to be more scale values than actual area/complexity values, completely disrupting the data table. Blocking: #86
priority
plotdata open does not correctly handle scale discrepencies if there are scales that may not be contained within all of the opened datasets then these scales will continue to be appended to the results scale list since the order of the scales is based on the files being imported the results scale list is completely out of order the open function then continues to append on to the relative area and complexity lists causing there to be more scale values than actual area complexity values completely disrupting the data table blocking
1
220,987
7,372,672,110
IssuesEvent
2018-03-13 15:18:39
springboilerplate/spring-rest-security-boilerplate
https://api.github.com/repos/springboilerplate/spring-rest-security-boilerplate
closed
RabbitMQ integration for email processes
High Priority enhancement
Currently, We send emails as normal Java methods, which is bad way. To understand why it is bad and what to do, you can read following links. In this task, we basically use RabbitMQ for our email sending services. What is RabbitMQ and why do we need it? - https://www.cloudamqp.com/blog/2015-05-18-part1-rabbitmq-for-beginners-what-is-rabbitmq.html How to make Spring and RabbitMQ work together? - http://www.bahadirakin.com/amqp-ve-rabbitmq/ - http://www.bahadirakin.com/spring-amqp-ve-rabbitmq/ Official Tutorial - https://www.rabbitmq.com/getstarted.html
1.0
RabbitMQ integration for email processes - Currently, We send emails as normal Java methods, which is bad way. To understand why it is bad and what to do, you can read following links. In this task, we basically use RabbitMQ for our email sending services. What is RabbitMQ and why do we need it? - https://www.cloudamqp.com/blog/2015-05-18-part1-rabbitmq-for-beginners-what-is-rabbitmq.html How to make Spring and RabbitMQ work together? - http://www.bahadirakin.com/amqp-ve-rabbitmq/ - http://www.bahadirakin.com/spring-amqp-ve-rabbitmq/ Official Tutorial - https://www.rabbitmq.com/getstarted.html
priority
rabbitmq integration for email processes currently we send emails as normal java methods which is bad way to understand why it is bad and what to do you can read following links in this task we basically use rabbitmq for our email sending services what is rabbitmq and why do we need it how to make spring and rabbitmq work together official tutorial
1
472,290
13,622,330,550
IssuesEvent
2020-09-24 03:18:34
SE761Team4/jabref
https://api.github.com/repos/SE761Team4/jabref
closed
[4pt] Make BibTeX to MindMap parser
backend blocker high priority
**User Story** <!--As a [user role], I'd like to [do something], so that [some goal].--> As a developer, I would like to map BibTex to JSON so that I can parse information to JabMap **Acceptance Criteria** - BibTeX data can be parsed via HTTP to JabMap **Notes** --- **Why is this feature needed? Please describe the problem your requested feature wants to solve** <!-- Describe what the problem is. Ex. I'm always frustrated when ... --> **Describe the solution you'd like** <!--Describe what you want to happen -->
1.0
[4pt] Make BibTeX to MindMap parser - **User Story** <!--As a [user role], I'd like to [do something], so that [some goal].--> As a developer, I would like to map BibTex to JSON so that I can parse information to JabMap **Acceptance Criteria** - BibTeX data can be parsed via HTTP to JabMap **Notes** --- **Why is this feature needed? Please describe the problem your requested feature wants to solve** <!-- Describe what the problem is. Ex. I'm always frustrated when ... --> **Describe the solution you'd like** <!--Describe what you want to happen -->
priority
make bibtex to mindmap parser user story as a developer i would like to map bibtex to json so that i can parse information to jabmap acceptance criteria bibtex data can be parsed via http to jabmap notes why is this feature needed please describe the problem your requested feature wants to solve describe the solution you d like
1
171,803
6,494,608,673
IssuesEvent
2017-08-21 22:48:57
cranndarach/lifetracker
https://api.github.com/repos/cranndarach/lifetracker
closed
[docs] forms.cson needs documenting
priority/high task-size/medium
There is a wiki page for making and editing forms, but it is empty. As soon as possible, it should be explained in plain language so that LifeTracker is still accessible to non-programmers.
1.0
[docs] forms.cson needs documenting - There is a wiki page for making and editing forms, but it is empty. As soon as possible, it should be explained in plain language so that LifeTracker is still accessible to non-programmers.
priority
forms cson needs documenting there is a wiki page for making and editing forms but it is empty as soon as possible it should be explained in plain language so that lifetracker is still accessible to non programmers
1
86,427
3,715,118,741
IssuesEvent
2016-03-03 00:02:35
UCHIC/CKANDev
https://api.github.com/repos/UCHIC/CKANDev
closed
Photo Upload
HIGH PRIORITY question
Are there ways to upload a picture other than the Gravatar website? Users may want to add photos without using that website functionality.
1.0
Photo Upload - Are there ways to upload a picture other than the Gravatar website? Users may want to add photos without using that website functionality.
priority
photo upload are there ways to upload a picture other than the gravatar website users may want to add photos without using that website functionality
1
535,893
15,700,641,130
IssuesEvent
2021-03-26 10:07:13
Systems-Learning-and-Development-Lab/MMM
https://api.github.com/repos/Systems-Learning-and-Development-Lab/MMM
opened
Remove mark not always works
priority-high
. המחיקה לא תמיד עובדת, אני לא ממש מצליחה לשחזר בדיוק. הפעולות שאני עושה הן לשים סמן, להריץ, למחוק, לשנות סמן, להריץ, למחוק את הסמן. ברוב המקרים המחיקה לא עובדת. @Ron-Teller
1.0
Remove mark not always works - . המחיקה לא תמיד עובדת, אני לא ממש מצליחה לשחזר בדיוק. הפעולות שאני עושה הן לשים סמן, להריץ, למחוק, לשנות סמן, להריץ, למחוק את הסמן. ברוב המקרים המחיקה לא עובדת. @Ron-Teller
priority
remove mark not always works המחיקה לא תמיד עובדת אני לא ממש מצליחה לשחזר בדיוק הפעולות שאני עושה הן לשים סמן להריץ למחוק לשנות סמן להריץ למחוק את הסמן ברוב המקרים המחיקה לא עובדת ron teller
1
558,834
16,543,398,181
IssuesEvent
2021-05-27 19:59:30
ampproject/amphtml
https://api.github.com/repos/ampproject/amphtml
opened
AMPHTML ads using gwd-animation are broken
Component: AMPHTML ads P1: High Priority Type: Bug
### Description Regression introduced with the launch of no-signing rendering path. ### Reproduction Steps Example page that will render as a blank ad: https://gist.github.com/calebcordry/8194102522f2b50821ea48d70bc2583e ### Relevant Logs _No response_ ### Browser(s) Affected Chrome, Firefox, Safari, Edge, UC Browser ### OS(s) Affected All ### Device(s) Affected All ### AMP Version Affected 2104170104000 and after
1.0
AMPHTML ads using gwd-animation are broken - ### Description Regression introduced with the launch of no-signing rendering path. ### Reproduction Steps Example page that will render as a blank ad: https://gist.github.com/calebcordry/8194102522f2b50821ea48d70bc2583e ### Relevant Logs _No response_ ### Browser(s) Affected Chrome, Firefox, Safari, Edge, UC Browser ### OS(s) Affected All ### Device(s) Affected All ### AMP Version Affected 2104170104000 and after
priority
amphtml ads using gwd animation are broken description regression introduced with the launch of no signing rendering path reproduction steps example page that will render as a blank ad relevant logs no response browser s affected chrome firefox safari edge uc browser os s affected all device s affected all amp version affected and after
1
675,854
23,109,762,005
IssuesEvent
2022-07-27 12:03:00
fadeinside/s3air-achievements-plus
https://api.github.com/repos/fadeinside/s3air-achievements-plus
closed
Achievement Guides
Type: Feature Priority: Highest
**Description** \- **Screenshots/References** \- **Additional context** \-
1.0
Achievement Guides - **Description** \- **Screenshots/References** \- **Additional context** \-
priority
achievement guides description screenshots references additional context
1
679,503
23,235,340,714
IssuesEvent
2022-08-03 11:16:13
chloebrett/mlvet
https://api.github.com/repos/chloebrett/mlvet
closed
Browsing for a file to import and then cancelling causes a blank selected file to appear
high priority
<img width="685" alt="Screen Shot 2022-05-28 at 3 23 14 pm" src="https://user-images.githubusercontent.com/6735055/170811415-e13b0d83-2935-4afb-a77a-51ba4bd069bc.png"> The transcription button then gets enabled, but clicking it doesn't do anything
1.0
Browsing for a file to import and then cancelling causes a blank selected file to appear - <img width="685" alt="Screen Shot 2022-05-28 at 3 23 14 pm" src="https://user-images.githubusercontent.com/6735055/170811415-e13b0d83-2935-4afb-a77a-51ba4bd069bc.png"> The transcription button then gets enabled, but clicking it doesn't do anything
priority
browsing for a file to import and then cancelling causes a blank selected file to appear img width alt screen shot at pm src the transcription button then gets enabled but clicking it doesn t do anything
1
748,851
26,140,449,085
IssuesEvent
2022-12-29 17:37:26
netdata/netdata-cloud
https://api.github.com/repos/netdata/netdata-cloud
closed
View all Active alerts in a user friendly format (table) from a mobile device.
priority/high Customer Success cloud-frontend alerts-team Q1 GOAL
##Issue As a user of Netdata Cloud, I would like to view all active alerts in my Space/Room/Node on my mobile device. ##Use-case On receiving a notification on the mobile, the user needs to be able to see the specific alert and also other active alerts on a mobile device.
1.0
View all Active alerts in a user friendly format (table) from a mobile device. - ##Issue As a user of Netdata Cloud, I would like to view all active alerts in my Space/Room/Node on my mobile device. ##Use-case On receiving a notification on the mobile, the user needs to be able to see the specific alert and also other active alerts on a mobile device.
priority
view all active alerts in a user friendly format table from a mobile device issue as a user of netdata cloud i would like to view all active alerts in my space room node on my mobile device use case on receiving a notification on the mobile the user needs to be able to see the specific alert and also other active alerts on a mobile device
1
155,819
5,961,329,165
IssuesEvent
2017-05-29 16:56:13
WatzekDigitalInitiatives/fitbit-ror
https://api.github.com/repos/WatzekDigitalInitiatives/fitbit-ror
closed
Cannot join private teams (extreme high priority)
front end high priority
Users should see `user_teams/new.html.erb` when joining a private team they are not part of
1.0
Cannot join private teams (extreme high priority) - Users should see `user_teams/new.html.erb` when joining a private team they are not part of
priority
cannot join private teams extreme high priority users should see user teams new html erb when joining a private team they are not part of
1
793,842
28,013,023,569
IssuesEvent
2023-03-27 20:08:49
RadeonOpenCompute/ROCm-CompilerSupport
https://api.github.com/repos/RadeonOpenCompute/ROCm-CompilerSupport
closed
comgr lacks MT safety
enhancement priority_high
It seems like comgr is not working properly in multithreaded contexts. This prevents it from full-blown usage in MIOpen. ## How to reproduce Prerequisites: A linux machine with gfx900/906/908 GPU card installed (Radeon VII preferred), vanilla ROCm 3.3. - (1) Pull the https://github.com/ROCmSoftwarePlatform/MIOpen/tree/comgr-attach branch from MIOpen repo and checkout https://github.com/ROCmSoftwarePlatform/MIOpen/commit/9a3a2e7142ea7651c9e0056ed8a3a1d860a8d702 - (2) Build MIOpen library and MIOpenDriver with HIP backend. Use developer build, `cmake -DBUILD_DEV=On...` (see [here](https://github.com/ROCmSoftwarePlatform/MIOpen/blob/develop/README.md) for detailed instruction), then make `MIOpenDriver -j`. Do not install. - (3) Optional: modify environment to skip non-relevant kernels in MIOpen. This would speed-up execution and clean MIOpen logs from non-relevant information. ``` export MIOPEN_DEBUG_CONV_GEMM=0 export MIOPEN_DEBUG_CONV_FFT=0 export MIOPEN_DEBUG_CONV_IMPLICIT_GEMM=0 export MIOPEN_DEBUG_CONV_SCGEMM=0 export MIOPEN_DEBUG_CONV_WINOGRAD=0 export MIOPEN_DEBUG_CONV_DIRECT=1 export MIOPEN_DEBUG_GCN_ASM_KERNELS=0 export MIOPEN_DEBUG_OPENCL_CONVOLUTIONS=1 ``` - (4) Run the following command and make sure that it works fine: ``` $ ./bin/MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1 MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1 MIOpen Forward Conv. Algorithm: 1, Solution: 9/ConvOclDirectFwdGen GPU Kernel Time Forward Conv. Elapsed: 0.307048 ms (average) stats: name, n, c, ho, wo, x, y, k, flopCnt, bytesRead, bytesWritten, GFLOPs, GB/s, timeMs stats: fwd-conv5x20u2, 8, 1, 79, 341, 5, 20, 32, 1379276800, 3619200, 27585536, 4492, 102, 0.307048 ``` - (5) Optional: re-run the same command prefixed with `MIOPEN_LOG_LEVEL=5` and make sure that the following lines present on the console: ``` Info [EvaluateInvokers] ConvOclDirectFwdGen: MIOpenCDFGen4: 0.323367 < 3.40282e+38 Info [EvaluateInvokers] ConvOclDirectFwd: MIOpenConvUni: 1.67044 >= 0.323367 ... Info [FindConvFwdAlgorithm] FW Chosen Algorithm: ConvOclDirectFwdGen , 0, 0.323367 ``` These upper two lines tell us that two kernels (MIOpenCDFGen4 and MIOpenConvUni) obtained from two Solutions (ConvOclDirectFwdGen and ConvOclDirectFwd) were built and run by the library. The third line says that ConvOclDirectFwdGen (the fastest) was selected. The library builds the kernels in parallel; you may wish to export `MIOPEN_ENABLE_LOGGING_MPMT=1` and `MIOPEN_LOG_LEVEL=6` and see these details in the console log. - (6) Now it's time to use comgr. Clean build directory (you can skip cleaning the build directory, but I am not sure if this would rebuild all the necessary components). Add `-DMIOPEN_USE_COMGR=On` to cmake command line (see [here](https://github.com/ROCmSoftwarePlatform/MIOpen/pull/177#issue-408736873) for details) and rebuild library and driver, - (7) Re-run the command from step (4). Output: ``` $ ./bin/MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1 MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1 clang (LLVM option parsing): for the --amdgpu-early-inline-all option: may only occur zero or one times! clang (LLVM option parsing): for the --amdgpu-prelink option: may only occur zero or one times! clang (LLVM option parsing): for the --amdgpu-internalize-symbols option: may only occur zero or one times! clang (LLVM option parsing): for the --pgo-warn-misexpect option: may only occur zero or one times! ``` ### :red_circle: There are two problems: - Build does not succeed. - Build kills the driver. Neither error codes returned to MIOpen, nor exceptions thrown. You can disable MT builds in MIOpen by the following env setting: ``` export MIOPEN_COMPILE_PARALLEL_LEVEL=1 ``` and see that driver works normally in this case.
1.0
comgr lacks MT safety - It seems like comgr is not working properly in multithreaded contexts. This prevents it from full-blown usage in MIOpen. ## How to reproduce Prerequisites: A linux machine with gfx900/906/908 GPU card installed (Radeon VII preferred), vanilla ROCm 3.3. - (1) Pull the https://github.com/ROCmSoftwarePlatform/MIOpen/tree/comgr-attach branch from MIOpen repo and checkout https://github.com/ROCmSoftwarePlatform/MIOpen/commit/9a3a2e7142ea7651c9e0056ed8a3a1d860a8d702 - (2) Build MIOpen library and MIOpenDriver with HIP backend. Use developer build, `cmake -DBUILD_DEV=On...` (see [here](https://github.com/ROCmSoftwarePlatform/MIOpen/blob/develop/README.md) for detailed instruction), then make `MIOpenDriver -j`. Do not install. - (3) Optional: modify environment to skip non-relevant kernels in MIOpen. This would speed-up execution and clean MIOpen logs from non-relevant information. ``` export MIOPEN_DEBUG_CONV_GEMM=0 export MIOPEN_DEBUG_CONV_FFT=0 export MIOPEN_DEBUG_CONV_IMPLICIT_GEMM=0 export MIOPEN_DEBUG_CONV_SCGEMM=0 export MIOPEN_DEBUG_CONV_WINOGRAD=0 export MIOPEN_DEBUG_CONV_DIRECT=1 export MIOPEN_DEBUG_GCN_ASM_KERNELS=0 export MIOPEN_DEBUG_OPENCL_CONVOLUTIONS=1 ``` - (4) Run the following command and make sure that it works fine: ``` $ ./bin/MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1 MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1 MIOpen Forward Conv. Algorithm: 1, Solution: 9/ConvOclDirectFwdGen GPU Kernel Time Forward Conv. Elapsed: 0.307048 ms (average) stats: name, n, c, ho, wo, x, y, k, flopCnt, bytesRead, bytesWritten, GFLOPs, GB/s, timeMs stats: fwd-conv5x20u2, 8, 1, 79, 341, 5, 20, 32, 1379276800, 3619200, 27585536, 4492, 102, 0.307048 ``` - (5) Optional: re-run the same command prefixed with `MIOPEN_LOG_LEVEL=5` and make sure that the following lines present on the console: ``` Info [EvaluateInvokers] ConvOclDirectFwdGen: MIOpenCDFGen4: 0.323367 < 3.40282e+38 Info [EvaluateInvokers] ConvOclDirectFwd: MIOpenConvUni: 1.67044 >= 0.323367 ... Info [FindConvFwdAlgorithm] FW Chosen Algorithm: ConvOclDirectFwdGen , 0, 0.323367 ``` These upper two lines tell us that two kernels (MIOpenCDFGen4 and MIOpenConvUni) obtained from two Solutions (ConvOclDirectFwdGen and ConvOclDirectFwd) were built and run by the library. The third line says that ConvOclDirectFwdGen (the fastest) was selected. The library builds the kernels in parallel; you may wish to export `MIOPEN_ENABLE_LOGGING_MPMT=1` and `MIOPEN_LOG_LEVEL=6` and see these details in the console log. - (6) Now it's time to use comgr. Clean build directory (you can skip cleaning the build directory, but I am not sure if this would rebuild all the necessary components). Add `-DMIOPEN_USE_COMGR=On` to cmake command line (see [here](https://github.com/ROCmSoftwarePlatform/MIOpen/pull/177#issue-408736873) for details) and rebuild library and driver, - (7) Re-run the command from step (4). Output: ``` $ ./bin/MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1 MIOpenDriver conv -x 20 -y 5 -W 700 -H 161 -c 1 -n 8 -k 32 -p 0 -q 0 -u 2 -v 2 -l 1 -j 1 -g 1 -F 1 -V 0 -s 0 -t 1 -i 1 clang (LLVM option parsing): for the --amdgpu-early-inline-all option: may only occur zero or one times! clang (LLVM option parsing): for the --amdgpu-prelink option: may only occur zero or one times! clang (LLVM option parsing): for the --amdgpu-internalize-symbols option: may only occur zero or one times! clang (LLVM option parsing): for the --pgo-warn-misexpect option: may only occur zero or one times! ``` ### :red_circle: There are two problems: - Build does not succeed. - Build kills the driver. Neither error codes returned to MIOpen, nor exceptions thrown. You can disable MT builds in MIOpen by the following env setting: ``` export MIOPEN_COMPILE_PARALLEL_LEVEL=1 ``` and see that driver works normally in this case.
priority
comgr lacks mt safety it seems like comgr is not working properly in multithreaded contexts this prevents it from full blown usage in miopen how to reproduce prerequisites a linux machine with gpu card installed radeon vii preferred vanilla rocm pull the branch from miopen repo and checkout build miopen library and miopendriver with hip backend use developer build cmake dbuild dev on see for detailed instruction then make miopendriver j do not install optional modify environment to skip non relevant kernels in miopen this would speed up execution and clean miopen logs from non relevant information export miopen debug conv gemm export miopen debug conv fft export miopen debug conv implicit gemm export miopen debug conv scgemm export miopen debug conv winograd export miopen debug conv direct export miopen debug gcn asm kernels export miopen debug opencl convolutions run the following command and make sure that it works fine bin miopendriver conv x y w h c n k p q u v l j g f v s t i miopendriver conv x y w h c n k p q u v l j g f v s t i miopen forward conv algorithm solution convocldirectfwdgen gpu kernel time forward conv elapsed ms average stats name n c ho wo x y k flopcnt bytesread byteswritten gflops gb s timems stats fwd optional re run the same command prefixed with miopen log level and make sure that the following lines present on the console info convocldirectfwdgen info convocldirectfwd miopenconvuni info fw chosen algorithm convocldirectfwdgen these upper two lines tell us that two kernels and miopenconvuni obtained from two solutions convocldirectfwdgen and convocldirectfwd were built and run by the library the third line says that convocldirectfwdgen the fastest was selected the library builds the kernels in parallel you may wish to export miopen enable logging mpmt and miopen log level and see these details in the console log now it s time to use comgr clean build directory you can skip cleaning the build directory but i am not sure if this would rebuild all the necessary components add dmiopen use comgr on to cmake command line see for details and rebuild library and driver re run the command from step output bin miopendriver conv x y w h c n k p q u v l j g f v s t i miopendriver conv x y w h c n k p q u v l j g f v s t i clang llvm option parsing for the amdgpu early inline all option may only occur zero or one times clang llvm option parsing for the amdgpu prelink option may only occur zero or one times clang llvm option parsing for the amdgpu internalize symbols option may only occur zero or one times clang llvm option parsing for the pgo warn misexpect option may only occur zero or one times red circle there are two problems build does not succeed build kills the driver neither error codes returned to miopen nor exceptions thrown you can disable mt builds in miopen by the following env setting export miopen compile parallel level and see that driver works normally in this case
1
755,492
26,430,515,971
IssuesEvent
2023-01-14 19:04:00
banco-alimentar/alimentestaideia.pt
https://api.github.com/repos/banco-alimentar/alimentestaideia.pt
opened
Management of donors (outside the site)
High.Priority
The food bank manages all donations that are not made in the site, using a “old” access database. They register the information about all the donors, and each donation. Then produce a invoice (a word doc) that is now not legal anymore. I thought that we have “everything” to replace this access with our site. For that we would need to be able to have a backoffice to enter the donor information, and add donations. One donor might have many donations, so we don’t need to reenter the donor info everytime we add a donation.” 2 times per year we need to export a list of donors, so we can send them a letter. If we do move the solution, I would need to also bring the past information to the site to provide continuity. ![image](https://user-images.githubusercontent.com/42462/212491625-2bf8f011-b790-45cd-99b3-f6200627dcfb.png) For Each donation: ![image](https://user-images.githubusercontent.com/42462/212491630-2ce85183-0ca6-4ca6-b269-7b069e938511.png) ### Fields: - Donor - DonationDate - DonationAmount - TypeOfDonation - Aleatorio - Espontanio - Personalizado - Regular - Tribunal - Um amigo - Site ### PaymentMethods - Cheque - Dep. Conta - Multibanco - Numerário - Vale Postal - MbWay - Site (the current and existing way!?) - Paypal - CreditCard This feature should be accessed in the Backoffice or it could be implemented in the Backoffice Powerapp.
1.0
Management of donors (outside the site) - The food bank manages all donations that are not made in the site, using a “old” access database. They register the information about all the donors, and each donation. Then produce a invoice (a word doc) that is now not legal anymore. I thought that we have “everything” to replace this access with our site. For that we would need to be able to have a backoffice to enter the donor information, and add donations. One donor might have many donations, so we don’t need to reenter the donor info everytime we add a donation.” 2 times per year we need to export a list of donors, so we can send them a letter. If we do move the solution, I would need to also bring the past information to the site to provide continuity. ![image](https://user-images.githubusercontent.com/42462/212491625-2bf8f011-b790-45cd-99b3-f6200627dcfb.png) For Each donation: ![image](https://user-images.githubusercontent.com/42462/212491630-2ce85183-0ca6-4ca6-b269-7b069e938511.png) ### Fields: - Donor - DonationDate - DonationAmount - TypeOfDonation - Aleatorio - Espontanio - Personalizado - Regular - Tribunal - Um amigo - Site ### PaymentMethods - Cheque - Dep. Conta - Multibanco - Numerário - Vale Postal - MbWay - Site (the current and existing way!?) - Paypal - CreditCard This feature should be accessed in the Backoffice or it could be implemented in the Backoffice Powerapp.
priority
management of donors outside the site the food bank manages all donations that are not made in the site using a “old” access database they register the information about all the donors and each donation then produce a invoice a word doc that is now not legal anymore i thought that we have “everything” to replace this access with our site for that we would need to be able to have a backoffice to enter the donor information and add donations one donor might have many donations so we don’t need to reenter the donor info everytime we add a donation ” times per year we need to export a list of donors so we can send them a letter if we do move the solution i would need to also bring the past information to the site to provide continuity for each donation fields donor donationdate donationamount typeofdonation aleatorio espontanio personalizado regular tribunal um amigo site paymentmethods cheque dep conta multibanco numerário vale postal mbway site the current and existing way paypal creditcard this feature should be accessed in the backoffice or it could be implemented in the backoffice powerapp
1
89,633
3,798,164,861
IssuesEvent
2016-03-23 11:17:04
bedita/bedita
https://api.github.com/repos/bedita/bedita
closed
cache: add nickname and object_type_id object cache
Priority - High Topic - Core Topic - Frontend Type - Enhancement
Add cache (using object cache or similar method) for two intensively used methods in frontends: * `BEObject::getType / findObjectTypeId` - performs a query like ```sql SELECT `BEObject`.`object_type_id` FROM `objects` AS `BEObject` WHERE `BEObject`.`id` = ### LIMIT 1 ``` * `BEObject::getIdFromNickname` - performs a query like ```sql SELECT `BEObject`.`id` FROM `objects` AS `BEObject` WHERE `nickname` = '###' AND status = ('on') LIMIT 1 ```
1.0
cache: add nickname and object_type_id object cache - Add cache (using object cache or similar method) for two intensively used methods in frontends: * `BEObject::getType / findObjectTypeId` - performs a query like ```sql SELECT `BEObject`.`object_type_id` FROM `objects` AS `BEObject` WHERE `BEObject`.`id` = ### LIMIT 1 ``` * `BEObject::getIdFromNickname` - performs a query like ```sql SELECT `BEObject`.`id` FROM `objects` AS `BEObject` WHERE `nickname` = '###' AND status = ('on') LIMIT 1 ```
priority
cache add nickname and object type id object cache add cache using object cache or similar method for two intensively used methods in frontends beobject gettype findobjecttypeid performs a query like sql select beobject object type id from objects as beobject where beobject id limit beobject getidfromnickname performs a query like sql select beobject id from objects as beobject where nickname and status on limit
1
757,804
26,529,508,954
IssuesEvent
2023-01-19 11:20:57
ballerina-platform/ballerina-standard-library
https://api.github.com/repos/ballerina-platform/ballerina-standard-library
opened
GraalVM Check is failing in FTP
Priority/High Type/Task module/ftp graalvm
**Description:** > $Subject > Part of https://github.com/ballerina-platform/ballerina-standard-library/issues/3755 GraalVM Check on ftp is failing due to `org.apache.sshd` dependency which is used in `test-utils`. And replacing this dependency or using a docker image for testing may solve this issue since I could not find a solution wrt to GraalVM native-image. Need to check whether it is possible or not This separate [`ftp-mock-server`](https://github.com/TharmiganK/ftp-mock-server) project may help to engage docker.
1.0
GraalVM Check is failing in FTP - **Description:** > $Subject > Part of https://github.com/ballerina-platform/ballerina-standard-library/issues/3755 GraalVM Check on ftp is failing due to `org.apache.sshd` dependency which is used in `test-utils`. And replacing this dependency or using a docker image for testing may solve this issue since I could not find a solution wrt to GraalVM native-image. Need to check whether it is possible or not This separate [`ftp-mock-server`](https://github.com/TharmiganK/ftp-mock-server) project may help to engage docker.
priority
graalvm check is failing in ftp description subject part of graalvm check on ftp is failing due to org apache sshd dependency which is used in test utils and replacing this dependency or using a docker image for testing may solve this issue since i could not find a solution wrt to graalvm native image need to check whether it is possible or not this separate project may help to engage docker
1
154,082
5,908,775,671
IssuesEvent
2017-05-19 21:22:53
Polymer/polymer-build
https://api.github.com/repos/Polymer/polymer-build
closed
[1.2.0] Unable to get document FILENAME.HTML:
Priority: High Status: Accepted Type: Bug
### Description My build script simply attempts to build a single-file version of Polymer 2.0. There are no custom files imported. After upgrading to 1.2.0, my Gulp build script fails to build my Polymer bundle. The message is > Unable to get document polymer-bundle.html: 1.1.0 works as I would expect. I tracked it down to `deps-index.js` inside of the `buildDepsIndex()` function. I realize this is a dependency of this project, so if I need to move this report, just let me know. ### Versions & Environment - polymer-build: v1.2.0 - node: v7.9.0 - Operating System: OSX 10.12.4 #### Steps to Reproduce My `polymer.json` file: ``` { "entrypoint": "index.html", "fragments": [ "polymer-bundle.html" ], "sources": [], "extraDependencies": [], "builds": [ { "name": "bundled", "bundle": true, "js": { "minify": true }, "css": { "minify": true }, "html": { "minify": true } } ], "lint": { "rules": [ "polymer-2-hybrid" ] } } ``` My `polymer-bundle.html` file: ``` <!-- polymer library --> <link rel="import" href="bower_components/polymer/polymer.html"> <!-- extra polymer behaviors --> <link rel="import" href="bower_components/polymer/lib/mixins/gesture-event-listeners.html"> ``` #### Expected Results A bundled version of Polymer and its imported files. #### Actual Results An error is thrown and my build fails.
1.0
[1.2.0] Unable to get document FILENAME.HTML: - ### Description My build script simply attempts to build a single-file version of Polymer 2.0. There are no custom files imported. After upgrading to 1.2.0, my Gulp build script fails to build my Polymer bundle. The message is > Unable to get document polymer-bundle.html: 1.1.0 works as I would expect. I tracked it down to `deps-index.js` inside of the `buildDepsIndex()` function. I realize this is a dependency of this project, so if I need to move this report, just let me know. ### Versions & Environment - polymer-build: v1.2.0 - node: v7.9.0 - Operating System: OSX 10.12.4 #### Steps to Reproduce My `polymer.json` file: ``` { "entrypoint": "index.html", "fragments": [ "polymer-bundle.html" ], "sources": [], "extraDependencies": [], "builds": [ { "name": "bundled", "bundle": true, "js": { "minify": true }, "css": { "minify": true }, "html": { "minify": true } } ], "lint": { "rules": [ "polymer-2-hybrid" ] } } ``` My `polymer-bundle.html` file: ``` <!-- polymer library --> <link rel="import" href="bower_components/polymer/polymer.html"> <!-- extra polymer behaviors --> <link rel="import" href="bower_components/polymer/lib/mixins/gesture-event-listeners.html"> ``` #### Expected Results A bundled version of Polymer and its imported files. #### Actual Results An error is thrown and my build fails.
priority
unable to get document filename html description my build script simply attempts to build a single file version of polymer there are no custom files imported after upgrading to my gulp build script fails to build my polymer bundle the message is unable to get document polymer bundle html works as i would expect i tracked it down to deps index js inside of the builddepsindex function i realize this is a dependency of this project so if i need to move this report just let me know versions environment polymer build node operating system osx steps to reproduce my polymer json file entrypoint index html fragments polymer bundle html sources extradependencies builds name bundled bundle true js minify true css minify true html minify true lint rules polymer hybrid my polymer bundle html file expected results a bundled version of polymer and its imported files actual results an error is thrown and my build fails
1
789,626
27,798,749,802
IssuesEvent
2023-03-17 14:24:27
PakizeBozkurt/hugsforbugs
https://api.github.com/repos/PakizeBozkurt/hugsforbugs
closed
Contact trainee via email
enhancement Part2 new high priority
## Story As a trainee I want to contact other trainees ## Acceptance Criteria When I see other trainees availability, I click on email. That goes to contact page where I can email them. I can see success screen when finished ## Implementation Notes 1. Link to the trainee list page (card with trainee details) 2. onClick => 'send calendar request', pass (user logged in) and tranee data via local storage. 3. Use ender's demo email implementation 4. Extra - (not essential) = display a success banner when done - https://getbootstrap.com/docs/4.0/components/alerts/ ![image](https://user-images.githubusercontent.com/38047481/222129425-d8bcc4e8-dbdd-460f-ba72-a6820c676fe8.png) ## Checklist: - [ Y ] I have defined the user story - [ Y ] I have defined some acceptance criteria - [ Y ] I have included any implementation notes - [ Y ] I have included all relevant links
1.0
Contact trainee via email - ## Story As a trainee I want to contact other trainees ## Acceptance Criteria When I see other trainees availability, I click on email. That goes to contact page where I can email them. I can see success screen when finished ## Implementation Notes 1. Link to the trainee list page (card with trainee details) 2. onClick => 'send calendar request', pass (user logged in) and tranee data via local storage. 3. Use ender's demo email implementation 4. Extra - (not essential) = display a success banner when done - https://getbootstrap.com/docs/4.0/components/alerts/ ![image](https://user-images.githubusercontent.com/38047481/222129425-d8bcc4e8-dbdd-460f-ba72-a6820c676fe8.png) ## Checklist: - [ Y ] I have defined the user story - [ Y ] I have defined some acceptance criteria - [ Y ] I have included any implementation notes - [ Y ] I have included all relevant links
priority
contact trainee via email story as a trainee i want to contact other trainees acceptance criteria when i see other trainees availability i click on email that goes to contact page where i can email them i can see success screen when finished implementation notes link to the trainee list page card with trainee details onclick send calendar request pass user logged in and tranee data via local storage use ender s demo email implementation extra not essential display a success banner when done checklist i have defined the user story i have defined some acceptance criteria i have included any implementation notes i have included all relevant links
1
717,462
24,676,275,673
IssuesEvent
2022-10-18 17:16:28
statechannels/go-nitro
https://api.github.com/repos/statechannels/go-nitro
closed
generate-adjudicator-bindings.sh is broken
🐛 bug :bangbang: high-priority
I think we may have merged some broken code, here. - The "check bindings" action did not run on this PR because we made no changes to the contracts: https://github.com/statechannels/go-nitro/blob/d9f17f0d93c09514f6f5cc113dee7f6882c040b8/.github/workflows/bindings-check.yml#L6-L7 (We should run the action when the shell script is modified, too) - the "check bindings" action does not run on `main`, so we haven't yet detected this problem (but soon will on a pull request) _Originally posted by @geoknee in https://github.com/statechannels/go-nitro/issues/921#issuecomment-1282345894_
1.0
generate-adjudicator-bindings.sh is broken - I think we may have merged some broken code, here. - The "check bindings" action did not run on this PR because we made no changes to the contracts: https://github.com/statechannels/go-nitro/blob/d9f17f0d93c09514f6f5cc113dee7f6882c040b8/.github/workflows/bindings-check.yml#L6-L7 (We should run the action when the shell script is modified, too) - the "check bindings" action does not run on `main`, so we haven't yet detected this problem (but soon will on a pull request) _Originally posted by @geoknee in https://github.com/statechannels/go-nitro/issues/921#issuecomment-1282345894_
priority
generate adjudicator bindings sh is broken i think we may have merged some broken code here the check bindings action did not run on this pr because we made no changes to the contracts we should run the action when the shell script is modified too the check bindings action does not run on main so we haven t yet detected this problem but soon will on a pull request originally posted by geoknee in
1
203,528
7,065,311,575
IssuesEvent
2018-01-06 18:28:31
mattbdean/Helium
https://api.github.com/repos/mattbdean/Helium
closed
Inserting zero entries to a part table
bug high priority
If I have a master__part table (e.g. `session__trial`), if I try to insert only to the `session` table with nothing to the `trial` part table I get the following error: ``` bash { message: 'generated invalid SQL', data: { session: [ [Object] ], session__trial: [] } } { Error: ER_PARSE_ERROR: You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near '' at line 1 at Query.Sequence._packetToError (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/sequences/Sequence.js:52:14) at Query.ErrorPacket (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/sequences/Query.js:77:18) at Protocol._parsePacket (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:279:23) at Parser.write (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Parser.js:76:12) at Protocol.write (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:39:16) at Socket.<anonymous> (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/Connection.js:103:28) at emitOne (events.js:116:13) at Socket.emit (events.js:211:7) at addChunk (_stream_readable.js:263:12) at readableAddChunk (_stream_readable.js:250:11) at Socket.Readable.push (_stream_readable.js:208:10) at TCP.onread (net.js:594:20) -------------------- at Protocol._enqueue (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:145:48) at PoolConnection.query (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/Connection.js:208:25) at /Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/helper.js:16:24 at Promise._execute (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/debuggability.js:303:9) at Promise._resolveFromExecutor (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/promise.js:483:18) at new Promise (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/promise.js:79:10) at PoolConnection.promiseCallback (/Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/helper.js:6:10) at poolConnection.connection.query (/Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/connection.js:64:28) at QueryHelper.<anonymous> (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:24:54) at Generator.next (<anonymous>) at /Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:7:71 at new Promise (<anonymous>) at __awaiter (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:3:12) at QueryHelper.execute (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:22:16) at SchemaDao.<anonymous> (/Users/moharb/Documents/Repos/Helium/dist/routes/api/schemas/schema.dao.js:177:39) at Generator.next (<anonymous>) code: 'ER_PARSE_ERROR', errno: 1064, sqlMessage: 'You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near \'\' at line 1', sqlState: '42000', index: 0, sql: 'INSERT INTO `boazmohar_experiment`.`session__trial`' } PUT /api/v1/schemas/boazmohar_experiment/data 500 104.906 ms - 3155 ``` If I add a single entry to the `trial` table the form works, also the same inset to the `session` table works from the dj python API.
1.0
Inserting zero entries to a part table - If I have a master__part table (e.g. `session__trial`), if I try to insert only to the `session` table with nothing to the `trial` part table I get the following error: ``` bash { message: 'generated invalid SQL', data: { session: [ [Object] ], session__trial: [] } } { Error: ER_PARSE_ERROR: You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near '' at line 1 at Query.Sequence._packetToError (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/sequences/Sequence.js:52:14) at Query.ErrorPacket (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/sequences/Query.js:77:18) at Protocol._parsePacket (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:279:23) at Parser.write (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Parser.js:76:12) at Protocol.write (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:39:16) at Socket.<anonymous> (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/Connection.js:103:28) at emitOne (events.js:116:13) at Socket.emit (events.js:211:7) at addChunk (_stream_readable.js:263:12) at readableAddChunk (_stream_readable.js:250:11) at Socket.Readable.push (_stream_readable.js:208:10) at TCP.onread (net.js:594:20) -------------------- at Protocol._enqueue (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/protocol/Protocol.js:145:48) at PoolConnection.query (/Users/moharb/Documents/Repos/Helium/node_modules/mysql/lib/Connection.js:208:25) at /Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/helper.js:16:24 at Promise._execute (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/debuggability.js:303:9) at Promise._resolveFromExecutor (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/promise.js:483:18) at new Promise (/Users/moharb/Documents/Repos/Helium/node_modules/bluebird/js/release/promise.js:79:10) at PoolConnection.promiseCallback (/Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/helper.js:6:10) at poolConnection.connection.query (/Users/moharb/Documents/Repos/Helium/node_modules/promise-mysql/lib/connection.js:64:28) at QueryHelper.<anonymous> (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:24:54) at Generator.next (<anonymous>) at /Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:7:71 at new Promise (<anonymous>) at __awaiter (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:3:12) at QueryHelper.execute (/Users/moharb/Documents/Repos/Helium/dist/db/query-helper.js:22:16) at SchemaDao.<anonymous> (/Users/moharb/Documents/Repos/Helium/dist/routes/api/schemas/schema.dao.js:177:39) at Generator.next (<anonymous>) code: 'ER_PARSE_ERROR', errno: 1064, sqlMessage: 'You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near \'\' at line 1', sqlState: '42000', index: 0, sql: 'INSERT INTO `boazmohar_experiment`.`session__trial`' } PUT /api/v1/schemas/boazmohar_experiment/data 500 104.906 ms - 3155 ``` If I add a single entry to the `trial` table the form works, also the same inset to the `session` table works from the dj python API.
priority
inserting zero entries to a part table if i have a master part table e g session trial if i try to insert only to the session table with nothing to the trial part table i get the following error bash message generated invalid sql data session session trial error er parse error you have an error in your sql syntax check the manual that corresponds to your mariadb server version for the right syntax to use near at line at query sequence packettoerror users moharb documents repos helium node modules mysql lib protocol sequences sequence js at query errorpacket users moharb documents repos helium node modules mysql lib protocol sequences query js at protocol parsepacket users moharb documents repos helium node modules mysql lib protocol protocol js at parser write users moharb documents repos helium node modules mysql lib protocol parser js at protocol write users moharb documents repos helium node modules mysql lib protocol protocol js at socket users moharb documents repos helium node modules mysql lib connection js at emitone events js at socket emit events js at addchunk stream readable js at readableaddchunk stream readable js at socket readable push stream readable js at tcp onread net js at protocol enqueue users moharb documents repos helium node modules mysql lib protocol protocol js at poolconnection query users moharb documents repos helium node modules mysql lib connection js at users moharb documents repos helium node modules promise mysql lib helper js at promise execute users moharb documents repos helium node modules bluebird js release debuggability js at promise resolvefromexecutor users moharb documents repos helium node modules bluebird js release promise js at new promise users moharb documents repos helium node modules bluebird js release promise js at poolconnection promisecallback users moharb documents repos helium node modules promise mysql lib helper js at poolconnection connection query users moharb documents repos helium node modules promise mysql lib connection js at queryhelper users moharb documents repos helium dist db query helper js at generator next at users moharb documents repos helium dist db query helper js at new promise at awaiter users moharb documents repos helium dist db query helper js at queryhelper execute users moharb documents repos helium dist db query helper js at schemadao users moharb documents repos helium dist routes api schemas schema dao js at generator next code er parse error errno sqlmessage you have an error in your sql syntax check the manual that corresponds to your mariadb server version for the right syntax to use near at line sqlstate index sql insert into boazmohar experiment session trial put api schemas boazmohar experiment data ms if i add a single entry to the trial table the form works also the same inset to the session table works from the dj python api
1
267,148
8,379,544,506
IssuesEvent
2018-10-07 03:27:29
doroxz/test
https://api.github.com/repos/doroxz/test
opened
Add awesome function
Priority/high status/pending type/feature
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. **Describe the solution you'd like** A clear and concise description of what you want to happen. **Additional context** Add any other context or screenshots about the feature request here.
1.0
Add awesome function - **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. **Describe the solution you'd like** A clear and concise description of what you want to happen. **Additional context** Add any other context or screenshots about the feature request here.
priority
add awesome function is your feature request related to a problem please describe a clear and concise description of what the problem is describe the solution you d like a clear and concise description of what you want to happen additional context add any other context or screenshots about the feature request here
1
407,233
11,908,488,975
IssuesEvent
2020-03-31 01:07:10
theintern/intern
https://api.github.com/repos/theintern/intern
opened
Call emit() listeners when events are emitted
bug effort-low priority-high
Currently (Intern 4.8 and Intern 5.0-pre), listener calls are chained in `Executor#emit`. ```ts emit(event, data) { for (const listener of listeners) { notifications = notifications .then(() => Task.resolve(listener(data))) .then(handleErrorEvent) .catch(handleListenerError); } // ... return notifications; } ``` This can lead to handlers receiving events out of order -- a new event may be emitted before the chain for the previous event has finished processing. It's also problematic for situations where timing is important since event handlers may be called at some arbitrary time after an event was actually emitted. Listeners are not inter-dependent; it shouldn't matter (and doesn't currently) in which order they're called, and there's no need to wait on one before calling the other, so they should all be called at the same time, something more like: ```ts const notifications = []; for (const listener of listeners) { notifications.push( Task.resolve(listener(data)) .then(handleErrorEvent) .catch(handleListenerError) ); } // ... return Promise.all(notifications); ```
1.0
Call emit() listeners when events are emitted - Currently (Intern 4.8 and Intern 5.0-pre), listener calls are chained in `Executor#emit`. ```ts emit(event, data) { for (const listener of listeners) { notifications = notifications .then(() => Task.resolve(listener(data))) .then(handleErrorEvent) .catch(handleListenerError); } // ... return notifications; } ``` This can lead to handlers receiving events out of order -- a new event may be emitted before the chain for the previous event has finished processing. It's also problematic for situations where timing is important since event handlers may be called at some arbitrary time after an event was actually emitted. Listeners are not inter-dependent; it shouldn't matter (and doesn't currently) in which order they're called, and there's no need to wait on one before calling the other, so they should all be called at the same time, something more like: ```ts const notifications = []; for (const listener of listeners) { notifications.push( Task.resolve(listener(data)) .then(handleErrorEvent) .catch(handleListenerError) ); } // ... return Promise.all(notifications); ```
priority
call emit listeners when events are emitted currently intern and intern pre listener calls are chained in executor emit ts emit event data for const listener of listeners notifications notifications then task resolve listener data then handleerrorevent catch handlelistenererror return notifications this can lead to handlers receiving events out of order a new event may be emitted before the chain for the previous event has finished processing it s also problematic for situations where timing is important since event handlers may be called at some arbitrary time after an event was actually emitted listeners are not inter dependent it shouldn t matter and doesn t currently in which order they re called and there s no need to wait on one before calling the other so they should all be called at the same time something more like ts const notifications for const listener of listeners notifications push task resolve listener data then handleerrorevent catch handlelistenererror return promise all notifications
1
138,376
5,332,719,679
IssuesEvent
2017-02-15 22:51:28
influxdata/docs.influxdata.com
https://api.github.com/repos/influxdata/docs.influxdata.com
opened
[Enterprise] Clarify influxd-ctl options
Priority: high
https://docs.influxdata.com/enterprise/v1.2/features/cluster-commands/ and https://docs.influxdata.com/enterprise/v1.2/guides/backup-and-restore/ Link backup and restore to the cluster-commands page to show that every option is applicable to the backup/restore influxd-ctl commands.
1.0
[Enterprise] Clarify influxd-ctl options - https://docs.influxdata.com/enterprise/v1.2/features/cluster-commands/ and https://docs.influxdata.com/enterprise/v1.2/guides/backup-and-restore/ Link backup and restore to the cluster-commands page to show that every option is applicable to the backup/restore influxd-ctl commands.
priority
clarify influxd ctl options and link backup and restore to the cluster commands page to show that every option is applicable to the backup restore influxd ctl commands
1
498,023
14,399,022,243
IssuesEvent
2020-12-03 10:21:45
wso2/kubernetes-is
https://api.github.com/repos/wso2/kubernetes-is
closed
[5.11.0] Introduce Kubernetes Resources for Simplified Setup of WSO2 IAM
Priority/High Type/Task
**Description:** Introduce Kubernetes resources for a simplified setup of WSO2 Identity and Access Management version `5.11.0` deployment.
1.0
[5.11.0] Introduce Kubernetes Resources for Simplified Setup of WSO2 IAM - **Description:** Introduce Kubernetes resources for a simplified setup of WSO2 Identity and Access Management version `5.11.0` deployment.
priority
introduce kubernetes resources for simplified setup of iam description introduce kubernetes resources for a simplified setup of identity and access management version deployment
1
742,896
25,876,436,118
IssuesEvent
2022-12-14 08:15:39
minio/minio
https://api.github.com/repos/minio/minio
closed
Rebalancing transfers almost all data to other pool
community fixed priority: high
Site-replication between sites with different capacity. The site with the smallest capacity had one server pool (one node, 4 disks of 1GB each) which reached full capacity. Expanded this site by adding a new server pool. (same capacity in new pool, 4 disks of 1GB each) We tried 'mc admin rebalance' to rebalance data over pools. It didn't balance, it just transferred as much possible data to the new pool. ## Expected Behavior 'mc admin rebalance status' Per-pool usage before rebalancing: - Pool-0 : 0.97% - Pool-1 : 0.13% Per-pool usage after rebalancing: - Pool-0: 0.55% - Pool-1: 0.55% Not exactly 0.55%-0.55%, but at least somewhat close. ## Current Behavior 'mc admin rebalance status' Per-pool usage before rebalancing: - Pool-0 : 0.97% - Pool-1 : 0.13% Per-pool usage after rebalancing: - Pool-0: 0.11% - Pool-1: 0.99% ## Steps to Reproduce (for bugs) 1. Fill one server pool with data 2. Expand by adding a server pool with same capacity 3. mc admin rebalance ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used (`minio --version`): RELEASE.2022-11-17T23-20-09Z (commit-id=a22b4adf4c5cc3e4db13fe92da683ef1ce45cd5a) * Operating System and version (`uname -a`):Linux minio01 5.10.0-15-amd64 SMP Debian 5.10.120-1 (2022-06-09) x86_64 GNU/Linux
1.0
Rebalancing transfers almost all data to other pool - Site-replication between sites with different capacity. The site with the smallest capacity had one server pool (one node, 4 disks of 1GB each) which reached full capacity. Expanded this site by adding a new server pool. (same capacity in new pool, 4 disks of 1GB each) We tried 'mc admin rebalance' to rebalance data over pools. It didn't balance, it just transferred as much possible data to the new pool. ## Expected Behavior 'mc admin rebalance status' Per-pool usage before rebalancing: - Pool-0 : 0.97% - Pool-1 : 0.13% Per-pool usage after rebalancing: - Pool-0: 0.55% - Pool-1: 0.55% Not exactly 0.55%-0.55%, but at least somewhat close. ## Current Behavior 'mc admin rebalance status' Per-pool usage before rebalancing: - Pool-0 : 0.97% - Pool-1 : 0.13% Per-pool usage after rebalancing: - Pool-0: 0.11% - Pool-1: 0.99% ## Steps to Reproduce (for bugs) 1. Fill one server pool with data 2. Expand by adding a server pool with same capacity 3. mc admin rebalance ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used (`minio --version`): RELEASE.2022-11-17T23-20-09Z (commit-id=a22b4adf4c5cc3e4db13fe92da683ef1ce45cd5a) * Operating System and version (`uname -a`):Linux minio01 5.10.0-15-amd64 SMP Debian 5.10.120-1 (2022-06-09) x86_64 GNU/Linux
priority
rebalancing transfers almost all data to other pool site replication between sites with different capacity the site with the smallest capacity had one server pool one node disks of each which reached full capacity expanded this site by adding a new server pool same capacity in new pool disks of each we tried mc admin rebalance to rebalance data over pools it didn t balance it just transferred as much possible data to the new pool expected behavior mc admin rebalance status per pool usage before rebalancing pool pool per pool usage after rebalancing pool pool not exactly but at least somewhat close current behavior mc admin rebalance status per pool usage before rebalancing pool pool per pool usage after rebalancing pool pool steps to reproduce for bugs fill one server pool with data expand by adding a server pool with same capacity mc admin rebalance your environment version used minio version release commit id operating system and version uname a linux smp debian gnu linux
1
713,819
24,540,990,655
IssuesEvent
2022-10-12 03:43:23
ipfs-force-community/venus-cluster
https://api.github.com/repos/ipfs-force-community/venus-cluster
closed
[venus-cluster] 扇区重建开发规划
enhancement priority:high
本 issue 用于规划和跟踪 `扇区重建` 功能的开发 对于以下的检查项,根据必要性,可能进行重组织、继续拆分、转换成独立 issue 等不同处理 - [x] venus-sector-manager - [x] 重建任务 #373 - [x] 数据结构 - [x] 增删改查 - [x] 重建过程相关接口 - [x] venus-worker - [x] 热更新 `sealing_thread` 配置 - [x] 改造 planner 的状态转移,使 `sealing_thread` 具备跳出无任务的空循环的能力 #387 - [x] 提供 `sealing_thread` 在每个循环初始重载配置的能力 #376 - [x] 实现 `rebuild` planner
1.0
[venus-cluster] 扇区重建开发规划 - 本 issue 用于规划和跟踪 `扇区重建` 功能的开发 对于以下的检查项,根据必要性,可能进行重组织、继续拆分、转换成独立 issue 等不同处理 - [x] venus-sector-manager - [x] 重建任务 #373 - [x] 数据结构 - [x] 增删改查 - [x] 重建过程相关接口 - [x] venus-worker - [x] 热更新 `sealing_thread` 配置 - [x] 改造 planner 的状态转移,使 `sealing_thread` 具备跳出无任务的空循环的能力 #387 - [x] 提供 `sealing_thread` 在每个循环初始重载配置的能力 #376 - [x] 实现 `rebuild` planner
priority
扇区重建开发规划 本 issue 用于规划和跟踪 扇区重建 功能的开发 对于以下的检查项,根据必要性,可能进行重组织、继续拆分、转换成独立 issue 等不同处理 venus sector manager 重建任务 数据结构 增删改查 重建过程相关接口 venus worker 热更新 sealing thread 配置 改造 planner 的状态转移,使 sealing thread 具备跳出无任务的空循环的能力 提供 sealing thread 在每个循环初始重载配置的能力 实现 rebuild planner
1
90,092
3,810,570,615
IssuesEvent
2016-03-26 06:17:27
osmosacademy/vr-campaign
https://api.github.com/repos/osmosacademy/vr-campaign
opened
Arrange coral reefs into a pretty scene
help wanted priority:high status:blocked
Once the coral reefs are created: https://github.com/osmosacademy/vr-campaign/issues/11 https://github.com/osmosacademy/vr-campaign/issues/6 Someone has to arrange them into a pretty looking underwater scene in Unity. We may end up just using that scene as our one scene. Or, it will be used to inspire a programmer to randomly generate new levels, based on how you arrange the coral reefs and other things. It will also greatly help us understand and visualize the flow and mechanics of the game. To set up the scene, please learn about nesting game objects in unity to organize your scene. Look it up or ask someone for help. Like these pictures: ![messy1](https://cloud.githubusercontent.com/assets/1461690/14058512/dbe1f376-f2f8-11e5-82ac-1929a58c038a.png) ![messy2](https://cloud.githubusercontent.com/assets/1461690/14058513/dbe639ae-f2f8-11e5-8753-c9c39bb12b20.png) ![messy3](https://cloud.githubusercontent.com/assets/1461690/14058514/dbea9c10-f2f8-11e5-96e6-a21b92ed834f.png)
1.0
Arrange coral reefs into a pretty scene - Once the coral reefs are created: https://github.com/osmosacademy/vr-campaign/issues/11 https://github.com/osmosacademy/vr-campaign/issues/6 Someone has to arrange them into a pretty looking underwater scene in Unity. We may end up just using that scene as our one scene. Or, it will be used to inspire a programmer to randomly generate new levels, based on how you arrange the coral reefs and other things. It will also greatly help us understand and visualize the flow and mechanics of the game. To set up the scene, please learn about nesting game objects in unity to organize your scene. Look it up or ask someone for help. Like these pictures: ![messy1](https://cloud.githubusercontent.com/assets/1461690/14058512/dbe1f376-f2f8-11e5-82ac-1929a58c038a.png) ![messy2](https://cloud.githubusercontent.com/assets/1461690/14058513/dbe639ae-f2f8-11e5-8753-c9c39bb12b20.png) ![messy3](https://cloud.githubusercontent.com/assets/1461690/14058514/dbea9c10-f2f8-11e5-96e6-a21b92ed834f.png)
priority
arrange coral reefs into a pretty scene once the coral reefs are created someone has to arrange them into a pretty looking underwater scene in unity we may end up just using that scene as our one scene or it will be used to inspire a programmer to randomly generate new levels based on how you arrange the coral reefs and other things it will also greatly help us understand and visualize the flow and mechanics of the game to set up the scene please learn about nesting game objects in unity to organize your scene look it up or ask someone for help like these pictures
1
47,856
2,986,639,654
IssuesEvent
2015-07-20 05:56:37
UnifiedViews/Plugins
https://api.github.com/repos/UnifiedViews/Plugins
closed
e-rdfFromSparql has to be reworked
priority: High resolution: fixed severity: enhancement
So far, e-rdfFromSparql was not update to helpers 2.0 because updating the DPU would take more effort than rewriting it. So we should rewrite it using helpers 2.0. TODO: Add reference to the @skodape DPU in mff-uk/DPUs which may be used as a base.
1.0
e-rdfFromSparql has to be reworked - So far, e-rdfFromSparql was not update to helpers 2.0 because updating the DPU would take more effort than rewriting it. So we should rewrite it using helpers 2.0. TODO: Add reference to the @skodape DPU in mff-uk/DPUs which may be used as a base.
priority
e rdffromsparql has to be reworked so far e rdffromsparql was not update to helpers because updating the dpu would take more effort than rewriting it so we should rewrite it using helpers todo add reference to the skodape dpu in mff uk dpus which may be used as a base
1
41,120
2,868,977,171
IssuesEvent
2015-06-05 22:19:49
SCIInstitute/shapeworks
https://api.github.com/repos/SCIInstitute/shapeworks
closed
Allow View2 to open without command line.
enhancement High Priority IBBM
The first thing to happen with view2 is a dialog to pop up asking for the parameter file. If it is a good parameter file, then everything should work as normal. If not, program will keep asking for a good one, or to quit.
1.0
Allow View2 to open without command line. - The first thing to happen with view2 is a dialog to pop up asking for the parameter file. If it is a good parameter file, then everything should work as normal. If not, program will keep asking for a good one, or to quit.
priority
allow to open without command line the first thing to happen with is a dialog to pop up asking for the parameter file if it is a good parameter file then everything should work as normal if not program will keep asking for a good one or to quit
1
822,475
30,874,155,311
IssuesEvent
2023-08-03 13:18:54
unlock-protocol/unlock
https://api.github.com/repos/unlock-protocol/unlock
closed
Can't find a way to disconnect from the Events dapp using Brave
🚨 High Priority
**Describe the bug** Using Brave on laptop, cannot see any way to disconnect from the app. **To Reproduce** Using Brave, go to https://events.unlock-protocol.com For some reason (not sure why) the Events app was showing me as connected. Tried to disconnect and could not find a way to do so. (On the other hand, using the Chrome browser and proper wallet, things work as expected.) **Expected behavior** I would expect to see some way to "disconnect." **Screenshots** https://www.loom.com/share/fd220211250049778398413225fe83af?sid=fef511c6-865b-4a16-802c-a4154d50583f
1.0
Can't find a way to disconnect from the Events dapp using Brave - **Describe the bug** Using Brave on laptop, cannot see any way to disconnect from the app. **To Reproduce** Using Brave, go to https://events.unlock-protocol.com For some reason (not sure why) the Events app was showing me as connected. Tried to disconnect and could not find a way to do so. (On the other hand, using the Chrome browser and proper wallet, things work as expected.) **Expected behavior** I would expect to see some way to "disconnect." **Screenshots** https://www.loom.com/share/fd220211250049778398413225fe83af?sid=fef511c6-865b-4a16-802c-a4154d50583f
priority
can t find a way to disconnect from the events dapp using brave describe the bug using brave on laptop cannot see any way to disconnect from the app to reproduce using brave go to for some reason not sure why the events app was showing me as connected tried to disconnect and could not find a way to do so on the other hand using the chrome browser and proper wallet things work as expected expected behavior i would expect to see some way to disconnect screenshots
1