Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
747,402
26,083,275,573
IssuesEvent
2022-12-25 18:15:53
bounswe/bounswe2022group1
https://api.github.com/repos/bounswe/bounswe2022group1
opened
Adding 'upvote' Feature for a Learning Space
Priority: High Type: Task Status: In Progress Frontend
**Issue Description:** I will add `upvote` feature for a learning space and create a pull request for this new feature if it will be completed. **Tasks to Do:** - [ ] add issue labels - [ ] add reviewer - [ ] add related links - [ ] search for `upvote` feature - [ ] implement `upvote` feature - [ ] create a PR *Task Deadline: 25/12/2022 11:59 pm* *Final Situation:* *Reviewer: @kamilkorkut*
1.0
Adding 'upvote' Feature for a Learning Space - **Issue Description:** I will add `upvote` feature for a learning space and create a pull request for this new feature if it will be completed. **Tasks to Do:** - [ ] add issue labels - [ ] add reviewer - [ ] add related links - [ ] search for `upvote` feature - [ ] implement `upvote` feature - [ ] create a PR *Task Deadline: 25/12/2022 11:59 pm* *Final Situation:* *Reviewer: @kamilkorkut*
priority
adding upvote feature for a learning space issue description i will add upvote feature for a learning space and create a pull request for this new feature if it will be completed tasks to do add issue labels add reviewer add related links search for upvote feature implement upvote feature create a pr task deadline pm final situation reviewer kamilkorkut
1
764,348
26,796,608,963
IssuesEvent
2023-02-01 12:21:37
SuddenDevelopment/StopMotion
https://api.github.com/repos/SuddenDevelopment/StopMotion
opened
Keyframing
enhancement Priority High
Let's have a chat about this. **Some things to think about?** 1. Process of keyframing. 2. Hotkeys. 3. Setting the first key. 4. Storing a default pose for keys.
1.0
Keyframing - Let's have a chat about this. **Some things to think about?** 1. Process of keyframing. 2. Hotkeys. 3. Setting the first key. 4. Storing a default pose for keys.
priority
keyframing let s have a chat about this some things to think about process of keyframing hotkeys setting the first key storing a default pose for keys
1
805,875
29,670,855,630
IssuesEvent
2023-06-11 11:44:43
OJ-lab/judger
https://api.github.com/repos/OJ-lab/judger
closed
Relation fix with judge-test-collection
good first issue high priority
https://github.com/OJ-lab/judger-test-collection/issues/2 After this issue fixed, we should also apply related changes for judger.
1.0
Relation fix with judge-test-collection - https://github.com/OJ-lab/judger-test-collection/issues/2 After this issue fixed, we should also apply related changes for judger.
priority
relation fix with judge test collection after this issue fixed we should also apply related changes for judger
1
112,626
4,535,183,856
IssuesEvent
2016-09-08 16:33:06
openml/openml-r
https://api.github.com/repos/openml/openml-r
closed
Diagnosing data download issues
bug high priority
I encountered some problems while downloading a few datasets, e.g.: ``` Downloading from 'http://www.openml.org/api/v1/data/1223' to '/Users/joa/.openml/cache/datasets/1223/description.xml'. Downloading from 'http://www.openml.org/data/download/249042/letter-challenge-labeled.arff' to '/Users/joa/.openml/cache/datasets/1223/dataset.arff' Error in parseHeader(path) : Invalid column specification line found in ARFF header: <!doctype html> ``` It looks like the server is returning an error/warning instead of the actual dataset. However, since only the first line is shown (I'm using verbosity=2) I cannot read the error message. I also cannot reproduce the error because it works fine when I use the REST API directly. There must be something different with the request made by the R interface. How do I get the full message returned by the server to the R interface? Thanks! Joaquin
1.0
Diagnosing data download issues - I encountered some problems while downloading a few datasets, e.g.: ``` Downloading from 'http://www.openml.org/api/v1/data/1223' to '/Users/joa/.openml/cache/datasets/1223/description.xml'. Downloading from 'http://www.openml.org/data/download/249042/letter-challenge-labeled.arff' to '/Users/joa/.openml/cache/datasets/1223/dataset.arff' Error in parseHeader(path) : Invalid column specification line found in ARFF header: <!doctype html> ``` It looks like the server is returning an error/warning instead of the actual dataset. However, since only the first line is shown (I'm using verbosity=2) I cannot read the error message. I also cannot reproduce the error because it works fine when I use the REST API directly. There must be something different with the request made by the R interface. How do I get the full message returned by the server to the R interface? Thanks! Joaquin
priority
diagnosing data download issues i encountered some problems while downloading a few datasets e g downloading from to users joa openml cache datasets description xml downloading from to users joa openml cache datasets dataset arff error in parseheader path invalid column specification line found in arff header it looks like the server is returning an error warning instead of the actual dataset however since only the first line is shown i m using verbosity i cannot read the error message i also cannot reproduce the error because it works fine when i use the rest api directly there must be something different with the request made by the r interface how do i get the full message returned by the server to the r interface thanks joaquin
1
28,076
2,699,293,984
IssuesEvent
2015-04-03 15:54:52
CenterForOpenScience/osf.io
https://api.github.com/repos/CenterForOpenScience/osf.io
closed
Email links not logged to console when running `invoke server`
5 - Pending Review Bug: Staging Community Priority - High
h/t @cosenal for helping us find this issue on IRC. When `USE_EMAIL` is set to `False` in `website.settings` and a new user is registered, the email that would have been sent to the user is printed to the console as part of the log output. This is important for development, as copying and pasting this link is how a developer can confirm the registration of an account on their local copy. This log no longer prints. Upon investigation, @HarryRybacki and I discovered that while it doesn't print when the server is run through `invoke server`, it does print when it is invoke through `python main.py`. The log lever for the email is `DEBUG`, and it appears that the logger's threshold is being changed somewhere when `inv server` is run, as no logs of that level are being printed. Since our documentation tells devs to run the OSF locally through Invoke, by running `inv server`, this issue is preventing community developers from creating and registering accounts locally.
1.0
Email links not logged to console when running `invoke server` - h/t @cosenal for helping us find this issue on IRC. When `USE_EMAIL` is set to `False` in `website.settings` and a new user is registered, the email that would have been sent to the user is printed to the console as part of the log output. This is important for development, as copying and pasting this link is how a developer can confirm the registration of an account on their local copy. This log no longer prints. Upon investigation, @HarryRybacki and I discovered that while it doesn't print when the server is run through `invoke server`, it does print when it is invoke through `python main.py`. The log lever for the email is `DEBUG`, and it appears that the logger's threshold is being changed somewhere when `inv server` is run, as no logs of that level are being printed. Since our documentation tells devs to run the OSF locally through Invoke, by running `inv server`, this issue is preventing community developers from creating and registering accounts locally.
priority
email links not logged to console when running invoke server h t cosenal for helping us find this issue on irc when use email is set to false in website settings and a new user is registered the email that would have been sent to the user is printed to the console as part of the log output this is important for development as copying and pasting this link is how a developer can confirm the registration of an account on their local copy this log no longer prints upon investigation harryrybacki and i discovered that while it doesn t print when the server is run through invoke server it does print when it is invoke through python main py the log lever for the email is debug and it appears that the logger s threshold is being changed somewhere when inv server is run as no logs of that level are being printed since our documentation tells devs to run the osf locally through invoke by running inv server this issue is preventing community developers from creating and registering accounts locally
1
378,815
11,209,146,276
IssuesEvent
2020-01-06 09:47:57
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Using torch.multiprocessing along with torch.distributed (mpi-backend)
high priority module: distributed module: mpi triage review triaged
## 🐛 Bug Trying to communicate between spawned sub-processes using Pytorch's multiprocessing class within an Openmpi distributed Backend process group fails. ## To Reproduce Steps to reproduce the behavior: 1. Install Openmpi 2. Build / install Pytorch from source 3. Run the following code using "mpirun -np 2 python test_code.py" ``` def run(i, *args): print("I WAS SPAWNED BY:", args[0]) tsr = torch.zeros(1) if args[0] == 0: tsr += 100 dist.send(tsr, dst=1) else: dist.recv(tsr) print ("RECEIVED VALUE =", tsr) if __name__ == '__main__': # Initialize Process Group dist.init_process_group(backend="mpi") mp.set_start_method('spawn') # get current process information world_size = dist.get_world_size() rank = dist.get_rank() # spawn sub-processes mp.spawn(run, args=(rank, world_size,), nprocs=1) ``` Following exception is raised "Default process group is not initialized" ``` Traceback (most recent call last): File "test_code_2.py", line 29, in <module> mp.spawn(run, args=(rank, world_size,), nprocs=1) File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 167, in spawn while not spawn_context.join(): File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 114, in join raise Exception(msg) Exception: -- Process 0 terminated with the following error: Traceback (most recent call last): File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 19, in _wrap fn(i, *args) File "/media/usama/Personal/Study/LAB/experiments/codes/test_code_2.py", line 12, in run dist.send(tsr, dst=1) File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/distributed/distributed_c10d.py", line 660, in send _check_default_pg() File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/distributed/distributed_c10d.py", line 185, in _check_default_pg "Default process group is not initialized" AssertionError: Default process group is not initialized ``` ## Expected behavior Each sub-process should be able to communicate not only with other sub-processes but also with parent processes. ## Environment PyTorch version: 1.1.0a0+8f0603b Is debug build: No CUDA used to build PyTorch: 10.1.105 OS: Ubuntu 18.04.3 LTS GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0 CMake version: version 3.14.3 Python version: 3.7 Is CUDA available: Yes CUDA runtime version: Could not collect GPU models and configuration: GPU 0: GeForce GTX 960M Nvidia driver version: 430.64 cuDNN version: /usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn.so.7 Versions of relevant libraries: [pip] numpy==1.17.4 [pip] numpydoc==0.9.1 [pip] torch==1.1.0a0+c182824 [pip] torchvision==0.2.2 [conda] _tflow_select 2.3.0 mkl [conda] blas 1.0 mkl [conda] mkl 2019.4 243 [conda] mkl-service 2.3.0 py37he904b0f_0 [conda] mkl_fft 1.0.15 py37ha843d7b_0 [conda] mkl_random 1.1.0 py37hd6b4f25_0 [conda] pytorch 1.0.1 cuda100py37he554f03_0 [conda] tensorflow 1.15.0 mkl_py37h28c19af_0 [conda] tensorflow-base 1.15.0 mkl_py37he1670d9_0 [conda] torch 1.1.0a0+c182824 pypi_0 pypi [conda] torchvision 0.2.2 py_3 pytorch ## Additional context Is it even possible to communicate between sub-process to two different parent processes? cc @ezyang @gchanan @zou3519 @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @xush6528
1.0
Using torch.multiprocessing along with torch.distributed (mpi-backend) - ## 🐛 Bug Trying to communicate between spawned sub-processes using Pytorch's multiprocessing class within an Openmpi distributed Backend process group fails. ## To Reproduce Steps to reproduce the behavior: 1. Install Openmpi 2. Build / install Pytorch from source 3. Run the following code using "mpirun -np 2 python test_code.py" ``` def run(i, *args): print("I WAS SPAWNED BY:", args[0]) tsr = torch.zeros(1) if args[0] == 0: tsr += 100 dist.send(tsr, dst=1) else: dist.recv(tsr) print ("RECEIVED VALUE =", tsr) if __name__ == '__main__': # Initialize Process Group dist.init_process_group(backend="mpi") mp.set_start_method('spawn') # get current process information world_size = dist.get_world_size() rank = dist.get_rank() # spawn sub-processes mp.spawn(run, args=(rank, world_size,), nprocs=1) ``` Following exception is raised "Default process group is not initialized" ``` Traceback (most recent call last): File "test_code_2.py", line 29, in <module> mp.spawn(run, args=(rank, world_size,), nprocs=1) File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 167, in spawn while not spawn_context.join(): File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 114, in join raise Exception(msg) Exception: -- Process 0 terminated with the following error: Traceback (most recent call last): File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 19, in _wrap fn(i, *args) File "/media/usama/Personal/Study/LAB/experiments/codes/test_code_2.py", line 12, in run dist.send(tsr, dst=1) File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/distributed/distributed_c10d.py", line 660, in send _check_default_pg() File "/home/usama/anaconda3/lib/python3.7/site-packages/torch/distributed/distributed_c10d.py", line 185, in _check_default_pg "Default process group is not initialized" AssertionError: Default process group is not initialized ``` ## Expected behavior Each sub-process should be able to communicate not only with other sub-processes but also with parent processes. ## Environment PyTorch version: 1.1.0a0+8f0603b Is debug build: No CUDA used to build PyTorch: 10.1.105 OS: Ubuntu 18.04.3 LTS GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0 CMake version: version 3.14.3 Python version: 3.7 Is CUDA available: Yes CUDA runtime version: Could not collect GPU models and configuration: GPU 0: GeForce GTX 960M Nvidia driver version: 430.64 cuDNN version: /usr/local/cuda-10.1/targets/x86_64-linux/lib/libcudnn.so.7 Versions of relevant libraries: [pip] numpy==1.17.4 [pip] numpydoc==0.9.1 [pip] torch==1.1.0a0+c182824 [pip] torchvision==0.2.2 [conda] _tflow_select 2.3.0 mkl [conda] blas 1.0 mkl [conda] mkl 2019.4 243 [conda] mkl-service 2.3.0 py37he904b0f_0 [conda] mkl_fft 1.0.15 py37ha843d7b_0 [conda] mkl_random 1.1.0 py37hd6b4f25_0 [conda] pytorch 1.0.1 cuda100py37he554f03_0 [conda] tensorflow 1.15.0 mkl_py37h28c19af_0 [conda] tensorflow-base 1.15.0 mkl_py37he1670d9_0 [conda] torch 1.1.0a0+c182824 pypi_0 pypi [conda] torchvision 0.2.2 py_3 pytorch ## Additional context Is it even possible to communicate between sub-process to two different parent processes? cc @ezyang @gchanan @zou3519 @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @xush6528
priority
using torch multiprocessing along with torch distributed mpi backend 🐛 bug trying to communicate between spawned sub processes using pytorch s multiprocessing class within an openmpi distributed backend process group fails to reproduce steps to reproduce the behavior install openmpi build install pytorch from source run the following code using mpirun np python test code py def run i args print i was spawned by args tsr torch zeros if args tsr dist send tsr dst else dist recv tsr print received value tsr if name main initialize process group dist init process group backend mpi mp set start method spawn get current process information world size dist get world size rank dist get rank spawn sub processes mp spawn run args rank world size nprocs following exception is raised default process group is not initialized traceback most recent call last file test code py line in mp spawn run args rank world size nprocs file home usama lib site packages torch multiprocessing spawn py line in spawn while not spawn context join file home usama lib site packages torch multiprocessing spawn py line in join raise exception msg exception process terminated with the following error traceback most recent call last file home usama lib site packages torch multiprocessing spawn py line in wrap fn i args file media usama personal study lab experiments codes test code py line in run dist send tsr dst file home usama lib site packages torch distributed distributed py line in send check default pg file home usama lib site packages torch distributed distributed py line in check default pg default process group is not initialized assertionerror default process group is not initialized expected behavior each sub process should be able to communicate not only with other sub processes but also with parent processes environment pytorch version is debug build no cuda used to build pytorch os ubuntu lts gcc version ubuntu cmake version version python version is cuda available yes cuda runtime version could not collect gpu models and configuration gpu geforce gtx nvidia driver version cudnn version usr local cuda targets linux lib libcudnn so versions of relevant libraries numpy numpydoc torch torchvision tflow select mkl blas mkl mkl mkl service mkl fft mkl random pytorch tensorflow mkl tensorflow base mkl torch pypi pypi torchvision py pytorch additional context is it even possible to communicate between sub process to two different parent processes cc ezyang gchanan pietern mrshenli zhaojuanmao satgera rohan varma gqchen aazzolini
1
594,960
18,058,092,471
IssuesEvent
2021-09-20 10:50:25
cusodede/dpl
https://api.github.com/repos/cusodede/dpl
opened
Переезд в гитлаб
priority:high
Планируем на конец этой/начало следующей недели. Гитхаб остаётся в качестве резервной площадки. Доделываем здесь всю текучку, с момента переезда всю новую активность ведём в гитлабе. Что нужно сделать: - [ ] Все PR - влить/закрыть. - [ ] Все тикеты - актуализировать. Выполненное - закрыть, нужное, но не реализованное - перенести в jira, косвенные задачи - по обстоятельствам (лучше всего - выполнить и влить). - [ ] Все ветки - проверить, влитые удалить, не влитые - актуализировать, надо/не надо. Не надо - удалить, надо - оставить тут для доделки.
1.0
Переезд в гитлаб - Планируем на конец этой/начало следующей недели. Гитхаб остаётся в качестве резервной площадки. Доделываем здесь всю текучку, с момента переезда всю новую активность ведём в гитлабе. Что нужно сделать: - [ ] Все PR - влить/закрыть. - [ ] Все тикеты - актуализировать. Выполненное - закрыть, нужное, но не реализованное - перенести в jira, косвенные задачи - по обстоятельствам (лучше всего - выполнить и влить). - [ ] Все ветки - проверить, влитые удалить, не влитые - актуализировать, надо/не надо. Не надо - удалить, надо - оставить тут для доделки.
priority
переезд в гитлаб планируем на конец этой начало следующей недели гитхаб остаётся в качестве резервной площадки доделываем здесь всю текучку с момента переезда всю новую активность ведём в гитлабе что нужно сделать все pr влить закрыть все тикеты актуализировать выполненное закрыть нужное но не реализованное перенести в jira косвенные задачи по обстоятельствам лучше всего выполнить и влить все ветки проверить влитые удалить не влитые актуализировать надо не надо не надо удалить надо оставить тут для доделки
1
271,137
8,476,585,831
IssuesEvent
2018-10-24 22:32:24
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[beta 7.8.0 b7709677] Sorting for match does not work
Fixed High Priority
When you sort for the "match" of an server this will not work, the other sorting features do work.
1.0
[beta 7.8.0 b7709677] Sorting for match does not work - When you sort for the "match" of an server this will not work, the other sorting features do work.
priority
sorting for match does not work when you sort for the match of an server this will not work the other sorting features do work
1
274,868
8,568,910,621
IssuesEvent
2018-11-11 03:33:47
CS2103-AY1819S1-F10-4/main
https://api.github.com/repos/CS2103-AY1819S1-F10-4/main
closed
UI component mentions Person
component.docs priority.high type.bug
**Describe the bug** As illustrated, it still mentions Person. ![image](https://user-images.githubusercontent.com/7473075/48301294-a1bf8000-e526-11e8-8499-bbded75c4dfb.png) **Expected behavior** Should be updated to fit the state of our application.
1.0
UI component mentions Person - **Describe the bug** As illustrated, it still mentions Person. ![image](https://user-images.githubusercontent.com/7473075/48301294-a1bf8000-e526-11e8-8499-bbded75c4dfb.png) **Expected behavior** Should be updated to fit the state of our application.
priority
ui component mentions person describe the bug as illustrated it still mentions person expected behavior should be updated to fit the state of our application
1
745,752
25,999,163,552
IssuesEvent
2022-12-20 14:03:21
nf-core/taxprofiler
https://api.github.com/repos/nf-core/taxprofiler
closed
Revert back to python samplesheet checking
bug high-priority
### Description of the bug We need a release very soon, and unfortunately the current version of `eido` is not producing good enough errors. We should retain the current `eido` implementation in a separte branch however to maybe re-integrate it at a future date. ### Command used and terminal output _No response_ ### Relevant files _No response_ ### System information _No response_
1.0
Revert back to python samplesheet checking - ### Description of the bug We need a release very soon, and unfortunately the current version of `eido` is not producing good enough errors. We should retain the current `eido` implementation in a separte branch however to maybe re-integrate it at a future date. ### Command used and terminal output _No response_ ### Relevant files _No response_ ### System information _No response_
priority
revert back to python samplesheet checking description of the bug we need a release very soon and unfortunately the current version of eido is not producing good enough errors we should retain the current eido implementation in a separte branch however to maybe re integrate it at a future date command used and terminal output no response relevant files no response system information no response
1
368,410
10,878,381,829
IssuesEvent
2019-11-16 17:15:41
bounswe/bounswe2019group10
https://api.github.com/repos/bounswe/bounswe2019group10
closed
Fix evaluation of own writing
Priority: High Relation: Backend Type: Bug
Currently a user can evaluate his/her own quiz. This needs to be fixed.
1.0
Fix evaluation of own writing - Currently a user can evaluate his/her own quiz. This needs to be fixed.
priority
fix evaluation of own writing currently a user can evaluate his her own quiz this needs to be fixed
1
584,563
17,458,354,899
IssuesEvent
2021-08-06 06:47:35
kubesphere/ks-devops
https://api.github.com/repos/kubesphere/ks-devops
closed
Correct repository owner comparison condition in GitHub action configuration
kind/bug priority/high
Recently, we have transferred organization of ks-devops from `kubesphere-sigs` to `kubesphere`. But, [in GitHub action configuration](https://github.com/kubesphere/ks-devops/blob/7714e9c334ad2f2c68bad2260415047eea71b7e2/.github/workflows/build.yaml), `kubesphere-sigs` is still be used to compare with `github.repository_owner`. I'm not sure if we are supposed to label this issue as `good-first-issue`, because this problem will affect docker image build triggered by subsequent commit.
1.0
Correct repository owner comparison condition in GitHub action configuration - Recently, we have transferred organization of ks-devops from `kubesphere-sigs` to `kubesphere`. But, [in GitHub action configuration](https://github.com/kubesphere/ks-devops/blob/7714e9c334ad2f2c68bad2260415047eea71b7e2/.github/workflows/build.yaml), `kubesphere-sigs` is still be used to compare with `github.repository_owner`. I'm not sure if we are supposed to label this issue as `good-first-issue`, because this problem will affect docker image build triggered by subsequent commit.
priority
correct repository owner comparison condition in github action configuration recently we have transferred organization of ks devops from kubesphere sigs to kubesphere but kubesphere sigs is still be used to compare with github repository owner i m not sure if we are supposed to label this issue as good first issue because this problem will affect docker image build triggered by subsequent commit
1
301,342
9,219,243,362
IssuesEvent
2019-03-11 15:02:51
TrinityCore/TrinityCore
https://api.github.com/repos/TrinityCore/TrinityCore
closed
[3.3.5] Core/Spell: Negative auras displayed as buff instead of debuff
Branch-3.3.5a Comp-Core Priority-High Sub-Spells
**Description:** Spell: Leeching Swarm (https://www.wowhead.com/spell=53468) Leeching Swarm is a buff that Deals 2 Nature damage per second **Current behaviour:** It appears like a buff, instead of a debuff **Expected behaviour:** It should be a debuff. **Steps to reproduce the problem:** 1. .go c id 29120 2. Start the encounter and Anub'arak will cast this on the character. **Branch(es):** 3.3.5 **TC rev. hash/commit:** 7b40303a488ef3edcee0930aee5171bb3df97688 **Operating system:** Win 8.1
1.0
[3.3.5] Core/Spell: Negative auras displayed as buff instead of debuff - **Description:** Spell: Leeching Swarm (https://www.wowhead.com/spell=53468) Leeching Swarm is a buff that Deals 2 Nature damage per second **Current behaviour:** It appears like a buff, instead of a debuff **Expected behaviour:** It should be a debuff. **Steps to reproduce the problem:** 1. .go c id 29120 2. Start the encounter and Anub'arak will cast this on the character. **Branch(es):** 3.3.5 **TC rev. hash/commit:** 7b40303a488ef3edcee0930aee5171bb3df97688 **Operating system:** Win 8.1
priority
core spell negative auras displayed as buff instead of debuff description spell leeching swarm leeching swarm is a buff that deals nature damage per second current behaviour it appears like a buff instead of a debuff expected behaviour it should be a debuff steps to reproduce the problem go c id start the encounter and anub arak will cast this on the character branch es tc rev hash commit operating system win
1
832,060
32,071,131,852
IssuesEvent
2023-09-25 08:05:42
EnMAP-Box/enmap-box
https://api.github.com/repos/EnMAP-Box/enmap-box
closed
[Spectral resampling algos] add option for linear interpolation
feature request priority: high
_Requested by Akpona and discussed with @jakimowb and Sebastian van der Linden._ Currently all spectral resampling algos are using SRF for resampling. It is proposed to add an option that uses simple linear interpolation instead.
1.0
[Spectral resampling algos] add option for linear interpolation - _Requested by Akpona and discussed with @jakimowb and Sebastian van der Linden._ Currently all spectral resampling algos are using SRF for resampling. It is proposed to add an option that uses simple linear interpolation instead.
priority
add option for linear interpolation requested by akpona and discussed with jakimowb and sebastian van der linden currently all spectral resampling algos are using srf for resampling it is proposed to add an option that uses simple linear interpolation instead
1
248,353
7,929,578,230
IssuesEvent
2018-07-06 15:31:32
canonical-websites/www.ubuntu.com
https://api.github.com/repos/canonical-websites/www.ubuntu.com
opened
misspelled Support on MOBILE - DEV
Priority: High
<img width="702" alt="screen shot 2018-07-06 at 16 23 17" src="https://user-images.githubusercontent.com/36884067/42387348-f4e3f1c2-8139-11e8-8a08-aab35e41a0e5.png"> --- *Reported from: http://mongoose.staging.ubuntu.com/*
1.0
misspelled Support on MOBILE - DEV - <img width="702" alt="screen shot 2018-07-06 at 16 23 17" src="https://user-images.githubusercontent.com/36884067/42387348-f4e3f1c2-8139-11e8-8a08-aab35e41a0e5.png"> --- *Reported from: http://mongoose.staging.ubuntu.com/*
priority
misspelled support on mobile dev img width alt screen shot at src reported from
1
410,614
11,994,483,788
IssuesEvent
2020-04-08 13:46:59
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
closed
Update All In app links to new help center URL
Bug Needed for V2 launch Priority: High
All Help Cneter links need to be updated to new urls https://docs.google.com/spreadsheets/d/1UHaFfDF9LBNgDcQx4U9ZyQXW8F6OD0MQfnpAvt_O0JI/edit#gid=603506167
1.0
Update All In app links to new help center URL - All Help Cneter links need to be updated to new urls https://docs.google.com/spreadsheets/d/1UHaFfDF9LBNgDcQx4U9ZyQXW8F6OD0MQfnpAvt_O0JI/edit#gid=603506167
priority
update all in app links to new help center url all help cneter links need to be updated to new urls
1
283,800
8,723,523,243
IssuesEvent
2018-12-09 22:33:35
noamross/redoc
https://api.github.com/repos/noamross/redoc
closed
Improve word -> md critic markup lua filter comment handling
high-priority
The lua filter that creates Critic Markup comments from docx is not complete, as it inserts comments at the start location of where they are in the document, but does not properly wrap the selected area with `{==` `==}`.
1.0
Improve word -> md critic markup lua filter comment handling - The lua filter that creates Critic Markup comments from docx is not complete, as it inserts comments at the start location of where they are in the document, but does not properly wrap the selected area with `{==` `==}`.
priority
improve word md critic markup lua filter comment handling the lua filter that creates critic markup comments from docx is not complete as it inserts comments at the start location of where they are in the document but does not properly wrap the selected area with
1
810,931
30,268,609,493
IssuesEvent
2023-07-07 13:45:07
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
opened
Multiple time the same attribute raise an error on the portal
Type: Bug Priority: High
**Describe the bug** When per example you try to hit the portal with this URL: http://10.0.0.1/Cisco::WLC/sidbef01b?&redirect=http://fireoscaptiveportal.com/generate_204&redirect=http://fireoscaptiveportal.com/generate_204 PacketFence trigger an error: httpd.portal-docker-wrapper[289117]: [Fri Jul 07 09:37:10.467422 2023] [perl:error] [pid 16] [client 100.64.0.1:49106] param(): usage error. Invalid syntax at /usr/local/pf/lib/pf/web/externalportal.pm line 208.\n
1.0
Multiple time the same attribute raise an error on the portal - **Describe the bug** When per example you try to hit the portal with this URL: http://10.0.0.1/Cisco::WLC/sidbef01b?&redirect=http://fireoscaptiveportal.com/generate_204&redirect=http://fireoscaptiveportal.com/generate_204 PacketFence trigger an error: httpd.portal-docker-wrapper[289117]: [Fri Jul 07 09:37:10.467422 2023] [perl:error] [pid 16] [client 100.64.0.1:49106] param(): usage error. Invalid syntax at /usr/local/pf/lib/pf/web/externalportal.pm line 208.\n
priority
multiple time the same attribute raise an error on the portal describe the bug when per example you try to hit the portal with this url packetfence trigger an error httpd portal docker wrapper param usage error invalid syntax at usr local pf lib pf web externalportal pm line n
1
67,037
3,265,501,950
IssuesEvent
2015-10-22 16:29:13
CoderDojo/community-platform
https://api.github.com/repos/CoderDojo/community-platform
opened
Capacity missing from Dojo
bug high priority
The capacity is empty for all events on the NSC Mahon Dojo. The Docklands one is working fine so I don't know why this is. This Dojo cannot use the events system because of the error. ![screen shot 2015-10-22 at 17 24 18](https://cloud.githubusercontent.com/assets/1136491/10671303/dae3a67c-78e1-11e5-966d-5a16b91855a9.png)
1.0
Capacity missing from Dojo - The capacity is empty for all events on the NSC Mahon Dojo. The Docklands one is working fine so I don't know why this is. This Dojo cannot use the events system because of the error. ![screen shot 2015-10-22 at 17 24 18](https://cloud.githubusercontent.com/assets/1136491/10671303/dae3a67c-78e1-11e5-966d-5a16b91855a9.png)
priority
capacity missing from dojo the capacity is empty for all events on the nsc mahon dojo the docklands one is working fine so i don t know why this is this dojo cannot use the events system because of the error
1
660,802
22,031,395,007
IssuesEvent
2022-05-28 00:14:46
depub-team/depub.space
https://api.github.com/repos/depub-team/depub.space
reopened
Edit and Delete feature of a single tweet
enhancement high priority
- Add [Edit] (pen icon) below a tweet for the author - Add [Share] (standard iOS share icon) below all tweets - Clicking Edit allows editing the tweet. When edited and reposted, the ISCN will stay same with the version number plus 1. - A [Delete Tweet] button is available in Edit mode. When used, a set of pre-defined words will be used as the new version of the original tweet. Also a parameter will be added to the ISCN will be added, e.g. Deleted: yes (to be discussed). depub.SPACE will not display tweets with such a parameter enabled. - The predefined tweet can be: > This tweet is removed from depub.SPACE frontend by the author. - [ ] Edit a tweet - [ ] Delete a tweet
1.0
Edit and Delete feature of a single tweet - - Add [Edit] (pen icon) below a tweet for the author - Add [Share] (standard iOS share icon) below all tweets - Clicking Edit allows editing the tweet. When edited and reposted, the ISCN will stay same with the version number plus 1. - A [Delete Tweet] button is available in Edit mode. When used, a set of pre-defined words will be used as the new version of the original tweet. Also a parameter will be added to the ISCN will be added, e.g. Deleted: yes (to be discussed). depub.SPACE will not display tweets with such a parameter enabled. - The predefined tweet can be: > This tweet is removed from depub.SPACE frontend by the author. - [ ] Edit a tweet - [ ] Delete a tweet
priority
edit and delete feature of a single tweet add pen icon below a tweet for the author add standard ios share icon below all tweets clicking edit allows editing the tweet when edited and reposted the iscn will stay same with the version number plus a button is available in edit mode when used a set of pre defined words will be used as the new version of the original tweet also a parameter will be added to the iscn will be added e g deleted yes to be discussed depub space will not display tweets with such a parameter enabled the predefined tweet can be this tweet is removed from depub space frontend by the author edit a tweet delete a tweet
1
268,747
8,411,186,580
IssuesEvent
2018-10-12 13:11:03
Kademi/kademi-dev
https://api.github.com/repos/Kademi/kademi-dev
opened
kcom2: possiblity to passthrough without shipping provider
High priority bug
When you do it quickly it will allow you to select shipping provider, but checkout form will not be updated https://drive.google.com/open?id=19vES5rHmKgMZsj0LxGAHXJ7rEHJExJWS http://vladtest54bweb-vladtest54b.kademi-ci.co/vladtest54bweb-ecomstore/cart
1.0
kcom2: possiblity to passthrough without shipping provider - When you do it quickly it will allow you to select shipping provider, but checkout form will not be updated https://drive.google.com/open?id=19vES5rHmKgMZsj0LxGAHXJ7rEHJExJWS http://vladtest54bweb-vladtest54b.kademi-ci.co/vladtest54bweb-ecomstore/cart
priority
possiblity to passthrough without shipping provider when you do it quickly it will allow you to select shipping provider but checkout form will not be updated
1
474,044
13,651,368,843
IssuesEvent
2020-09-27 00:43:23
labsquare/cutevariant
https://api.github.com/repos/labsquare/cutevariant
closed
Cutevariant not working when installed via pip
bug devel high-priority master
I installed cutevariant via pip (without any error) but when I tried to launch cutevariant I got the following error : ![Screenshot_20200918_122704](https://user-images.githubusercontent.com/24796531/93602470-f190ed00-f9c2-11ea-808b-6133e4d0f0ad.png) Requirements allow PySide2>=5.11.2 but cutevariant is not working with the last version of PySide2 (5.15.1). The error disappears if you force the installation of PySide2 == 5.11.2.
1.0
Cutevariant not working when installed via pip - I installed cutevariant via pip (without any error) but when I tried to launch cutevariant I got the following error : ![Screenshot_20200918_122704](https://user-images.githubusercontent.com/24796531/93602470-f190ed00-f9c2-11ea-808b-6133e4d0f0ad.png) Requirements allow PySide2>=5.11.2 but cutevariant is not working with the last version of PySide2 (5.15.1). The error disappears if you force the installation of PySide2 == 5.11.2.
priority
cutevariant not working when installed via pip i installed cutevariant via pip without any error but when i tried to launch cutevariant i got the following error requirements allow but cutevariant is not working with the last version of the error disappears if you force the installation of
1
613,260
19,085,107,339
IssuesEvent
2021-11-29 04:15:22
pycaret/pycaret
https://api.github.com/repos/pycaret/pycaret
closed
Add Prediction Interval coverage of the test set
enhancement time_series metrics priority_high
**Is your feature request related to a problem? Please describe.** Per this [comment on LinkedIn](https://www.linkedin.com/feed/update/urn:li:ugcPost:6814933350713741312?commentUrn=urn%3Ali%3Acomment%3A%28ugcPost%3A6814933350713741312%2C6814934502796779520%29&replyUrn=urn%3Ali%3Acomment%3A%28ugcPost%3A6814933350713741312%2C6815038277939273728%29) from @RamiKrispin **Describe the solution you'd like** If you are running a horse race between multiple models with backtesting, it is nice to add the PI coverage rate on the testing partitions.
1.0
Add Prediction Interval coverage of the test set - **Is your feature request related to a problem? Please describe.** Per this [comment on LinkedIn](https://www.linkedin.com/feed/update/urn:li:ugcPost:6814933350713741312?commentUrn=urn%3Ali%3Acomment%3A%28ugcPost%3A6814933350713741312%2C6814934502796779520%29&replyUrn=urn%3Ali%3Acomment%3A%28ugcPost%3A6814933350713741312%2C6815038277939273728%29) from @RamiKrispin **Describe the solution you'd like** If you are running a horse race between multiple models with backtesting, it is nice to add the PI coverage rate on the testing partitions.
priority
add prediction interval coverage of the test set is your feature request related to a problem please describe per this from ramikrispin describe the solution you d like if you are running a horse race between multiple models with backtesting it is nice to add the pi coverage rate on the testing partitions
1
765,007
26,828,365,520
IssuesEvent
2023-02-02 14:25:52
NIAEFEUP/tts-revamp-fe
https://api.github.com/repos/NIAEFEUP/tts-revamp-fe
closed
Hotfix: schedule editing after importing not applying
bug high priority medium effort
After importing a schedule, if the schedule is then edited, the changes are permanent: after switching to another schedule and then back to the imported schedule it's possible to notice that the changes were not applied. Note: this only happens once, after switching back and forth once the schedule can be edited without any problems.
1.0
Hotfix: schedule editing after importing not applying - After importing a schedule, if the schedule is then edited, the changes are permanent: after switching to another schedule and then back to the imported schedule it's possible to notice that the changes were not applied. Note: this only happens once, after switching back and forth once the schedule can be edited without any problems.
priority
hotfix schedule editing after importing not applying after importing a schedule if the schedule is then edited the changes are permanent after switching to another schedule and then back to the imported schedule it s possible to notice that the changes were not applied note this only happens once after switching back and forth once the schedule can be edited without any problems
1
664,752
22,287,075,428
IssuesEvent
2022-06-11 20:06:26
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
closed
Fail to pull FIO image, due to pull rate limit
bug High Priority lifecycle/stale
tests with io on background may fail if pull rate limit reached from docker.io test report [here](http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j002aud1c33-ua/j002aud1c33-ua_20210720T192204/logs/test_report_1626808655.html), see `test_add_capacity`
1.0
Fail to pull FIO image, due to pull rate limit - tests with io on background may fail if pull rate limit reached from docker.io test report [here](http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j002aud1c33-ua/j002aud1c33-ua_20210720T192204/logs/test_report_1626808655.html), see `test_add_capacity`
priority
fail to pull fio image due to pull rate limit tests with io on background may fail if pull rate limit reached from docker io test report see test add capacity
1
584,293
17,411,024,143
IssuesEvent
2021-08-03 12:21:37
OceanDataTools/openrvdas
https://api.github.com/repos/OceanDataTools/openrvdas
opened
Fix logger_manager stderr listening
bug high priority
**Describe the bug** Right now, the logger_manager sets up threads to listen to each logger's stderr file to relay those lines to the console and cached data server. When one or more loggers is generating a lot of stderr output (say, because of a bad parse format), the logger manager gets overloaded and starts timing out on other tasks. This causes cascading failures. The problem seems to be exacerbated when the logger_manager starts up if logger stderr files are already large.
1.0
Fix logger_manager stderr listening - **Describe the bug** Right now, the logger_manager sets up threads to listen to each logger's stderr file to relay those lines to the console and cached data server. When one or more loggers is generating a lot of stderr output (say, because of a bad parse format), the logger manager gets overloaded and starts timing out on other tasks. This causes cascading failures. The problem seems to be exacerbated when the logger_manager starts up if logger stderr files are already large.
priority
fix logger manager stderr listening describe the bug right now the logger manager sets up threads to listen to each logger s stderr file to relay those lines to the console and cached data server when one or more loggers is generating a lot of stderr output say because of a bad parse format the logger manager gets overloaded and starts timing out on other tasks this causes cascading failures the problem seems to be exacerbated when the logger manager starts up if logger stderr files are already large
1
743,231
25,892,074,481
IssuesEvent
2022-12-14 18:49:53
dmwm/CRABServer
https://api.github.com/repos/dmwm/CRABServer
opened
add support for user containers
Type: Enhancement Type: Question Priority: High Status: Available
all we know about this so far: Katy had a meeting with convenors of new Common Analysis Tools (CAT) group. They already have questions about CRAB, one is: Running in an ad-hoc container instead of CMSSW standard singularity image. This may be possible already, Dario has info which he will share although the container needs to be on CVMFS (quite a limitation in a way). Dario passed this config https://gitlab.cern.ch/-/snippets/2139 to Shahzad some time ago We need to understand the use case, and decide if to support. We need to verify and understand details and decide how to document/offer this to users. Since request came from CAT convenors, put prio as high, at least meaning that we need to undrestand more and clarify before we define exactly what to do
1.0
add support for user containers - all we know about this so far: Katy had a meeting with convenors of new Common Analysis Tools (CAT) group. They already have questions about CRAB, one is: Running in an ad-hoc container instead of CMSSW standard singularity image. This may be possible already, Dario has info which he will share although the container needs to be on CVMFS (quite a limitation in a way). Dario passed this config https://gitlab.cern.ch/-/snippets/2139 to Shahzad some time ago We need to understand the use case, and decide if to support. We need to verify and understand details and decide how to document/offer this to users. Since request came from CAT convenors, put prio as high, at least meaning that we need to undrestand more and clarify before we define exactly what to do
priority
add support for user containers all we know about this so far katy had a meeting with convenors of new common analysis tools cat group they already have questions about crab one is running in an ad hoc container instead of cmssw standard singularity image this may be possible already dario has info which he will share although the container needs to be on cvmfs quite a limitation in a way dario passed this config to shahzad some time ago we need to understand the use case and decide if to support we need to verify and understand details and decide how to document offer this to users since request came from cat convenors put prio as high at least meaning that we need to undrestand more and clarify before we define exactly what to do
1
64,393
3,211,199,108
IssuesEvent
2015-10-06 09:23:16
music-encoding/music-encoding
https://api.github.com/repos/music-encoding/music-encoding
closed
Permanent links to the schema
Component: Core Schema Priority: High Status: Needs Discussion
Since the Google Code repository will be dropped, the links to the schema previously hosted on Google Code will need to be changed to the new url. This is actually a pretty bad practice and I think we should avoid it in the future. So we should have a permanent url for the schema that does not rely on a repository system. We can use music-encoding.org and either copy each release there or redirect to GitHub. It would also be good to have the schema without version number referring to the latest version but also each version numbered explicitly for users who want to refer to a specific release. That is, something like: http://www.music-encoding.org/schema/mei-all.rng http://www.music-encoding.org/schema/mei-all-3.0.0.rng Any thoughts?
1.0
Permanent links to the schema - Since the Google Code repository will be dropped, the links to the schema previously hosted on Google Code will need to be changed to the new url. This is actually a pretty bad practice and I think we should avoid it in the future. So we should have a permanent url for the schema that does not rely on a repository system. We can use music-encoding.org and either copy each release there or redirect to GitHub. It would also be good to have the schema without version number referring to the latest version but also each version numbered explicitly for users who want to refer to a specific release. That is, something like: http://www.music-encoding.org/schema/mei-all.rng http://www.music-encoding.org/schema/mei-all-3.0.0.rng Any thoughts?
priority
permanent links to the schema since the google code repository will be dropped the links to the schema previously hosted on google code will need to be changed to the new url this is actually a pretty bad practice and i think we should avoid it in the future so we should have a permanent url for the schema that does not rely on a repository system we can use music encoding org and either copy each release there or redirect to github it would also be good to have the schema without version number referring to the latest version but also each version numbered explicitly for users who want to refer to a specific release that is something like any thoughts
1
638,308
20,721,097,226
IssuesEvent
2022-03-13 12:03:59
AY2122S2-CS2103T-W15-2/tp
https://api.github.com/repos/AY2122S2-CS2103T-W15-2/tp
opened
Update GUI Icon and Name
priority.High type.Task
Update Icon and name displayed in the GUI to match with the project name and icon.
1.0
Update GUI Icon and Name - Update Icon and name displayed in the GUI to match with the project name and icon.
priority
update gui icon and name update icon and name displayed in the gui to match with the project name and icon
1
504,320
14,616,756,193
IssuesEvent
2020-12-22 13:45:54
SAP/ownid-webapp
https://api.github.com/repos/SAP/ownid-webapp
closed
Reset password notify that phone already associated with an account
Priority: High Type: Bug
Reset password should overwrite the existing data. ![image.png](https://zenhub.mo.sap.corp/images/5ebd66da010eda23834dcc74/3e1b95fd-af69-420d-8575-9c1a0411de24)
1.0
Reset password notify that phone already associated with an account - Reset password should overwrite the existing data. ![image.png](https://zenhub.mo.sap.corp/images/5ebd66da010eda23834dcc74/3e1b95fd-af69-420d-8575-9c1a0411de24)
priority
reset password notify that phone already associated with an account reset password should overwrite the existing data
1
4,861
2,564,696,700
IssuesEvent
2015-02-06 21:42:29
prey/prey-android-client
https://api.github.com/repos/prey/prey-android-client
closed
Passwords stored in plaintext in com.prey_references.xml
bug priority: high
The preferences `PREFS_ADMIN_DEVICE_REVOKED_PASSWORD`, `PASSWORD`, `UNLOCK_PASS` store passwords in plaintext. Please encrypt these, using one-way hashes like sha1 if possible. Particularly, the file seems to contain enough information to log in to my prey account. And if someone uses the same password for prey.com as the email they use to log in, that email address is right there too. This file can probably be opened by root only, but root access on an android device is usually a matter of selecting "allow" - or not doing anything at all if the app was used before. So basically, anyone smart enough who steals my phone could avoid changing the SIM card (so prey doesn't activate itself), read that file, get the password, and detach prey / uninstall it / take over my prey account. Fun. (Of course this previous argument isn't very valid since root access with no auth also lets you remove prey with a `rm` and maybe a `kill`. Or factory reset right after stealing it, no root or passwords required. The fact that it's trivial to identify that a phone runs prey annoys me too. But I digress, this ticket isn't meant to cover all of my paranoia sources.)
1.0
Passwords stored in plaintext in com.prey_references.xml - The preferences `PREFS_ADMIN_DEVICE_REVOKED_PASSWORD`, `PASSWORD`, `UNLOCK_PASS` store passwords in plaintext. Please encrypt these, using one-way hashes like sha1 if possible. Particularly, the file seems to contain enough information to log in to my prey account. And if someone uses the same password for prey.com as the email they use to log in, that email address is right there too. This file can probably be opened by root only, but root access on an android device is usually a matter of selecting "allow" - or not doing anything at all if the app was used before. So basically, anyone smart enough who steals my phone could avoid changing the SIM card (so prey doesn't activate itself), read that file, get the password, and detach prey / uninstall it / take over my prey account. Fun. (Of course this previous argument isn't very valid since root access with no auth also lets you remove prey with a `rm` and maybe a `kill`. Or factory reset right after stealing it, no root or passwords required. The fact that it's trivial to identify that a phone runs prey annoys me too. But I digress, this ticket isn't meant to cover all of my paranoia sources.)
priority
passwords stored in plaintext in com prey references xml the preferences prefs admin device revoked password password unlock pass store passwords in plaintext please encrypt these using one way hashes like if possible particularly the file seems to contain enough information to log in to my prey account and if someone uses the same password for prey com as the email they use to log in that email address is right there too this file can probably be opened by root only but root access on an android device is usually a matter of selecting allow or not doing anything at all if the app was used before so basically anyone smart enough who steals my phone could avoid changing the sim card so prey doesn t activate itself read that file get the password and detach prey uninstall it take over my prey account fun of course this previous argument isn t very valid since root access with no auth also lets you remove prey with a rm and maybe a kill or factory reset right after stealing it no root or passwords required the fact that it s trivial to identify that a phone runs prey annoys me too but i digress this ticket isn t meant to cover all of my paranoia sources
1
517,869
15,020,503,843
IssuesEvent
2021-02-01 14:47:20
ansible/galaxy_ng
https://api.github.com/repos/ansible/galaxy_ng
closed
Remove legacy v3/_ui/ endpoint
area/api priority/high status/new type/bug
Remove old `v3/_ui/` endpoint once the UI on cloud.redhat.com has been updated to not require it anymore. This should happen after we push our current api to cloud.redhat.com
1.0
Remove legacy v3/_ui/ endpoint - Remove old `v3/_ui/` endpoint once the UI on cloud.redhat.com has been updated to not require it anymore. This should happen after we push our current api to cloud.redhat.com
priority
remove legacy ui endpoint remove old ui endpoint once the ui on cloud redhat com has been updated to not require it anymore this should happen after we push our current api to cloud redhat com
1
586,543
17,580,421,179
IssuesEvent
2021-08-16 06:31:13
ivpn/android-app
https://api.github.com/repos/ivpn/android-app
closed
Default trust status is always reverted to None
type: bug Network Protection priority: high
### Description: On version 2.4.1, when setting the default trust status to Untrusted/Trusted, the trust status is always reverted back to None. The issue does not happen the first time the trust status is changed, but when logging out and logging back in, from this moment on, it is not possible to change the default to Untrusted or Trusted. **Note:** The issue happens in all devices, Android OS. This issue might be related to the changes implemented on https://github.com/ivpn/android-app/issues/48 ### Actual result: Default trust status is always reverted to None. ### Expected result: User should always be able to change the trust status of default, mobile and current network. ### Steps to reproduce: 1. Install 2.4.1. 2. Login. 3. Enable Network Protection. 4. Set default to untrusted - app connects automatically. 5. Disconnect from VPN. 6. Enable KillSwitch. 7. Logout. 8. Log back in. 9. Set default to Untrusted - nothing happens. 10. Access network protection again, default is reverted to none. ### Environment: IVPN: 2.4.1 Devices: Samsung Galaxy S10/Android 11, Xiaomi Redmi 5/Android 8.1
1.0
Default trust status is always reverted to None - ### Description: On version 2.4.1, when setting the default trust status to Untrusted/Trusted, the trust status is always reverted back to None. The issue does not happen the first time the trust status is changed, but when logging out and logging back in, from this moment on, it is not possible to change the default to Untrusted or Trusted. **Note:** The issue happens in all devices, Android OS. This issue might be related to the changes implemented on https://github.com/ivpn/android-app/issues/48 ### Actual result: Default trust status is always reverted to None. ### Expected result: User should always be able to change the trust status of default, mobile and current network. ### Steps to reproduce: 1. Install 2.4.1. 2. Login. 3. Enable Network Protection. 4. Set default to untrusted - app connects automatically. 5. Disconnect from VPN. 6. Enable KillSwitch. 7. Logout. 8. Log back in. 9. Set default to Untrusted - nothing happens. 10. Access network protection again, default is reverted to none. ### Environment: IVPN: 2.4.1 Devices: Samsung Galaxy S10/Android 11, Xiaomi Redmi 5/Android 8.1
priority
default trust status is always reverted to none description on version when setting the default trust status to untrusted trusted the trust status is always reverted back to none the issue does not happen the first time the trust status is changed but when logging out and logging back in from this moment on it is not possible to change the default to untrusted or trusted note the issue happens in all devices android os this issue might be related to the changes implemented on actual result default trust status is always reverted to none expected result user should always be able to change the trust status of default mobile and current network steps to reproduce install login enable network protection set default to untrusted app connects automatically disconnect from vpn enable killswitch logout log back in set default to untrusted nothing happens access network protection again default is reverted to none environment ivpn devices samsung galaxy android xiaomi redmi android
1
506,045
14,657,130,731
IssuesEvent
2020-12-28 14:56:41
dmwm/CRABServer
https://api.github.com/repos/dmwm/CRABServer
closed
PublisherMaster print final message too soon
Area: StandalonePublish/ASOless Priority: High Type: Bug
e.g. here PublihserMaster print "iteration completed" before Process 58 stopped. I was looking at things and noticed that after log had this content still `ps fux` was showing one slave running ``` 2020-12-26 18:25:19,570:INFO:master:Starting process <Process(Process-56, started)> pid=1071 2020-12-26 18:25:29,581:INFO:master:Terminated: <Process(Process-55, stopped)> pid=1047 2020-12-26 18:25:29,585:INFO:master:Starting process <Process(Process-57, started)> pid=1078 2020-12-26 18:25:39,597:INFO:master:Terminated: <Process(Process-56, stopped)> pid=1071 2020-12-26 18:25:39,601:INFO:master:Starting process <Process(Process-58, started)> pid=1090 2020-12-26 18:25:49,613:INFO:master:Terminated: <Process(Process-58, stopped)> pid=1090 2020-12-26 18:25:49,617:INFO:master:Starting process <Process(Process-59, started)> pid=1204 2020-12-26 18:25:59,629:INFO:master:Terminated: <Process(Process-57, stopped)> pid=1078 2020-12-26 18:25:59,629:INFO:master:Algorithm iteration completed 2020-12-26 18:25:59,630:INFO:master:Wait 1498 sec for next cycle 2020-12-26 18:25:59,630:INFO:master:Next cycle will start at 18:50:57 [crab3@crab-prod-tw02 processes]$ ```
1.0
PublisherMaster print final message too soon - e.g. here PublihserMaster print "iteration completed" before Process 58 stopped. I was looking at things and noticed that after log had this content still `ps fux` was showing one slave running ``` 2020-12-26 18:25:19,570:INFO:master:Starting process <Process(Process-56, started)> pid=1071 2020-12-26 18:25:29,581:INFO:master:Terminated: <Process(Process-55, stopped)> pid=1047 2020-12-26 18:25:29,585:INFO:master:Starting process <Process(Process-57, started)> pid=1078 2020-12-26 18:25:39,597:INFO:master:Terminated: <Process(Process-56, stopped)> pid=1071 2020-12-26 18:25:39,601:INFO:master:Starting process <Process(Process-58, started)> pid=1090 2020-12-26 18:25:49,613:INFO:master:Terminated: <Process(Process-58, stopped)> pid=1090 2020-12-26 18:25:49,617:INFO:master:Starting process <Process(Process-59, started)> pid=1204 2020-12-26 18:25:59,629:INFO:master:Terminated: <Process(Process-57, stopped)> pid=1078 2020-12-26 18:25:59,629:INFO:master:Algorithm iteration completed 2020-12-26 18:25:59,630:INFO:master:Wait 1498 sec for next cycle 2020-12-26 18:25:59,630:INFO:master:Next cycle will start at 18:50:57 [crab3@crab-prod-tw02 processes]$ ```
priority
publishermaster print final message too soon e g here publihsermaster print iteration completed before process stopped i was looking at things and noticed that after log had this content still ps fux was showing one slave running info master starting process pid info master terminated pid info master starting process pid info master terminated pid info master starting process pid info master terminated pid info master starting process pid info master terminated pid info master algorithm iteration completed info master wait sec for next cycle info master next cycle will start at
1
788,049
27,741,736,054
IssuesEvent
2023-03-15 14:41:36
nlbdev/nordic-epub3-dtbook-migrator
https://api.github.com/repos/nlbdev/nordic-epub3-dtbook-migrator
closed
Add check for backlink presence in footnotes/endnotes
High priority validator-revision
Currently, we don't have a rule that checks backlink presence for footnotes/endnotes, so this should be added. The [current text of the Nordic Guidelines](https://format.mtm.se/nordic_epub/2020-1#notes-and-note-references) is not exact on which of the two methods for backlinks from the DAISY KB should be preferred in 2020-1, so to begin with, we should check both or any method (always including `@role="doc-backlink"`). Is link integrity check enough, or should we implement a cross-reference check that ensures that the backlink target is equal to the referencing noteref of the present footnote?
1.0
Add check for backlink presence in footnotes/endnotes - Currently, we don't have a rule that checks backlink presence for footnotes/endnotes, so this should be added. The [current text of the Nordic Guidelines](https://format.mtm.se/nordic_epub/2020-1#notes-and-note-references) is not exact on which of the two methods for backlinks from the DAISY KB should be preferred in 2020-1, so to begin with, we should check both or any method (always including `@role="doc-backlink"`). Is link integrity check enough, or should we implement a cross-reference check that ensures that the backlink target is equal to the referencing noteref of the present footnote?
priority
add check for backlink presence in footnotes endnotes currently we don t have a rule that checks backlink presence for footnotes endnotes so this should be added the is not exact on which of the two methods for backlinks from the daisy kb should be preferred in so to begin with we should check both or any method always including role doc backlink is link integrity check enough or should we implement a cross reference check that ensures that the backlink target is equal to the referencing noteref of the present footnote
1
739,739
25,715,736,790
IssuesEvent
2022-12-07 10:11:54
DataDog/guarddog
https://api.github.com/repos/DataDog/guarddog
closed
GuardDog may fail on `verify` if trying to unzip a file
bug high-priority
Can be reproduced with ``` guarddog verify requirements.txt ``` With the requirements file from the project (see [PR](https://github.com/DataDog/guarddog/pull/103))
1.0
GuardDog may fail on `verify` if trying to unzip a file - Can be reproduced with ``` guarddog verify requirements.txt ``` With the requirements file from the project (see [PR](https://github.com/DataDog/guarddog/pull/103))
priority
guarddog may fail on verify if trying to unzip a file can be reproduced with guarddog verify requirements txt with the requirements file from the project see
1
117,774
4,727,637,441
IssuesEvent
2016-10-18 14:02:02
zoonproject/zoon
https://api.github.com/repos/zoonproject/zoon
closed
User defined attributes in df not propagated
bug Priority - high
User defined attributes (as opposed to hard-coded attributes like `covCols`) don't propagate through the workflow. In line 387 (zoonHelpers.R), the `cbind` strips the attributes, and user defined attributes in occurrence of process modules never get recovered (only the `covCols` attribute get recovered). I'll write up a pull request for this today or tomorrow, as fixing this is necessary for allowing support of detection models.
1.0
User defined attributes in df not propagated - User defined attributes (as opposed to hard-coded attributes like `covCols`) don't propagate through the workflow. In line 387 (zoonHelpers.R), the `cbind` strips the attributes, and user defined attributes in occurrence of process modules never get recovered (only the `covCols` attribute get recovered). I'll write up a pull request for this today or tomorrow, as fixing this is necessary for allowing support of detection models.
priority
user defined attributes in df not propagated user defined attributes as opposed to hard coded attributes like covcols don t propagate through the workflow in line zoonhelpers r the cbind strips the attributes and user defined attributes in occurrence of process modules never get recovered only the covcols attribute get recovered i ll write up a pull request for this today or tomorrow as fixing this is necessary for allowing support of detection models
1
559,060
16,549,103,221
IssuesEvent
2021-05-28 06:11:33
bryntum/support
https://api.github.com/repos/bryntum/support
closed
Cannot assign multiple resources when creating new event
bug high-priority resolved
https://bryntum.com/examples/calendar/multiassign/ double click to create a new event notice the resource picker is single select (should be multi)
1.0
Cannot assign multiple resources when creating new event - https://bryntum.com/examples/calendar/multiassign/ double click to create a new event notice the resource picker is single select (should be multi)
priority
cannot assign multiple resources when creating new event double click to create a new event notice the resource picker is single select should be multi
1
311,647
9,537,031,686
IssuesEvent
2019-04-30 11:23:04
aartiukh/sph
https://api.github.com/repos/aartiukh/sph
opened
Setup coverage measure using Traciv CI and codecov
area: cmake priority: high type: enhancement
**Is your feature request related to a problem? Please describe.** Add coverage measure and badge to README
1.0
Setup coverage measure using Traciv CI and codecov - **Is your feature request related to a problem? Please describe.** Add coverage measure and badge to README
priority
setup coverage measure using traciv ci and codecov is your feature request related to a problem please describe add coverage measure and badge to readme
1
533,729
15,597,683,962
IssuesEvent
2021-03-18 17:11:50
AY2021S2-CS2113-F10-3/tp
https://api.github.com/repos/AY2021S2-CS2113-F10-3/tp
closed
As a user, I can store all my data locally
priority.High type.Epic
so that my saved schedules can be loaded whenever I load the application
1.0
As a user, I can store all my data locally - so that my saved schedules can be loaded whenever I load the application
priority
as a user i can store all my data locally so that my saved schedules can be loaded whenever i load the application
1
253,203
8,052,773,985
IssuesEvent
2018-08-01 20:25:57
children-of-gazimba/companion-qt-tool
https://api.github.com/repos/children-of-gazimba/companion-qt-tool
closed
visualize volume change when hover scrolling sound tiles
enhancement high priority on it!
The hover scrolling volume change has seen some good use, yet it is still difficult to make precise changes. We need visual feedback to aid this feature. It might make most sense to have a slider pop up and visualize the change whenever a tile is hover scrolled.
1.0
visualize volume change when hover scrolling sound tiles - The hover scrolling volume change has seen some good use, yet it is still difficult to make precise changes. We need visual feedback to aid this feature. It might make most sense to have a slider pop up and visualize the change whenever a tile is hover scrolled.
priority
visualize volume change when hover scrolling sound tiles the hover scrolling volume change has seen some good use yet it is still difficult to make precise changes we need visual feedback to aid this feature it might make most sense to have a slider pop up and visualize the change whenever a tile is hover scrolled
1
689,684
23,630,463,839
IssuesEvent
2022-08-25 08:53:23
apluslms/a-plus
https://api.github.com/repos/apluslms/a-plus
closed
Assignments using custom JavaScript stopped working in v1.16
type: bug priority: high effort: days experience: moderate requester: CS area: javascript
A+ v1.16 has a bug that breaks assignments that use custom JavaScript code in the assignment description/instructions. Clicking the submit button does not seem to do anything: it does not upload the submission, no data is saved and nothing is graded since there is no submission. > The "submit" button does not submit any data anymore, instead it just makes the button disabled. Private ticket: https://rt.cs.aalto.fi/Ticket/Display.html?id=21819 One affected course and assignment are listed in the RT ticket.
1.0
Assignments using custom JavaScript stopped working in v1.16 - A+ v1.16 has a bug that breaks assignments that use custom JavaScript code in the assignment description/instructions. Clicking the submit button does not seem to do anything: it does not upload the submission, no data is saved and nothing is graded since there is no submission. > The "submit" button does not submit any data anymore, instead it just makes the button disabled. Private ticket: https://rt.cs.aalto.fi/Ticket/Display.html?id=21819 One affected course and assignment are listed in the RT ticket.
priority
assignments using custom javascript stopped working in a has a bug that breaks assignments that use custom javascript code in the assignment description instructions clicking the submit button does not seem to do anything it does not upload the submission no data is saved and nothing is graded since there is no submission the submit button does not submit any data anymore instead it just makes the button disabled private ticket one affected course and assignment are listed in the rt ticket
1
594,710
18,051,826,273
IssuesEvent
2021-09-19 21:48:39
EdwinParra35/MinTIC_g02_TicDigitalG8
https://api.github.com/repos/EdwinParra35/MinTIC_g02_TicDigitalG8
closed
Genera un formulario que permita la administración de películas
High priority
Genera un formulario que permita la administración de películas que hay disponibles en cartelera para ver en cine. ¿Qué información es relevante? Antes de ponerlo en consideración con el grupo de trabajo, genera tu propia propuesta de interfaz. Luego, puedes generar una nueva versión del diseño.
1.0
Genera un formulario que permita la administración de películas - Genera un formulario que permita la administración de películas que hay disponibles en cartelera para ver en cine. ¿Qué información es relevante? Antes de ponerlo en consideración con el grupo de trabajo, genera tu propia propuesta de interfaz. Luego, puedes generar una nueva versión del diseño.
priority
genera un formulario que permita la administración de películas genera un formulario que permita la administración de películas que hay disponibles en cartelera para ver en cine ¿qué información es relevante antes de ponerlo en consideración con el grupo de trabajo genera tu propia propuesta de interfaz luego puedes generar una nueva versión del diseño
1
596,026
18,094,793,600
IssuesEvent
2021-09-22 07:51:13
turbot/steampipe-plugin-aws
https://api.github.com/repos/turbot/steampipe-plugin-aws
closed
Handling exceptions aws_macie2_classification_job table
bug priority:high
**Describe the bug** `aws_macie2_classification_job` table returns AccessDeniedException in case any one of the regions is configured in `aws.spc `file is not enabled with Macie. Also, provide the Macie service enabled/disabled status part of this table. **Steampipe version (`steampipe -v`)** Example: v0.3.0 **Plugin version (`steampipe plugin list`)** Example: v0.5.0 **To reproduce** **Error when IAM user with all required permissions and only 1 region is enabled with Macie** ``` > select * from aws_aab.aws_macie2_classification_job Error: AccessDeniedException: Macie is not enabled. > select * from aws_aab.aws_macie2_classification_job where region = 'us-east-1' +--------------+----------------------------------+-------------------------------------------------------------------------------------------+------------+----------+--------------------------------------+------ | name | job_id | arn | job_status | job_type | client_token | creat +--------------+----------------------------------+-------------------------------------------------------------------------------------------+------------+----------+--------------------------------------+------ | bucket-audit | f9d91733ace0a9sdfdsfsdfsdfsd98916e | arn:aws:macie2:us-east-1:123493682495:classification-job/f9d91733ace0a93c09e8da13f598916e | COMPLETE | ONE_TIME | cf6b6918-9c3f-4a33-887b-f4f834a1046d | 2021- +--------------+----------------------------------+-------------------------------------------------------------------------------------------+------------+----------+--------------------------------------+------ ``` **The exception remains the same with a different message when a user with no IAM privileges, tries to query** ``` > select * from aws_macie2_classification_job Error: AccessDeniedException: User: arn:aws:iam::533793682495:user/macie-test is not authorized to perform: macie2:ListClassificationJobs on resource: arn:aws:macie2:ap-south-1:533793682495:* > .exit ``` **CLI Output** ``` turbot-macpro-raj:steampipe-mod-aws-top10 raj$ aws macie2 list-classification-jobs --profile devaab --region us-east-1 { "items": [ { "bucketDefinitions": [ { "accountId": "123453682495", "buckets": [ "andrew-turbot-test-bucket", "aws-logs-123453682495-us-east-1" ] } ], "createdAt": "2021-09-06T18:13:46.335482+00:00", "jobId": "f9d91733ace0a93c09e8da13f598916e", "jobStatus": "COMPLETE", "jobType": "ONE_TIME", "lastRunErrorStatus": { "code": "NONE" }, "name": "bucket-audit" } ] } turbot-macpro-raj:steampipe-mod-aws-top10 raj$ aws macie2 list-classification-jobs --profile devaab --region ap-south-1 An error occurred (AccessDeniedException) when calling the ListClassificationJobs operation: Macie is not enabled. turbot-macpro-raj:steampipe-mod-aws-top10 raj$ aws macie2 list-classification-jobs --profile devaab-onlyrds --region ap-south-1 An error occurred (AccessDeniedException) when calling the ListClassificationJobs operation: User: arn:aws:iam::123453682495:user/macie-test is not authorized to perform: macie2:ListClassificationJobs on resource: arn:aws:macie2:ap-south-1:123453682495:* turbot-macpro-raj:steampipe-mod-aws-top10 raj$ ``` **Expected behavior** Do we need to handle the message, not the exception? **Additional context** Add any other context about the problem here.
1.0
Handling exceptions aws_macie2_classification_job table - **Describe the bug** `aws_macie2_classification_job` table returns AccessDeniedException in case any one of the regions is configured in `aws.spc `file is not enabled with Macie. Also, provide the Macie service enabled/disabled status part of this table. **Steampipe version (`steampipe -v`)** Example: v0.3.0 **Plugin version (`steampipe plugin list`)** Example: v0.5.0 **To reproduce** **Error when IAM user with all required permissions and only 1 region is enabled with Macie** ``` > select * from aws_aab.aws_macie2_classification_job Error: AccessDeniedException: Macie is not enabled. > select * from aws_aab.aws_macie2_classification_job where region = 'us-east-1' +--------------+----------------------------------+-------------------------------------------------------------------------------------------+------------+----------+--------------------------------------+------ | name | job_id | arn | job_status | job_type | client_token | creat +--------------+----------------------------------+-------------------------------------------------------------------------------------------+------------+----------+--------------------------------------+------ | bucket-audit | f9d91733ace0a9sdfdsfsdfsdfsd98916e | arn:aws:macie2:us-east-1:123493682495:classification-job/f9d91733ace0a93c09e8da13f598916e | COMPLETE | ONE_TIME | cf6b6918-9c3f-4a33-887b-f4f834a1046d | 2021- +--------------+----------------------------------+-------------------------------------------------------------------------------------------+------------+----------+--------------------------------------+------ ``` **The exception remains the same with a different message when a user with no IAM privileges, tries to query** ``` > select * from aws_macie2_classification_job Error: AccessDeniedException: User: arn:aws:iam::533793682495:user/macie-test is not authorized to perform: macie2:ListClassificationJobs on resource: arn:aws:macie2:ap-south-1:533793682495:* > .exit ``` **CLI Output** ``` turbot-macpro-raj:steampipe-mod-aws-top10 raj$ aws macie2 list-classification-jobs --profile devaab --region us-east-1 { "items": [ { "bucketDefinitions": [ { "accountId": "123453682495", "buckets": [ "andrew-turbot-test-bucket", "aws-logs-123453682495-us-east-1" ] } ], "createdAt": "2021-09-06T18:13:46.335482+00:00", "jobId": "f9d91733ace0a93c09e8da13f598916e", "jobStatus": "COMPLETE", "jobType": "ONE_TIME", "lastRunErrorStatus": { "code": "NONE" }, "name": "bucket-audit" } ] } turbot-macpro-raj:steampipe-mod-aws-top10 raj$ aws macie2 list-classification-jobs --profile devaab --region ap-south-1 An error occurred (AccessDeniedException) when calling the ListClassificationJobs operation: Macie is not enabled. turbot-macpro-raj:steampipe-mod-aws-top10 raj$ aws macie2 list-classification-jobs --profile devaab-onlyrds --region ap-south-1 An error occurred (AccessDeniedException) when calling the ListClassificationJobs operation: User: arn:aws:iam::123453682495:user/macie-test is not authorized to perform: macie2:ListClassificationJobs on resource: arn:aws:macie2:ap-south-1:123453682495:* turbot-macpro-raj:steampipe-mod-aws-top10 raj$ ``` **Expected behavior** Do we need to handle the message, not the exception? **Additional context** Add any other context about the problem here.
priority
handling exceptions aws classification job table describe the bug aws classification job table returns accessdeniedexception in case any one of the regions is configured in aws spc file is not enabled with macie also provide the macie service enabled disabled status part of this table steampipe version steampipe v example plugin version steampipe plugin list example to reproduce error when iam user with all required permissions and only region is enabled with macie select from aws aab aws classification job error accessdeniedexception macie is not enabled select from aws aab aws classification job where region us east name job id arn job status job type client token creat bucket audit arn aws us east classification job complete one time the exception remains the same with a different message when a user with no iam privileges tries to query select from aws classification job error accessdeniedexception user arn aws iam user macie test is not authorized to perform listclassificationjobs on resource arn aws ap south exit cli output turbot macpro raj steampipe mod aws raj aws list classification jobs profile devaab region us east items bucketdefinitions accountid buckets andrew turbot test bucket aws logs us east createdat jobid jobstatus complete jobtype one time lastrunerrorstatus code none name bucket audit turbot macpro raj steampipe mod aws raj aws list classification jobs profile devaab region ap south an error occurred accessdeniedexception when calling the listclassificationjobs operation macie is not enabled turbot macpro raj steampipe mod aws raj aws list classification jobs profile devaab onlyrds region ap south an error occurred accessdeniedexception when calling the listclassificationjobs operation user arn aws iam user macie test is not authorized to perform listclassificationjobs on resource arn aws ap south turbot macpro raj steampipe mod aws raj expected behavior do we need to handle the message not the exception additional context add any other context about the problem here
1
399,125
11,743,235,603
IssuesEvent
2020-03-12 03:43:58
wso2/devstudio-tooling-ei
https://api.github.com/repos/wso2/devstudio-tooling-ei
closed
[Windows]When it is selecting the Datasource type as Carbon Data source, it is asking to select the database engine
DSS Priority/High Severity/Major
**Steps to reproduce:** 1. Create a Data Service project and right click on the project and add new Data Service. 2. Click to add a new data source, insert the below information and click on save. Data Source type: Carbon Database Datasource ID: 100 Datasource name: test ![screen1](https://user-images.githubusercontent.com/31693103/76382166-d1ca9d80-637d-11ea-82f0-c38f849af645.PNG) It is asked to select a database engine but no drop down to select the database engine.
1.0
[Windows]When it is selecting the Datasource type as Carbon Data source, it is asking to select the database engine - **Steps to reproduce:** 1. Create a Data Service project and right click on the project and add new Data Service. 2. Click to add a new data source, insert the below information and click on save. Data Source type: Carbon Database Datasource ID: 100 Datasource name: test ![screen1](https://user-images.githubusercontent.com/31693103/76382166-d1ca9d80-637d-11ea-82f0-c38f849af645.PNG) It is asked to select a database engine but no drop down to select the database engine.
priority
when it is selecting the datasource type as carbon data source it is asking to select the database engine steps to reproduce create a data service project and right click on the project and add new data service click to add a new data source insert the below information and click on save data source type carbon database datasource id datasource name test it is asked to select a database engine but no drop down to select the database engine
1
325,810
9,936,037,498
IssuesEvent
2019-07-02 18:04:44
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
POI/Poly merge during review fails when more than one POI is merged separately into the same poly
Category: Core Priority: High Type: Bug
As described here: https://github.com/DigitalGlobe/VGI-team-repo/issues/1912 Not sure if this is a core, merge service, or UI issue yet. If it ends up being a UI issue will move this to the hootenanny-ui repo.
1.0
POI/Poly merge during review fails when more than one POI is merged separately into the same poly - As described here: https://github.com/DigitalGlobe/VGI-team-repo/issues/1912 Not sure if this is a core, merge service, or UI issue yet. If it ends up being a UI issue will move this to the hootenanny-ui repo.
priority
poi poly merge during review fails when more than one poi is merged separately into the same poly as described here not sure if this is a core merge service or ui issue yet if it ends up being a ui issue will move this to the hootenanny ui repo
1
325,038
9,915,865,960
IssuesEvent
2019-06-28 18:09:00
dojot/dojot
https://api.github.com/repos/dojot/dojot
closed
[GUI] Cache problem in device creation
Priority:High Team:Frontend Type:Bug
**Steps to reproduce the problem:** 1. Create a template with static attribute 2. Create a device of this template and edit the static attribute 3. Click on device to edit it 4. Click on "New device", without discarding the edition 5. Create a new device from the same template and edit the static attribute with the same value as the previous device ![Screenshot at 2019-05-03 10-19-23](https://user-images.githubusercontent.com/37310063/57140796-05688e80-6d8f-11e9-9567-178aca811f22.png) ![Screenshot at 2019-05-03 10-19-32](https://user-images.githubusercontent.com/37310063/57140800-08fc1580-6d8f-11e9-90b7-10f5f3ed100a.png) **Affected Version:** 61.1-20190423
1.0
[GUI] Cache problem in device creation - **Steps to reproduce the problem:** 1. Create a template with static attribute 2. Create a device of this template and edit the static attribute 3. Click on device to edit it 4. Click on "New device", without discarding the edition 5. Create a new device from the same template and edit the static attribute with the same value as the previous device ![Screenshot at 2019-05-03 10-19-23](https://user-images.githubusercontent.com/37310063/57140796-05688e80-6d8f-11e9-9567-178aca811f22.png) ![Screenshot at 2019-05-03 10-19-32](https://user-images.githubusercontent.com/37310063/57140800-08fc1580-6d8f-11e9-90b7-10f5f3ed100a.png) **Affected Version:** 61.1-20190423
priority
cache problem in device creation steps to reproduce the problem create a template with static attribute create a device of this template and edit the static attribute click on device to edit it click on new device without discarding the edition create a new device from the same template and edit the static attribute with the same value as the previous device affected version
1
606,057
18,753,972,103
IssuesEvent
2021-11-05 08:15:26
AY2122S1-CS2113T-F14-1/tp
https://api.github.com/repos/AY2122S1-CS2113T-F14-1/tp
closed
[PE-D] Updated End Date is before start date
priority.High
As shown, after updating, end date can occur before start date ![3.JPG](https://raw.githubusercontent.com/Poopies99/ped/main/files/7c43d6d3-e225-4115-992a-2e699acf16c6.JPG) <!--session: 1635497050512-d62d4e36-ba20-4c91-b3f3-922f5292632d--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Medium` `type.FeatureFlaw` original: Poopies99/ped#5
1.0
[PE-D] Updated End Date is before start date - As shown, after updating, end date can occur before start date ![3.JPG](https://raw.githubusercontent.com/Poopies99/ped/main/files/7c43d6d3-e225-4115-992a-2e699acf16c6.JPG) <!--session: 1635497050512-d62d4e36-ba20-4c91-b3f3-922f5292632d--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Medium` `type.FeatureFlaw` original: Poopies99/ped#5
priority
updated end date is before start date as shown after updating end date can occur before start date labels severity medium type featureflaw original ped
1
329,148
10,012,739,117
IssuesEvent
2019-07-15 13:50:34
umple/umple
https://api.github.com/repos/umple/umple
closed
Graphviz views in UmpleOnline create extraneous URL_SPLIT message at bottom
Component-UmpleOnline Diffic-Easy Priority-High ucosp
If there is an error in the code, the following appears when displaying the Graphviz diagrams: "URL_SPLIT An error occurred interpreting your Umple code, please review it and try again. If the problem persists, please email the Umple code to the umple-help google group: umple-help@googlegroups.com" For example, enter the following in a tab: ``` class X { a; b; d; c; isA X; } ```
1.0
Graphviz views in UmpleOnline create extraneous URL_SPLIT message at bottom - If there is an error in the code, the following appears when displaying the Graphviz diagrams: "URL_SPLIT An error occurred interpreting your Umple code, please review it and try again. If the problem persists, please email the Umple code to the umple-help google group: umple-help@googlegroups.com" For example, enter the following in a tab: ``` class X { a; b; d; c; isA X; } ```
priority
graphviz views in umpleonline create extraneous url split message at bottom if there is an error in the code the following appears when displaying the graphviz diagrams url split an error occurred interpreting your umple code please review it and try again if the problem persists please email the umple code to the umple help google group umple help googlegroups com for example enter the following in a tab class x a b d c isa x
1
278,755
8,649,927,234
IssuesEvent
2018-11-26 20:55:09
norestlabs/colab
https://api.github.com/repos/norestlabs/colab
closed
Rename `Steem` to `Knowledgr`
DONE Priority:HIGH enhancement
Rename all instances of Steem to Knowledgr in the project. Note that Steem -> Knowledgr and steem -> knowledgr. Project should continue to compile afterwards.
1.0
Rename `Steem` to `Knowledgr` - Rename all instances of Steem to Knowledgr in the project. Note that Steem -> Knowledgr and steem -> knowledgr. Project should continue to compile afterwards.
priority
rename steem to knowledgr rename all instances of steem to knowledgr in the project note that steem knowledgr and steem knowledgr project should continue to compile afterwards
1
795,881
28,090,695,796
IssuesEvent
2023-03-30 12:53:32
AY2223S2-CS2103T-W15-1/tp
https://api.github.com/repos/AY2223S2-CS2103T-W15-1/tp
closed
As an exco of a CCA, I can find the exco of my CCA
type.Story priority.High
..so that I can plan an event for the members of the CCA with them
1.0
As an exco of a CCA, I can find the exco of my CCA - ..so that I can plan an event for the members of the CCA with them
priority
as an exco of a cca i can find the exco of my cca so that i can plan an event for the members of the cca with them
1
453,580
13,082,881,384
IssuesEvent
2020-08-01 15:58:32
passwall/passwall-desktop
https://api.github.com/repos/passwall/passwall-desktop
opened
Send imported values to serverside
Priority: High Status: Available Type: Enhancement
**Is your feature request related to a problem? Please describe.** Thanks to community now we have an import and parse csv handler function in **/src/renderer/App.vue**. ```js onImport() { dialog.showOpenDialog({ properties: ['openFile'] }, files => { if (files.length === 0) return try { const fileContent = fs.readFileSync(files[0]).toString() const parsedCSV = this.parseCSV(fileContent) // TODO: Send http request to add every line to database console.log('parsedCSV', parsedCSV) } catch (error) { console.log(error) } }) }, ``` **Describe the solution you'd like** We have an async create method in **/src/renderer/views/Logins/store.js**. We need to use a method like this in onImport function to add csv datas to database. ```js async Create({ rootState }, data) { data.username = this._vm.$helpers.encrypt(data.username, rootState.master_hash) data.password = this._vm.$helpers.encrypt(data.password, rootState.master_hash) // Encrypt payload with transmission key const payload = { data: this._vm.$helpers.aesEncrypt(JSON.stringify(data), rootState.transmission_key) } await LoginsService.Create(payload) }, ``` Does anyone wants to work on this?
1.0
Send imported values to serverside - **Is your feature request related to a problem? Please describe.** Thanks to community now we have an import and parse csv handler function in **/src/renderer/App.vue**. ```js onImport() { dialog.showOpenDialog({ properties: ['openFile'] }, files => { if (files.length === 0) return try { const fileContent = fs.readFileSync(files[0]).toString() const parsedCSV = this.parseCSV(fileContent) // TODO: Send http request to add every line to database console.log('parsedCSV', parsedCSV) } catch (error) { console.log(error) } }) }, ``` **Describe the solution you'd like** We have an async create method in **/src/renderer/views/Logins/store.js**. We need to use a method like this in onImport function to add csv datas to database. ```js async Create({ rootState }, data) { data.username = this._vm.$helpers.encrypt(data.username, rootState.master_hash) data.password = this._vm.$helpers.encrypt(data.password, rootState.master_hash) // Encrypt payload with transmission key const payload = { data: this._vm.$helpers.aesEncrypt(JSON.stringify(data), rootState.transmission_key) } await LoginsService.Create(payload) }, ``` Does anyone wants to work on this?
priority
send imported values to serverside is your feature request related to a problem please describe thanks to community now we have an import and parse csv handler function in src renderer app vue js onimport dialog showopendialog properties files if files length return try const filecontent fs readfilesync files tostring const parsedcsv this parsecsv filecontent todo send http request to add every line to database console log parsedcsv parsedcsv catch error console log error describe the solution you d like we have an async create method in src renderer views logins store js we need to use a method like this in onimport function to add csv datas to database js async create rootstate data data username this vm helpers encrypt data username rootstate master hash data password this vm helpers encrypt data password rootstate master hash encrypt payload with transmission key const payload data this vm helpers aesencrypt json stringify data rootstate transmission key await loginsservice create payload does anyone wants to work on this
1
548,791
16,075,608,259
IssuesEvent
2021-04-25 09:39:02
sopra-fs21-group-08/sopra-fs21-group08-client
https://api.github.com/repos/sopra-fs21-group-08/sopra-fs21-group08-client
closed
Ticket component. Create different buttons for the transport tickets and show the amount of tickets that are left.
high priority task
Time estimate: 3h This task is part of user story #17
1.0
Ticket component. Create different buttons for the transport tickets and show the amount of tickets that are left. - Time estimate: 3h This task is part of user story #17
priority
ticket component create different buttons for the transport tickets and show the amount of tickets that are left time estimate this task is part of user story
1
812,533
30,340,479,034
IssuesEvent
2023-07-11 12:27:08
markendley/pprjne
https://api.github.com/repos/markendley/pprjne
closed
Pricing Collection Calculator – Whitelisting issue? (see comment)
high priority
Client says: The embed code for this is likely not working as we need to whitelist the site in order for it to properly function, just FYI! Not sure what they mean here, do you have any ideas?
1.0
Pricing Collection Calculator – Whitelisting issue? (see comment) - Client says: The embed code for this is likely not working as we need to whitelist the site in order for it to properly function, just FYI! Not sure what they mean here, do you have any ideas?
priority
pricing collection calculator – whitelisting issue see comment client says the embed code for this is likely not working as we need to whitelist the site in order for it to properly function just fyi not sure what they mean here do you have any ideas
1
123,074
4,856,835,841
IssuesEvent
2016-11-12 08:31:40
bounswe/bounswe2016group4
https://api.github.com/repos/bounswe/bounswe2016group4
opened
Search feature implementation
android backend frontend priority-high
ignore my eating preferences and nutrition filter is enough for now in the advanced search panel.
1.0
Search feature implementation - ignore my eating preferences and nutrition filter is enough for now in the advanced search panel.
priority
search feature implementation ignore my eating preferences and nutrition filter is enough for now in the advanced search panel
1
479,623
13,803,913,682
IssuesEvent
2020-10-11 06:09:40
AY2021S1-CS2113T-F12-3/tp
https://api.github.com/repos/AY2021S1-CS2113T-F12-3/tp
closed
Simplify interfaces for coupling and cohesion
priority.High
Need to simplify so that our UML diagram is less complex
1.0
Simplify interfaces for coupling and cohesion - Need to simplify so that our UML diagram is less complex
priority
simplify interfaces for coupling and cohesion need to simplify so that our uml diagram is less complex
1
375,545
11,104,949,434
IssuesEvent
2019-12-17 08:50:33
PopupMaker/Popup-Maker
https://api.github.com/repos/PopupMaker/Popup-Maker
closed
Contact Form 7 prompting to save unsave changes when Popup Maker active
component-admin integration-forms priority-high scope-ux type-bug
![image](https://user-images.githubusercontent.com/847818/70979655-f9acac80-207f-11ea-94e1-92d61e905ad8.png) This is due to us setting form default value of open_popup_id to `0`, but the default select field value is `''`
1.0
Contact Form 7 prompting to save unsave changes when Popup Maker active - ![image](https://user-images.githubusercontent.com/847818/70979655-f9acac80-207f-11ea-94e1-92d61e905ad8.png) This is due to us setting form default value of open_popup_id to `0`, but the default select field value is `''`
priority
contact form prompting to save unsave changes when popup maker active this is due to us setting form default value of open popup id to but the default select field value is
1
699,346
24,013,085,267
IssuesEvent
2022-09-14 20:47:06
ConservationInternational/trends.earth
https://api.github.com/repos/ConservationInternational/trends.earth
closed
Problem installing plugin
Bug Priority 2 - High
I'm trying to install in QGIS 10.13 on Debian Linux. First I tried directly thru the plugin manager. This is the error that appeared when trying to load the plugin: ``` Couldn't load plugin 'LDMP' due to an error when calling its classFactory() method ImportError: /home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/summary_numba.so: undefined symbol: PyInstance_Type Traceback (most recent call last): File "/usr/lib/python3/dist-packages/qgis/utils.py", line 334, in _startPlugin plugins[packageName] = package.classFactory(iface) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/__init__.py", line 53, in classFactory return LDMPPlugin(iface) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/plugin.py", line 100, in __init__ self.dlg_calculate = DlgCalculate() File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate.py", line 437, in __init__ self.dlg_calculate_ld = DlgCalculateLD() File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate.py", line 474, in __init__ from LDMP.calculate_ldn import DlgCalculateOneStep, DlgCalculateLDNSummaryTableAdmin File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate_ldn.py", line 48, in from LDMP.summary_numba import merge_xtabs_i16, xtab_i16 File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) ImportError: /home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/summary_numba.so: undefined symbol: PyInstance_Type Python version: 3.8.6 (default, Sep 25 2020, 09:36:53) [GCC 10.2.0] QGIS version: 3.10.11-A Coruña A Coruña, exported ``` Then I downloaded the zipfile from [https://s3.amazonaws.com/trends.earth/sharing/LDMP_QGIS3.zip](url) and got a different error: ``` Couldn't load plugin 'LDMP' due to an error when calling its classFactory() method KeyError: 'Can only index numba types with slices with no start or stop, got 4.' Traceback (most recent call last): File "/usr/lib/python3/dist-packages/qgis/utils.py", line 334, in _startPlugin plugins[packageName] = package.classFactory(iface) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/__init__.py", line 49, in classFactory from LDMP.plugin import LDMPPlugin File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/plugin.py", line 24, in from LDMP.download_data import DlgDownload File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/download_data.py", line 30, in from LDMP.calculate import DlgCalculateBase, get_script_slug File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate.py", line 59, in from LDMP.calculate_numba import * File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate_numba.py", line 130, in @cc.export('ldn_total_deg', 'f4[4](i2[:,:], b1[:,:], f4[:,:])') File "/usr/lib/python3/dist-packages/numba/pycc/cc.py", line 136, in export fn_args, fn_retty = sigutils.normalize_signature(sig) File "/usr/lib/python3/dist-packages/numba/core/sigutils.py", line 24, in normalize_signature parsed = _parse_signature_string(sig) File "/usr/lib/python3/dist-packages/numba/core/sigutils.py", line 14, in _parse_signature_string return eval(signature_str, {}, types.__dict__) File "", line 1, in File "/usr/lib/python3/dist-packages/numba/core/types/abstract.py", line 186, in __getitem__ ndim, layout = self._determine_array_spec(args) File "/usr/lib/python3/dist-packages/numba/core/types/abstract.py", line 211, in _determine_array_spec raise KeyError(f"Can only index numba types with slices with no start or stop, got {args}.") KeyError: 'Can only index numba types with slices with no start or stop, got 4.' Python version: 3.8.6 (default, Sep 25 2020, 09:36:53) [GCC 10.2.0] QGIS version: 3.10.11-A Coruña A Coruña, exported ``` Any ideas how to proceed? Thanks
1.0
Problem installing plugin - I'm trying to install in QGIS 10.13 on Debian Linux. First I tried directly thru the plugin manager. This is the error that appeared when trying to load the plugin: ``` Couldn't load plugin 'LDMP' due to an error when calling its classFactory() method ImportError: /home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/summary_numba.so: undefined symbol: PyInstance_Type Traceback (most recent call last): File "/usr/lib/python3/dist-packages/qgis/utils.py", line 334, in _startPlugin plugins[packageName] = package.classFactory(iface) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/__init__.py", line 53, in classFactory return LDMPPlugin(iface) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/plugin.py", line 100, in __init__ self.dlg_calculate = DlgCalculate() File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate.py", line 437, in __init__ self.dlg_calculate_ld = DlgCalculateLD() File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate.py", line 474, in __init__ from LDMP.calculate_ldn import DlgCalculateOneStep, DlgCalculateLDNSummaryTableAdmin File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate_ldn.py", line 48, in from LDMP.summary_numba import merge_xtabs_i16, xtab_i16 File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) ImportError: /home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/summary_numba.so: undefined symbol: PyInstance_Type Python version: 3.8.6 (default, Sep 25 2020, 09:36:53) [GCC 10.2.0] QGIS version: 3.10.11-A Coruña A Coruña, exported ``` Then I downloaded the zipfile from [https://s3.amazonaws.com/trends.earth/sharing/LDMP_QGIS3.zip](url) and got a different error: ``` Couldn't load plugin 'LDMP' due to an error when calling its classFactory() method KeyError: 'Can only index numba types with slices with no start or stop, got 4.' Traceback (most recent call last): File "/usr/lib/python3/dist-packages/qgis/utils.py", line 334, in _startPlugin plugins[packageName] = package.classFactory(iface) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/__init__.py", line 49, in classFactory from LDMP.plugin import LDMPPlugin File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/plugin.py", line 24, in from LDMP.download_data import DlgDownload File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/download_data.py", line 30, in from LDMP.calculate import DlgCalculateBase, get_script_slug File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate.py", line 59, in from LDMP.calculate_numba import * File "/usr/lib/python3/dist-packages/qgis/utils.py", line 792, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/home/micha/.local/share/QGIS/QGIS3/profiles/default/python/plugins/LDMP/calculate_numba.py", line 130, in @cc.export('ldn_total_deg', 'f4[4](i2[:,:], b1[:,:], f4[:,:])') File "/usr/lib/python3/dist-packages/numba/pycc/cc.py", line 136, in export fn_args, fn_retty = sigutils.normalize_signature(sig) File "/usr/lib/python3/dist-packages/numba/core/sigutils.py", line 24, in normalize_signature parsed = _parse_signature_string(sig) File "/usr/lib/python3/dist-packages/numba/core/sigutils.py", line 14, in _parse_signature_string return eval(signature_str, {}, types.__dict__) File "", line 1, in File "/usr/lib/python3/dist-packages/numba/core/types/abstract.py", line 186, in __getitem__ ndim, layout = self._determine_array_spec(args) File "/usr/lib/python3/dist-packages/numba/core/types/abstract.py", line 211, in _determine_array_spec raise KeyError(f"Can only index numba types with slices with no start or stop, got {args}.") KeyError: 'Can only index numba types with slices with no start or stop, got 4.' Python version: 3.8.6 (default, Sep 25 2020, 09:36:53) [GCC 10.2.0] QGIS version: 3.10.11-A Coruña A Coruña, exported ``` Any ideas how to proceed? Thanks
priority
problem installing plugin i m trying to install in qgis on debian linux first i tried directly thru the plugin manager this is the error that appeared when trying to load the plugin couldn t load plugin ldmp due to an error when calling its classfactory method importerror home micha local share qgis profiles default python plugins ldmp summary numba so undefined symbol pyinstance type traceback most recent call last file usr lib dist packages qgis utils py line in startplugin plugins package classfactory iface file home micha local share qgis profiles default python plugins ldmp init py line in classfactory return ldmpplugin iface file home micha local share qgis profiles default python plugins ldmp plugin py line in init self dlg calculate dlgcalculate file home micha local share qgis profiles default python plugins ldmp calculate py line in init self dlg calculate ld dlgcalculateld file home micha local share qgis profiles default python plugins ldmp calculate py line in init from ldmp calculate ldn import dlgcalculateonestep dlgcalculateldnsummarytableadmin file usr lib dist packages qgis utils py line in import mod builtin import name globals locals fromlist level file home micha local share qgis profiles default python plugins ldmp calculate ldn py line in from ldmp summary numba import merge xtabs xtab file usr lib dist packages qgis utils py line in import mod builtin import name globals locals fromlist level importerror home micha local share qgis profiles default python plugins ldmp summary numba so undefined symbol pyinstance type python version default sep qgis version a coruña a coruña exported then i downloaded the zipfile from url and got a different error couldn t load plugin ldmp due to an error when calling its classfactory method keyerror can only index numba types with slices with no start or stop got traceback most recent call last file usr lib dist packages qgis utils py line in startplugin plugins package classfactory iface file home micha local share qgis profiles default python plugins ldmp init py line in classfactory from ldmp plugin import ldmpplugin file usr lib dist packages qgis utils py line in import mod builtin import name globals locals fromlist level file home micha local share qgis profiles default python plugins ldmp plugin py line in from ldmp download data import dlgdownload file usr lib dist packages qgis utils py line in import mod builtin import name globals locals fromlist level file home micha local share qgis profiles default python plugins ldmp download data py line in from ldmp calculate import dlgcalculatebase get script slug file usr lib dist packages qgis utils py line in import mod builtin import name globals locals fromlist level file home micha local share qgis profiles default python plugins ldmp calculate py line in from ldmp calculate numba import file usr lib dist packages qgis utils py line in import mod builtin import name globals locals fromlist level file home micha local share qgis profiles default python plugins ldmp calculate numba py line in cc export ldn total deg file usr lib dist packages numba pycc cc py line in export fn args fn retty sigutils normalize signature sig file usr lib dist packages numba core sigutils py line in normalize signature parsed parse signature string sig file usr lib dist packages numba core sigutils py line in parse signature string return eval signature str types dict file line in file usr lib dist packages numba core types abstract py line in getitem ndim layout self determine array spec args file usr lib dist packages numba core types abstract py line in determine array spec raise keyerror f can only index numba types with slices with no start or stop got args keyerror can only index numba types with slices with no start or stop got python version default sep qgis version a coruña a coruña exported any ideas how to proceed thanks
1
302,914
9,300,463,269
IssuesEvent
2019-03-23 14:01:33
cs2103-ay1819s2-w09-1/main
https://api.github.com/repos/cs2103-ay1819s2-w09-1/main
closed
Implement DeleteNote for NotesManager
priority.High v1.2
Allows user to delete a note which they specified in CLI.
1.0
Implement DeleteNote for NotesManager - Allows user to delete a note which they specified in CLI.
priority
implement deletenote for notesmanager allows user to delete a note which they specified in cli
1
525,213
15,240,811,168
IssuesEvent
2021-02-19 07:22:57
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
unifi.ui.com - desktop site instead of mobile site
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal
<!-- @browser: Firefox Mobile 86.0 --> <!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:86.0) Gecko/86.0 Firefox/86.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/67285 --> <!-- @extra_labels: browser-fenix --> **URL**: https://unifi.ui.com/device/e809ca0f-3053-4d45-8c40-332fc684e292/network/site/default/settings/networks/list **Browser / Version**: Firefox Mobile 86.0 **Operating System**: Android 11 **Tested Another Browser**: Yes Other **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: The website did not scale for mobile correctly <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210209185733</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/2/9bc147a6-b9a2-4671-8ff8-01c8df85ba40) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
unifi.ui.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 86.0 --> <!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:86.0) Gecko/86.0 Firefox/86.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/67285 --> <!-- @extra_labels: browser-fenix --> **URL**: https://unifi.ui.com/device/e809ca0f-3053-4d45-8c40-332fc684e292/network/site/default/settings/networks/list **Browser / Version**: Firefox Mobile 86.0 **Operating System**: Android 11 **Tested Another Browser**: Yes Other **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: The website did not scale for mobile correctly <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210209185733</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/2/9bc147a6-b9a2-4671-8ff8-01c8df85ba40) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
unifi ui com desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes other problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce the website did not scale for mobile correctly browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
408,792
11,952,006,964
IssuesEvent
2020-04-03 17:57:02
oneinfra/oneinfra
https://api.github.com/repos/oneinfra/oneinfra
closed
Add a CI pipeline per supported managed Kubernetes version
area/ci area/release kind/enhancement priority/high
Add a CI pipeline per supported managed Kubernetes version
1.0
Add a CI pipeline per supported managed Kubernetes version - Add a CI pipeline per supported managed Kubernetes version
priority
add a ci pipeline per supported managed kubernetes version add a ci pipeline per supported managed kubernetes version
1
362,052
10,722,392,892
IssuesEvent
2019-10-27 11:41:39
AY1920S1-CS2103T-F14-1/main
https://api.github.com/repos/AY1920S1-CS2103T-F14-1/main
closed
Add home page
priority.High severity.Medium status.Ongoing type.Story
- [x] Integrate profile page and its various tabs into one single home page - [x] Allow home page to pull actual data from storage and display useful information to user
1.0
Add home page - - [x] Integrate profile page and its various tabs into one single home page - [x] Allow home page to pull actual data from storage and display useful information to user
priority
add home page integrate profile page and its various tabs into one single home page allow home page to pull actual data from storage and display useful information to user
1
633,342
20,252,316,610
IssuesEvent
2022-02-14 19:10:49
GimelStudio/GimelStudio
https://api.github.com/repos/GimelStudio/GimelStudio
closed
Fix glitching on Properties Panel update
bug feature ui python high priority ready
Currently, the Properties Panel has a noticeable lag in refreshing/updating. This issue should optimize the panel so that only the necessary widgets are destroyed and re-created, or find another solution for updating altogether.
1.0
Fix glitching on Properties Panel update - Currently, the Properties Panel has a noticeable lag in refreshing/updating. This issue should optimize the panel so that only the necessary widgets are destroyed and re-created, or find another solution for updating altogether.
priority
fix glitching on properties panel update currently the properties panel has a noticeable lag in refreshing updating this issue should optimize the panel so that only the necessary widgets are destroyed and re created or find another solution for updating altogether
1
350,939
10,510,849,124
IssuesEvent
2019-09-27 14:16:13
input-output-hk/jormungandr
https://api.github.com/repos/input-output-hk/jormungandr
closed
Node crash with "free(): invalid pointer"
Priority - High bug jörmungandr
**Describe the bug** Last logs of the node (r-a-3) before crash: ``` Sep 24 09:02:59.365 WARN error connecting to peer: ConnectError(Http(Connect(Os { code: 111, kind: ConnectionRefused, message: "Connection refused" }))), peer_addr: 52.57.157.167:3000, task: network Sep 24 09:02:59.365 DEBG dropping I/O source: 13, scope: global Sep 24 09:02:59.385 DEBG received; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.386 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.386 DEBG send; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 16384 }, scope: global Sep 24 09:02:59.386 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.387 DEBG received; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.387 DEBG received; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.387 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.387 DEBG send; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 16384 }, scope: global Sep 24 09:02:59.387 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.388 DEBG send; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 32765 }, scope: global Sep 24 09:02:59.388 DEBG send; frame=WindowUpdate { stream_id: StreamId(7), size_increment: 49149 }, scope: global Sep 24 09:02:59.453 DEBG received; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 21762 }, scope: global Sep 24 09:02:59.454 DEBG received; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 32768 }, scope: global Sep 24 09:02:59.454 DEBG received; frame=WindowUpdate { stream_id: StreamId(7), size_increment: 32768 }, scope: global Sep 24 09:02:59.454 DEBG scheduling Read for: 31, scope: global Sep 24 09:02:59.454 DEBG send; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.455 DEBG send; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.542 DEBG received; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 59609 }, scope: global Sep 24 09:02:59.543 DEBG received; frame=WindowUpdate { stream_id: StreamId(7), size_increment: 59609 }, scope: global Sep 24 09:02:59.543 DEBG scheduling Read for: 22, scope: global Sep 24 09:02:59.552 DEBG sending gossip to node 16444025286634269777, task: network Sep 24 09:02:59.553 INFO gossip propagation to peer 16444025286634269777 failed: StreamOverflow, task: network Sep 24 09:02:59.553 DEBG unsubscribing peer 16444025286634269777, task: network Sep 24 09:02:59.553 DEBG connecting to node 16444025286634269777, peer_addr: 52.194.90.179:3000, task: network Sep 24 09:02:59.553 DEBG selecting gossips for 14548023685822520157, task: network Sep 24 09:02:59.554 DEBG scheduling Read for: 1, scope: global Sep 24 09:02:59.554 DEBG send; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.554 DEBG send; frame=Headers { stream_id: StreamId(3), flags: (0x5: END_HEADERS | END_STREAM) }, scope: global Sep 24 09:02:59.554 DEBG connecting to 52.194.90.179:3000, scope: global Sep 24 09:02:59.554 DEBG adding I/O source: 612368397, scope: global Sep 24 09:02:59.554 DEBG scheduling Write for: 13, scope: global Sep 24 09:02:59.554 DEBG scheduling Read for: 1, scope: global Sep 24 09:02:59.554 DEBG send; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.554 DEBG send; frame=Headers { stream_id: StreamId(5), flags: (0x5: END_HEADERS | END_STREAM) }, scope: global Sep 24 09:02:59.562 DEBG received; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.563 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.617 DEBG received gossip: Gossip { nodes: [Node(Node { id: Id([1, 80, 233, 83, 6, 221, 13, 54]), address: None, subscriptions: Subscriptions({Topic(1): High, Topic(0): High}), subscribers: {}, last_gossip: SystemTime { tv> , 179]), Id([9, 211, 198, 50, 93, 30, 248, 137]), Id([9, 222, 244, 17, 206, 49, 109, 91]), Id([9, 247, 44, 60, 128, 76, 158, 123]), Id([9, 251, 3, 195, 52, 40, 185, 67]), Id([9, 251, 134, 43, 250, 104, 246, 27]), Id([10, 7, 238, 63, 12, > ]), Id([41, 254, 187, 25, 80, 5, 138, 32]), Id([42, 12, 193, 32, 118, 144, 93, 180]), Id([42, 21, 235, 144, 163, 17, 104, 64]), Id([42, 25, 148, 44, 105, 169, 155, 196]), Id([42, 33, 110, 37, 208, 244, 224, 209]), Id([42, 45, 84, 189, 14> 95, 8, 38, 106, 231, 136]), Id([73, 114, 20, 218, 103, 241, 152, 5]), Id([73, 115, 15, 28, 119, 57, 134, 178]), Id([73, 119, 193, 250, 247, 157, 162, 99]), Id([73, 126, 73, 105, 6, 18, 208, 227]), Id([73, 127, 88, 221, 138, 213, 177, 81]> , 208, 245]), Id([105, 118, 0, 3, 99, 46, 245, 170]), Id([105, 128, 184, 164, 212, 1, 167, 57]), Id([105, 131, 123, 70, 15, 13, 209, 191]), Id([105, 143, 11, 52, 161, 133, 91, 33]), Id([105, 149, 199, 92, 184, 242, 159, 152]), Id([105, 1> ), Id([137, 152, 142, 121, 178, 254, 89, 199]), Id([137, 159, 231, 239, 79, 204, 23, 213]), Id([137, 171, 103, 185, 143, 79, 176, 249]), Id([137, 180, 30, 29, 57, 56, 179, 249]), Id([137, 180, 71, 200, 45, 4, 126, 67]), Id([137, 181, 119> 4, 51, 168, 11, 170, 207]), Id([168, 75, 138, 36, 247, 254, 239, 213]), Id([168, 77, 52, 233, 160, 67, 215, 183]), Id([168, 110, 179, 142, 187, 35, 206, 239]), Id([168, 115, 4, 206, 223, 71, 198, 56]), Id([168, 124, 112, 35, 163, 141, 12> 70, 158, 253, 107, 93, 109]), Id([202, 43, 96, 246, 215, 131, 121, 86]), Id([202, 49, 133, 59, 32, 41, 129, 178]), Id([202, 53, 5, 198, 114, 248, 168, 244]), Id([202, 65, 149, 205, 72, 168, 69, 57]), Id([202, 86, 115, 122, 41, 248, 73, > 3, 0, 212, 128, 70, 64]), Id([234, 122, 55, 187, 4, 99, 180, 228]), Id([234, 134, 120, 209, 61, 217, 225, 149]), Id([234, 139, 133, 26, 244, 153, 87, 244]), Id([234, 148, 238, 229, 182, 100, 161, 10]), Id([234, 166, 41, 149, 77, 244, 10,> Sep 24 09:02:59.624 DEBG updating P2P local topology, task: network free(): invalid pointer /nix/store/s80hdddnqg1i4jlfxni6k9z8bq4gx1yn-unit-script-jormungandr-start: line 5: 3567 Aborted (core dumped) RUST_BACKTRACE=1 /nix/store/v49ggr37n0lg0qq9bwflwqhm546ifa7d-jormungandr-0.5.0/bin/jormungandr --genesis-block> jormungandr.service: Main process exited, code=exited, status=134/n/a ``` **Mandatory Information** 1. `jcli --full-version` output; jcli 0.5.0 (, release, linux [x86_64]) - [rustc 1.37.0 (eae3437df 2019-08-13)] 2. `jormungandr --full-version` output; jormungandr 0.5.0 (, release, linux [x86_64]) - [rustc 1.37.0 (eae3437df 2019-08-13)]
1.0
Node crash with "free(): invalid pointer" - **Describe the bug** Last logs of the node (r-a-3) before crash: ``` Sep 24 09:02:59.365 WARN error connecting to peer: ConnectError(Http(Connect(Os { code: 111, kind: ConnectionRefused, message: "Connection refused" }))), peer_addr: 52.57.157.167:3000, task: network Sep 24 09:02:59.365 DEBG dropping I/O source: 13, scope: global Sep 24 09:02:59.385 DEBG received; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.386 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.386 DEBG send; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 16384 }, scope: global Sep 24 09:02:59.386 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.387 DEBG received; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.387 DEBG received; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.387 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.387 DEBG send; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 16384 }, scope: global Sep 24 09:02:59.387 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.388 DEBG send; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 32765 }, scope: global Sep 24 09:02:59.388 DEBG send; frame=WindowUpdate { stream_id: StreamId(7), size_increment: 49149 }, scope: global Sep 24 09:02:59.453 DEBG received; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 21762 }, scope: global Sep 24 09:02:59.454 DEBG received; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 32768 }, scope: global Sep 24 09:02:59.454 DEBG received; frame=WindowUpdate { stream_id: StreamId(7), size_increment: 32768 }, scope: global Sep 24 09:02:59.454 DEBG scheduling Read for: 31, scope: global Sep 24 09:02:59.454 DEBG send; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.455 DEBG send; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.542 DEBG received; frame=WindowUpdate { stream_id: StreamId(0), size_increment: 59609 }, scope: global Sep 24 09:02:59.543 DEBG received; frame=WindowUpdate { stream_id: StreamId(7), size_increment: 59609 }, scope: global Sep 24 09:02:59.543 DEBG scheduling Read for: 22, scope: global Sep 24 09:02:59.552 DEBG sending gossip to node 16444025286634269777, task: network Sep 24 09:02:59.553 INFO gossip propagation to peer 16444025286634269777 failed: StreamOverflow, task: network Sep 24 09:02:59.553 DEBG unsubscribing peer 16444025286634269777, task: network Sep 24 09:02:59.553 DEBG connecting to node 16444025286634269777, peer_addr: 52.194.90.179:3000, task: network Sep 24 09:02:59.553 DEBG selecting gossips for 14548023685822520157, task: network Sep 24 09:02:59.554 DEBG scheduling Read for: 1, scope: global Sep 24 09:02:59.554 DEBG send; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.554 DEBG send; frame=Headers { stream_id: StreamId(3), flags: (0x5: END_HEADERS | END_STREAM) }, scope: global Sep 24 09:02:59.554 DEBG connecting to 52.194.90.179:3000, scope: global Sep 24 09:02:59.554 DEBG adding I/O source: 612368397, scope: global Sep 24 09:02:59.554 DEBG scheduling Write for: 13, scope: global Sep 24 09:02:59.554 DEBG scheduling Read for: 1, scope: global Sep 24 09:02:59.554 DEBG send; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.554 DEBG send; frame=Headers { stream_id: StreamId(5), flags: (0x5: END_HEADERS | END_STREAM) }, scope: global Sep 24 09:02:59.562 DEBG received; frame=Data { stream_id: StreamId(7) }, scope: global Sep 24 09:02:59.563 DEBG scheduling Read for: 4, scope: global Sep 24 09:02:59.617 DEBG received gossip: Gossip { nodes: [Node(Node { id: Id([1, 80, 233, 83, 6, 221, 13, 54]), address: None, subscriptions: Subscriptions({Topic(1): High, Topic(0): High}), subscribers: {}, last_gossip: SystemTime { tv> , 179]), Id([9, 211, 198, 50, 93, 30, 248, 137]), Id([9, 222, 244, 17, 206, 49, 109, 91]), Id([9, 247, 44, 60, 128, 76, 158, 123]), Id([9, 251, 3, 195, 52, 40, 185, 67]), Id([9, 251, 134, 43, 250, 104, 246, 27]), Id([10, 7, 238, 63, 12, > ]), Id([41, 254, 187, 25, 80, 5, 138, 32]), Id([42, 12, 193, 32, 118, 144, 93, 180]), Id([42, 21, 235, 144, 163, 17, 104, 64]), Id([42, 25, 148, 44, 105, 169, 155, 196]), Id([42, 33, 110, 37, 208, 244, 224, 209]), Id([42, 45, 84, 189, 14> 95, 8, 38, 106, 231, 136]), Id([73, 114, 20, 218, 103, 241, 152, 5]), Id([73, 115, 15, 28, 119, 57, 134, 178]), Id([73, 119, 193, 250, 247, 157, 162, 99]), Id([73, 126, 73, 105, 6, 18, 208, 227]), Id([73, 127, 88, 221, 138, 213, 177, 81]> , 208, 245]), Id([105, 118, 0, 3, 99, 46, 245, 170]), Id([105, 128, 184, 164, 212, 1, 167, 57]), Id([105, 131, 123, 70, 15, 13, 209, 191]), Id([105, 143, 11, 52, 161, 133, 91, 33]), Id([105, 149, 199, 92, 184, 242, 159, 152]), Id([105, 1> ), Id([137, 152, 142, 121, 178, 254, 89, 199]), Id([137, 159, 231, 239, 79, 204, 23, 213]), Id([137, 171, 103, 185, 143, 79, 176, 249]), Id([137, 180, 30, 29, 57, 56, 179, 249]), Id([137, 180, 71, 200, 45, 4, 126, 67]), Id([137, 181, 119> 4, 51, 168, 11, 170, 207]), Id([168, 75, 138, 36, 247, 254, 239, 213]), Id([168, 77, 52, 233, 160, 67, 215, 183]), Id([168, 110, 179, 142, 187, 35, 206, 239]), Id([168, 115, 4, 206, 223, 71, 198, 56]), Id([168, 124, 112, 35, 163, 141, 12> 70, 158, 253, 107, 93, 109]), Id([202, 43, 96, 246, 215, 131, 121, 86]), Id([202, 49, 133, 59, 32, 41, 129, 178]), Id([202, 53, 5, 198, 114, 248, 168, 244]), Id([202, 65, 149, 205, 72, 168, 69, 57]), Id([202, 86, 115, 122, 41, 248, 73, > 3, 0, 212, 128, 70, 64]), Id([234, 122, 55, 187, 4, 99, 180, 228]), Id([234, 134, 120, 209, 61, 217, 225, 149]), Id([234, 139, 133, 26, 244, 153, 87, 244]), Id([234, 148, 238, 229, 182, 100, 161, 10]), Id([234, 166, 41, 149, 77, 244, 10,> Sep 24 09:02:59.624 DEBG updating P2P local topology, task: network free(): invalid pointer /nix/store/s80hdddnqg1i4jlfxni6k9z8bq4gx1yn-unit-script-jormungandr-start: line 5: 3567 Aborted (core dumped) RUST_BACKTRACE=1 /nix/store/v49ggr37n0lg0qq9bwflwqhm546ifa7d-jormungandr-0.5.0/bin/jormungandr --genesis-block> jormungandr.service: Main process exited, code=exited, status=134/n/a ``` **Mandatory Information** 1. `jcli --full-version` output; jcli 0.5.0 (, release, linux [x86_64]) - [rustc 1.37.0 (eae3437df 2019-08-13)] 2. `jormungandr --full-version` output; jormungandr 0.5.0 (, release, linux [x86_64]) - [rustc 1.37.0 (eae3437df 2019-08-13)]
priority
node crash with free invalid pointer describe the bug last logs of the node r a before crash sep warn error connecting to peer connecterror http connect os code kind connectionrefused message connection refused peer addr task network sep debg dropping i o source scope global sep debg received frame data stream id streamid scope global sep debg scheduling read for scope global sep debg send frame windowupdate stream id streamid size increment scope global sep debg scheduling read for scope global sep debg received frame data stream id streamid scope global sep debg received frame data stream id streamid scope global sep debg scheduling read for scope global sep debg send frame windowupdate stream id streamid size increment scope global sep debg scheduling read for scope global sep debg send frame windowupdate stream id streamid size increment scope global sep debg send frame windowupdate stream id streamid size increment scope global sep debg received frame windowupdate stream id streamid size increment scope global sep debg received frame windowupdate stream id streamid size increment scope global sep debg received frame windowupdate stream id streamid size increment scope global sep debg scheduling read for scope global sep debg send frame data stream id streamid scope global sep debg send frame data stream id streamid scope global sep debg received frame windowupdate stream id streamid size increment scope global sep debg received frame windowupdate stream id streamid size increment scope global sep debg scheduling read for scope global sep debg sending gossip to node task network sep info gossip propagation to peer failed streamoverflow task network sep debg unsubscribing peer task network sep debg connecting to node peer addr task network sep debg selecting gossips for task network sep debg scheduling read for scope global sep debg send frame data stream id streamid scope global sep debg send frame headers stream id streamid flags end headers end stream scope global sep debg connecting to scope global sep debg adding i o source scope global sep debg scheduling write for scope global sep debg scheduling read for scope global sep debg send frame data stream id streamid scope global sep debg send frame headers stream id streamid flags end headers end stream scope global sep debg received frame data stream id streamid scope global sep debg scheduling read for scope global sep debg received gossip gossip nodes address none subscriptions subscriptions topic high topic high subscribers last gossip systemtime tv id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id id sep debg updating local topology task network free invalid pointer nix store unit script jormungandr start line aborted core dumped rust backtrace nix store jormungandr bin jormungandr genesis block jormungandr service main process exited code exited status n a mandatory information jcli full version output jcli release linux jormungandr full version output jormungandr release linux
1
415,444
12,129,478,341
IssuesEvent
2020-04-22 22:42:06
autodo-app/autodo
https://api.github.com/repos/autodo-app/autodo
closed
Add a new todo at latestCompleted + interval after new account setup
bug crossplatform priority: high
Adding a completed todo in the new account setup screen should create a new todo for the given repeat.
1.0
Add a new todo at latestCompleted + interval after new account setup - Adding a completed todo in the new account setup screen should create a new todo for the given repeat.
priority
add a new todo at latestcompleted interval after new account setup adding a completed todo in the new account setup screen should create a new todo for the given repeat
1
209,283
7,167,193,950
IssuesEvent
2018-01-29 19:43:14
yandex/pandora
https://api.github.com/repos/yandex/pandora
opened
Incorrect signal handling
Priority: High Status: Available Type: Bug
Current signal handling have bugs and doesn't work properly. Neded: first SIGINT soft cancel or shooting. Any other - hard exit(1)
1.0
Incorrect signal handling - Current signal handling have bugs and doesn't work properly. Neded: first SIGINT soft cancel or shooting. Any other - hard exit(1)
priority
incorrect signal handling current signal handling have bugs and doesn t work properly neded first sigint soft cancel or shooting any other hard exit
1
677,168
23,153,328,443
IssuesEvent
2022-07-29 10:27:54
navikt/kviss
https://api.github.com/repos/navikt/kviss
closed
Feedback on result of choosen answer
frontend feature priority: high
When a participant has choosen their answer for a question, the result of choosen answer should be displayed.
1.0
Feedback on result of choosen answer - When a participant has choosen their answer for a question, the result of choosen answer should be displayed.
priority
feedback on result of choosen answer when a participant has choosen their answer for a question the result of choosen answer should be displayed
1
830,243
31,997,315,113
IssuesEvent
2023-09-21 09:59:51
flowforge/flowforge-nr-dashboard
https://api.github.com/repos/flowforge/flowforge-nr-dashboard
closed
/nodes/widgets Images are Broken
size:XS - 1 task bug area:docs priority:high
### Description Despite rendering locally, they do not render correctly on the production documentation: https://dashboard.flowfuse.com/nodes/widgets.html ### Epic/Story _No response_ ### Have you provided an initial effort estimate for this issue? I have provided an initial effort estimate
1.0
/nodes/widgets Images are Broken - ### Description Despite rendering locally, they do not render correctly on the production documentation: https://dashboard.flowfuse.com/nodes/widgets.html ### Epic/Story _No response_ ### Have you provided an initial effort estimate for this issue? I have provided an initial effort estimate
priority
nodes widgets images are broken description despite rendering locally they do not render correctly on the production documentation epic story no response have you provided an initial effort estimate for this issue i have provided an initial effort estimate
1
61,279
3,143,585,327
IssuesEvent
2015-09-14 08:08:52
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Cannot upload a files over a certain size to the script repository
Component: Framework Misc: Bugfix Priority: High State: Review Required
From James Lord: > Files 1,2,4 worked OK (and one .py_bck I uploaded as well by mistake :-) > > File 3, which I hadn't tried earlier, gives a new error (dialog box > "Upload Failed", "Bad answer from the Server") which happened again when > re-tried after successfully uploading file 4. > > ScriptRepository uploading > muon/Quantum/Algorithms_Autoload/quantumtabledriven.py ... > Uploading to folder: ./muon/Quantum/Algorithms_Autoload/ > Sending http POST request to: > http://upload.mantidproject.org/scriptrepository/payload/publish > Answer from web: 411 Length Required > Answer from web: 411 Length Required > ScriptRepository upload status: 411 > Form Output: <!DOCTYPE HTML PUBLIC "-//IETF//DTD HTML 2.0//EN"> > <html><head> > <title>411 Length Required</title> > </head><body> > <h1>Length Required</h1> > <p>A request of the requested method POST requires a valid Content- > length.<br /> </p> <hr> > <address>Apache/2.2.15 (Red Hat) Server at <a href="mailto:mantid- > webmaster@mantidproject.org">upload.mantidproject.org</a> Port > 80</address> </body></html> > > The file quantumtabledriven.py is the largest of the set, 73,527 bytes > on my computer (not empty!). The others are smaller (1, 7, and 55 kB).
1.0
Cannot upload a files over a certain size to the script repository - From James Lord: > Files 1,2,4 worked OK (and one .py_bck I uploaded as well by mistake :-) > > File 3, which I hadn't tried earlier, gives a new error (dialog box > "Upload Failed", "Bad answer from the Server") which happened again when > re-tried after successfully uploading file 4. > > ScriptRepository uploading > muon/Quantum/Algorithms_Autoload/quantumtabledriven.py ... > Uploading to folder: ./muon/Quantum/Algorithms_Autoload/ > Sending http POST request to: > http://upload.mantidproject.org/scriptrepository/payload/publish > Answer from web: 411 Length Required > Answer from web: 411 Length Required > ScriptRepository upload status: 411 > Form Output: <!DOCTYPE HTML PUBLIC "-//IETF//DTD HTML 2.0//EN"> > <html><head> > <title>411 Length Required</title> > </head><body> > <h1>Length Required</h1> > <p>A request of the requested method POST requires a valid Content- > length.<br /> </p> <hr> > <address>Apache/2.2.15 (Red Hat) Server at <a href="mailto:mantid- > webmaster@mantidproject.org">upload.mantidproject.org</a> Port > 80</address> </body></html> > > The file quantumtabledriven.py is the largest of the set, 73,527 bytes > on my computer (not empty!). The others are smaller (1, 7, and 55 kB).
priority
cannot upload a files over a certain size to the script repository from james lord files worked ok and one py bck i uploaded as well by mistake file which i hadn t tried earlier gives a new error dialog box upload failed bad answer from the server which happened again when re tried after successfully uploading file scriptrepository uploading muon quantum algorithms autoload quantumtabledriven py uploading to folder muon quantum algorithms autoload sending http post request to answer from web length required answer from web length required scriptrepository upload status form output length required length required a request of the requested method post requires a valid content length apache red hat server at a href mailto mantid webmaster mantidproject org upload mantidproject org port the file quantumtabledriven py is the largest of the set bytes on my computer not empty the others are smaller and kb
1
665,497
22,320,092,664
IssuesEvent
2022-06-14 05:06:40
opencrvs/opencrvs-core
https://api.github.com/repos/opencrvs/opencrvs-core
closed
Registration Agent title appearing on certificate
👹Bug Priority: high
**Describe the bug** A clear and concise description of what the bug is. **To Reproduce** Steps to reproduce the behaviour: 1. Login as a registration agent 2. Click on ready to print tab, select a record, download 3. Click print certificate, follow prompts 4. View the certificate 4. See error on the title below the signature **Expected behaviour** Only Registrar Title should appear in the certificates **Screenshots** ![Screenshot 2022-06-09 at 13.54.26.png](https://images.zenhubusercontent.com/61d3cc06859638b515717606/9962b6ce-d582-49b6-9317-a589c62e7952) https://www.loom.com/share/7ec1518881924f35a58c96bf2848a80d **Desktop (please complete the following information):** -chrome **Additional context** Add any other context about the problem here. **Possible fixes** (If you can, link to the line of code that might be responsible for the problem)
1.0
Registration Agent title appearing on certificate - **Describe the bug** A clear and concise description of what the bug is. **To Reproduce** Steps to reproduce the behaviour: 1. Login as a registration agent 2. Click on ready to print tab, select a record, download 3. Click print certificate, follow prompts 4. View the certificate 4. See error on the title below the signature **Expected behaviour** Only Registrar Title should appear in the certificates **Screenshots** ![Screenshot 2022-06-09 at 13.54.26.png](https://images.zenhubusercontent.com/61d3cc06859638b515717606/9962b6ce-d582-49b6-9317-a589c62e7952) https://www.loom.com/share/7ec1518881924f35a58c96bf2848a80d **Desktop (please complete the following information):** -chrome **Additional context** Add any other context about the problem here. **Possible fixes** (If you can, link to the line of code that might be responsible for the problem)
priority
registration agent title appearing on certificate describe the bug a clear and concise description of what the bug is to reproduce steps to reproduce the behaviour login as a registration agent click on ready to print tab select a record download click print certificate follow prompts view the certificate see error on the title below the signature expected behaviour only registrar title should appear in the certificates screenshots desktop please complete the following information chrome additional context add any other context about the problem here possible fixes if you can link to the line of code that might be responsible for the problem
1
403,317
11,839,144,067
IssuesEvent
2020-03-23 16:41:58
conan-io/conan
https://api.github.com/repos/conan-io/conan
closed
Feature: model intra-package dependencies
complex: high priority: medium stage: queue type: feature type: look into
From https://github.com/conan-io/conan/issues/2382 This could help both to model system dependencies ``self.cpp_info.libs = ["mylib", "pthread"]``, and then some consumer can depend on pthread by other means, and the order mylib->pthread is not defined. It can be generalized for your own libs deps "Mylib1"->"MyLib2" inside the same package.
1.0
Feature: model intra-package dependencies - From https://github.com/conan-io/conan/issues/2382 This could help both to model system dependencies ``self.cpp_info.libs = ["mylib", "pthread"]``, and then some consumer can depend on pthread by other means, and the order mylib->pthread is not defined. It can be generalized for your own libs deps "Mylib1"->"MyLib2" inside the same package.
priority
feature model intra package dependencies from this could help both to model system dependencies self cpp info libs and then some consumer can depend on pthread by other means and the order mylib pthread is not defined it can be generalized for your own libs deps inside the same package
1
98,268
4,019,296,424
IssuesEvent
2016-05-16 14:28:05
fgpv-vpgf/fgpv-vpgf
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
opened
Integrate Service Based Layer Import with UI Component
addition: feature priority: high
Integrate Esri & WMS based layer import code with UI component.
1.0
Integrate Service Based Layer Import with UI Component - Integrate Esri & WMS based layer import code with UI component.
priority
integrate service based layer import with ui component integrate esri wms based layer import code with ui component
1
534,055
15,608,395,651
IssuesEvent
2021-03-19 10:32:57
ballerina-platform/nballerina
https://api.github.com/repos/ballerina-platform/nballerina
opened
Incorrect output from float value
Priority/High Type/Bug
**Description:** $subject. **Steps to reproduce:** ``` public function printf32(float val) = external; public function main() { printf32(2.2); // prints wrong value like 4612136000000000000 } ```
1.0
Incorrect output from float value - **Description:** $subject. **Steps to reproduce:** ``` public function printf32(float val) = external; public function main() { printf32(2.2); // prints wrong value like 4612136000000000000 } ```
priority
incorrect output from float value description subject steps to reproduce public function float val external public function main prints wrong value like
1
727,845
25,048,355,210
IssuesEvent
2022-11-05 15:04:21
AY2223S1-CS2103T-W15-1/tp
https://api.github.com/repos/AY2223S1-CS2103T-W15-1/tp
closed
[PE-D][Tester B] Invalid command result for tag remove
bug priority.High
![image.png](https://raw.githubusercontent.com/PeiYee88/ped/main/files/0c1afba6-798e-4ebc-a412-ca4a2d44d063.png) Tag `family` does exist in the contact book but command result shows up `this tag does not exist` <!--session: 1666942944315-ae116179-fb18-4cf4-94b2-7dad168c8a9a--> <!--Version: Web v3.4.4--> ------------- Labels: `type.FunctionalityBug` `severity.Medium` original: PeiYee88/ped#15
1.0
[PE-D][Tester B] Invalid command result for tag remove - ![image.png](https://raw.githubusercontent.com/PeiYee88/ped/main/files/0c1afba6-798e-4ebc-a412-ca4a2d44d063.png) Tag `family` does exist in the contact book but command result shows up `this tag does not exist` <!--session: 1666942944315-ae116179-fb18-4cf4-94b2-7dad168c8a9a--> <!--Version: Web v3.4.4--> ------------- Labels: `type.FunctionalityBug` `severity.Medium` original: PeiYee88/ped#15
priority
invalid command result for tag remove tag family does exist in the contact book but command result shows up this tag does not exist labels type functionalitybug severity medium original ped
1
815,440
30,554,522,580
IssuesEvent
2023-07-20 10:44:47
Shelf-nu/shelf.nu
https://api.github.com/repos/Shelf-nu/shelf.nu
closed
[Bug]: Performing some action on asset index, disables clicking on table elements
bug High Priority
### Contact Details _No response_ ### What happened? Steps to reproduce: 1. Go to Asset index 2. Click on any asset 3. Delete the asset 4. You will be redirected back to `/assets` 5. Try clicking on any asset on the list and it won't work 6. Refreshing the page fixes the issue ### What is the expected behaviour? Being able to click on asset ### Version Hosted (app.shelf.nu) ### What browsers are you seeing the problem on? Firefox, Chrome, Safari, Microsoft Edge ### Relevant log output _No response_
1.0
[Bug]: Performing some action on asset index, disables clicking on table elements - ### Contact Details _No response_ ### What happened? Steps to reproduce: 1. Go to Asset index 2. Click on any asset 3. Delete the asset 4. You will be redirected back to `/assets` 5. Try clicking on any asset on the list and it won't work 6. Refreshing the page fixes the issue ### What is the expected behaviour? Being able to click on asset ### Version Hosted (app.shelf.nu) ### What browsers are you seeing the problem on? Firefox, Chrome, Safari, Microsoft Edge ### Relevant log output _No response_
priority
performing some action on asset index disables clicking on table elements contact details no response what happened steps to reproduce go to asset index click on any asset delete the asset you will be redirected back to assets try clicking on any asset on the list and it won t work refreshing the page fixes the issue what is the expected behaviour being able to click on asset version hosted app shelf nu what browsers are you seeing the problem on firefox chrome safari microsoft edge relevant log output no response
1
724,728
24,940,434,210
IssuesEvent
2022-10-31 18:27:26
geneontology/noctua-visual-pathway-editor
https://api.github.com/repos/geneontology/noctua-visual-pathway-editor
closed
Creating new model VPE on landing page
high priority
Add a way to create new model VPE on landing page Maybe ![image](https://user-images.githubusercontent.com/2273929/165428653-bdf85b24-d8a2-43f0-aca2-c75055229f20.png) tagging @pgaudet @thomaspd @vanaukenk this can be a landing page issue
1.0
Creating new model VPE on landing page - Add a way to create new model VPE on landing page Maybe ![image](https://user-images.githubusercontent.com/2273929/165428653-bdf85b24-d8a2-43f0-aca2-c75055229f20.png) tagging @pgaudet @thomaspd @vanaukenk this can be a landing page issue
priority
creating new model vpe on landing page add a way to create new model vpe on landing page maybe tagging pgaudet thomaspd vanaukenk this can be a landing page issue
1
133,540
5,205,246,007
IssuesEvent
2017-01-24 17:26:00
roundware/roundware-server
https://api.github.com/repos/roundware/roundware-server
closed
Add /api/v2/tagcategories
enhancement high priority now
For the new admin, I'll need `/api/v2/tagcategories` and `/api/v2/tagcategories/:id` routes. `GET` is fine for the moment. Ensuring that RESTful `POST` routes play well with Django's authentication is a separate issue, and ensuring that there is a token-based API `User` authentication for `POST`-ing is also a separate issue.
1.0
Add /api/v2/tagcategories - For the new admin, I'll need `/api/v2/tagcategories` and `/api/v2/tagcategories/:id` routes. `GET` is fine for the moment. Ensuring that RESTful `POST` routes play well with Django's authentication is a separate issue, and ensuring that there is a token-based API `User` authentication for `POST`-ing is also a separate issue.
priority
add api tagcategories for the new admin i ll need api tagcategories and api tagcategories id routes get is fine for the moment ensuring that restful post routes play well with django s authentication is a separate issue and ensuring that there is a token based api user authentication for post ing is also a separate issue
1
591,298
17,836,114,485
IssuesEvent
2021-09-03 01:28:40
harvester/harvester
https://api.github.com/repos/harvester/harvester
closed
[FEATURE] Replace K3OS
enhancement priority/1 highlight
We need to replace K3OS with an OS that can be supported by Rancher/SUSE. Disclaimer: Somehow this comment was picked up by people and was interpreted in an unintended way. So let me clarify: 1. We're replacing K3OS since K3OS has never been an officially supported OS by either Rancher or SUSE. We need something we can support as the base OS for the Harvester. 2. This change has **ABSOLUTELY NOTHING to do with the future of K3S**. On the one hand, K3S is a CNCF project, which is not even owned by Rancher/SUSE. On the other hand, K3S is one of the most popular K8s distros in the world, and people love it. There is zero reasons to be concerned about Rancher/SUSE's dedication to the K3S. 3. As you might have seen, Harvester is switching from K3S to [RKE2](https://docs.rke2.io/#how-is-this-different-from-rke-or-k3s), which is a technology that has a strong connection with K3S. K3S is in fact one of the upstream projects for RKE2. As we mentioned [here](https://docs.rke2.io/#how-is-this-different-from-rke-or-k3s): >How is this different from RKE or K3s? > >RKE2 combines the best-of-both-worlds from the 1.x version of RKE (hereafter referred to as RKE1) and K3s. > >From K3s, it inherits the usability, ease-of-operations, and deployment model. > >From RKE1, it inherits close alignment with upstream Kubernetes. In places K3s has diverged from upstream Kubernetes in order to optimize for edge deployments, but RKE1 and RKE2 can stay closely aligned with upstream. > >Importantly, RKE2 does not rely on Docker as RKE1 does. RKE1 leveraged Docker for deploying and managing the control plane components as well as the container runtime for Kubernetes. RKE2 launches control plane components as static pods, managed by the kubelet. The embedded container runtime is contained.
1.0
[FEATURE] Replace K3OS - We need to replace K3OS with an OS that can be supported by Rancher/SUSE. Disclaimer: Somehow this comment was picked up by people and was interpreted in an unintended way. So let me clarify: 1. We're replacing K3OS since K3OS has never been an officially supported OS by either Rancher or SUSE. We need something we can support as the base OS for the Harvester. 2. This change has **ABSOLUTELY NOTHING to do with the future of K3S**. On the one hand, K3S is a CNCF project, which is not even owned by Rancher/SUSE. On the other hand, K3S is one of the most popular K8s distros in the world, and people love it. There is zero reasons to be concerned about Rancher/SUSE's dedication to the K3S. 3. As you might have seen, Harvester is switching from K3S to [RKE2](https://docs.rke2.io/#how-is-this-different-from-rke-or-k3s), which is a technology that has a strong connection with K3S. K3S is in fact one of the upstream projects for RKE2. As we mentioned [here](https://docs.rke2.io/#how-is-this-different-from-rke-or-k3s): >How is this different from RKE or K3s? > >RKE2 combines the best-of-both-worlds from the 1.x version of RKE (hereafter referred to as RKE1) and K3s. > >From K3s, it inherits the usability, ease-of-operations, and deployment model. > >From RKE1, it inherits close alignment with upstream Kubernetes. In places K3s has diverged from upstream Kubernetes in order to optimize for edge deployments, but RKE1 and RKE2 can stay closely aligned with upstream. > >Importantly, RKE2 does not rely on Docker as RKE1 does. RKE1 leveraged Docker for deploying and managing the control plane components as well as the container runtime for Kubernetes. RKE2 launches control plane components as static pods, managed by the kubelet. The embedded container runtime is contained.
priority
replace we need to replace with an os that can be supported by rancher suse disclaimer somehow this comment was picked up by people and was interpreted in an unintended way so let me clarify we re replacing since has never been an officially supported os by either rancher or suse we need something we can support as the base os for the harvester this change has absolutely nothing to do with the future of on the one hand is a cncf project which is not even owned by rancher suse on the other hand is one of the most popular distros in the world and people love it there is zero reasons to be concerned about rancher suse s dedication to the as you might have seen harvester is switching from to which is a technology that has a strong connection with is in fact one of the upstream projects for as we mentioned how is this different from rke or combines the best of both worlds from the x version of rke hereafter referred to as and from it inherits the usability ease of operations and deployment model from it inherits close alignment with upstream kubernetes in places has diverged from upstream kubernetes in order to optimize for edge deployments but and can stay closely aligned with upstream importantly does not rely on docker as does leveraged docker for deploying and managing the control plane components as well as the container runtime for kubernetes launches control plane components as static pods managed by the kubelet the embedded container runtime is contained
1
471,699
13,608,444,451
IssuesEvent
2020-09-23 02:23:10
a2000-erp-team/WEBERP
https://api.github.com/repos/a2000-erp-team/WEBERP
opened
JAMES WILSON Sales Quotation is free text. Is there a setting for me to call out the inventory lists and all its attributes?
ABIGAIL High Priority
SALES-SALES-SALES QUOTATION ![image](https://user-images.githubusercontent.com/62376141/93957312-9f671780-fd86-11ea-8edd-77c4748045f5.png)
1.0
JAMES WILSON Sales Quotation is free text. Is there a setting for me to call out the inventory lists and all its attributes? - SALES-SALES-SALES QUOTATION ![image](https://user-images.githubusercontent.com/62376141/93957312-9f671780-fd86-11ea-8edd-77c4748045f5.png)
priority
james wilson sales quotation is free text is there a setting for me to call out the inventory lists and all its attributes sales sales sales quotation
1
490,337
14,118,391,981
IssuesEvent
2020-11-08 13:28:42
AY2021S1-CS2113-T13-4/tp
https://api.github.com/repos/AY2021S1-CS2113-T13-4/tp
closed
update addCommand to handle schedule when there is clash in timetable
priority.High
When user wants to add a task to the schedule manager and there is a clash in timings with another task, the schedule manager should inform the user about it and display the task that has clashes
1.0
update addCommand to handle schedule when there is clash in timetable - When user wants to add a task to the schedule manager and there is a clash in timings with another task, the schedule manager should inform the user about it and display the task that has clashes
priority
update addcommand to handle schedule when there is clash in timetable when user wants to add a task to the schedule manager and there is a clash in timings with another task the schedule manager should inform the user about it and display the task that has clashes
1
693,264
23,769,960,172
IssuesEvent
2022-09-01 15:30:29
Dessia-tech/dessia_common
https://api.github.com/repos/Dessia-tech/dessia_common
closed
Workflow copy breaks non block variables
type: bug priority: High status: done
**Note: for support questions, please use https://nextcloud.dessia.tech/call/hr9z9bif * **I'm submitting a ...** - [x] bug report - [ ] feature request * **What is the current behavior?** When copying a workflow, if a NBV is connected to several inputs, it is broke in several non_block_variables original : ![image](https://user-images.githubusercontent.com/42537234/153662242-63672e49-8b9c-4b1e-8c4a-2e1cd3dca490.png) copy : ![image](https://user-images.githubusercontent.com/42537234/153662349-07500c10-dc07-44dc-be68-544cb744d1f3.png) * **If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem** Avoid reference to other packages * **What is the expected behavior?** Copy without breaking NBVs * **What is the motivation / use case for changing the behavior?** It breaks the current (temporary ?) workflow state copying process. => If such a workflow and its copy are started and run with the same values, it doesn't give the same results and workflow state aren't equal anymore * **Possible fixes** * **Please tell us about your environment:** - branch: dev - commit: - python version: 3.9 Can be reproduced with forms_workflow. The minimal step is to copy it
1.0
Workflow copy breaks non block variables - **Note: for support questions, please use https://nextcloud.dessia.tech/call/hr9z9bif * **I'm submitting a ...** - [x] bug report - [ ] feature request * **What is the current behavior?** When copying a workflow, if a NBV is connected to several inputs, it is broke in several non_block_variables original : ![image](https://user-images.githubusercontent.com/42537234/153662242-63672e49-8b9c-4b1e-8c4a-2e1cd3dca490.png) copy : ![image](https://user-images.githubusercontent.com/42537234/153662349-07500c10-dc07-44dc-be68-544cb744d1f3.png) * **If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem** Avoid reference to other packages * **What is the expected behavior?** Copy without breaking NBVs * **What is the motivation / use case for changing the behavior?** It breaks the current (temporary ?) workflow state copying process. => If such a workflow and its copy are started and run with the same values, it doesn't give the same results and workflow state aren't equal anymore * **Possible fixes** * **Please tell us about your environment:** - branch: dev - commit: - python version: 3.9 Can be reproduced with forms_workflow. The minimal step is to copy it
priority
workflow copy breaks non block variables note for support questions please use i m submitting a bug report feature request what is the current behavior when copying a workflow if a nbv is connected to several inputs it is broke in several non block variables original copy if the current behavior is a bug please provide the steps to reproduce and if possible a minimal demo of the problem avoid reference to other packages what is the expected behavior copy without breaking nbvs what is the motivation use case for changing the behavior it breaks the current temporary workflow state copying process if such a workflow and its copy are started and run with the same values it doesn t give the same results and workflow state aren t equal anymore possible fixes please tell us about your environment branch dev commit python version can be reproduced with forms workflow the minimal step is to copy it
1
603,116
18,527,776,167
IssuesEvent
2021-10-20 23:19:39
DrakeHunterK7/SOEN-341-Fall-2021-Team-9-Stack-Bubbling-
https://api.github.com/repos/DrakeHunterK7/SOEN-341-Fall-2021-Team-9-Stack-Bubbling-
closed
Travis Ci set up
Env Setup Task Low Risk High Value 1 point Priority: Low
this is one of the requirements from the grading schema For the current stage, we don't need to write any test case, set up the env only.
1.0
Travis Ci set up - this is one of the requirements from the grading schema For the current stage, we don't need to write any test case, set up the env only.
priority
travis ci set up this is one of the requirements from the grading schema for the current stage we don t need to write any test case set up the env only
1
606,127
18,755,400,233
IssuesEvent
2021-11-05 10:05:08
wasmerio/wasmer
https://api.github.com/repos/wasmerio/wasmer
closed
CLI args in run subcommand: wasm module can't follow `--dir` or `--mapdir` with no `=`
❓ question 📦 lib-cli priority-high
<!-- Thanks for the bug report! --> ### Describe the bug `<FILE>` has to be first for options to work. <!-- A clear and concise description of what the bug is. Copy and paste the result of executing the following in your shell, so we can know the version of wasmer, Rust (if available) and architecture of your environment. --> ```sh wasmer 2.0.0 | rustc 1.52.1 (9bc8c42bb 2021-05-09) | x86_64 ``` ### Steps to reproduce ```sh wasmer run --dir foo bar.wasm ``` ### Expected behavior Program runs with the given directory ### Actual behavior The `<FILE>` is interpreted as another directory. ``` error: The following required arguments were not provided: <FILE> USAGE: wasmer run <FILE> <--dir <DIR>...> ```
1.0
CLI args in run subcommand: wasm module can't follow `--dir` or `--mapdir` with no `=` - <!-- Thanks for the bug report! --> ### Describe the bug `<FILE>` has to be first for options to work. <!-- A clear and concise description of what the bug is. Copy and paste the result of executing the following in your shell, so we can know the version of wasmer, Rust (if available) and architecture of your environment. --> ```sh wasmer 2.0.0 | rustc 1.52.1 (9bc8c42bb 2021-05-09) | x86_64 ``` ### Steps to reproduce ```sh wasmer run --dir foo bar.wasm ``` ### Expected behavior Program runs with the given directory ### Actual behavior The `<FILE>` is interpreted as another directory. ``` error: The following required arguments were not provided: <FILE> USAGE: wasmer run <FILE> <--dir <DIR>...> ```
priority
cli args in run subcommand wasm module can t follow dir or mapdir with no describe the bug has to be first for options to work a clear and concise description of what the bug is copy and paste the result of executing the following in your shell so we can know the version of wasmer rust if available and architecture of your environment sh wasmer rustc steps to reproduce sh wasmer run dir foo bar wasm expected behavior program runs with the given directory actual behavior the is interpreted as another directory error the following required arguments were not provided usage wasmer run
1
361,767
10,720,070,876
IssuesEvent
2019-10-26 15:04:56
apache/incubator-echarts
https://api.github.com/repos/apache/incubator-echarts
closed
Bar chart: Log y-Axis not working
bug priority: high
### Version 4.4.0-rc.1 ### Reproduction link [https://jsfiddle.net/tmtron/4fguo0wn/5/](https://jsfiddle.net/tmtron/4fguo0wn/5/) ### Steps to reproduce Just execute the jsfiddle example: it does not show any bars ![image](https://user-images.githubusercontent.com/20422904/66051177-be91e700-e52e-11e9-9129-4050d6e5c61d.png) ### What is expected? It should show bars on the logarithmic y-axis. ![image](https://user-images.githubusercontent.com/20422904/66051203-c8b3e580-e52e-11e9-9c03-20cb428195b4.png) ### What is actually happening? No bars are shown --- The same options worked fine in version 4.3.0: see this [js-fiddle example](https://jsfiddle.net/tmtron/4fguo0wn/4/): <!-- This issue is generated by echarts-issue-helper. DO NOT REMOVE --> <!-- This issue is in English. DO NOT REMOVE -->
1.0
Bar chart: Log y-Axis not working - ### Version 4.4.0-rc.1 ### Reproduction link [https://jsfiddle.net/tmtron/4fguo0wn/5/](https://jsfiddle.net/tmtron/4fguo0wn/5/) ### Steps to reproduce Just execute the jsfiddle example: it does not show any bars ![image](https://user-images.githubusercontent.com/20422904/66051177-be91e700-e52e-11e9-9129-4050d6e5c61d.png) ### What is expected? It should show bars on the logarithmic y-axis. ![image](https://user-images.githubusercontent.com/20422904/66051203-c8b3e580-e52e-11e9-9c03-20cb428195b4.png) ### What is actually happening? No bars are shown --- The same options worked fine in version 4.3.0: see this [js-fiddle example](https://jsfiddle.net/tmtron/4fguo0wn/4/): <!-- This issue is generated by echarts-issue-helper. DO NOT REMOVE --> <!-- This issue is in English. DO NOT REMOVE -->
priority
bar chart log y axis not working version rc reproduction link steps to reproduce just execute the jsfiddle example it does not show any bars what is expected it should show bars on the logarithmic y axis what is actually happening no bars are shown the same options worked fine in version see this
1
315,487
9,621,259,372
IssuesEvent
2019-05-14 10:11:52
aau-giraf/weekplanner
https://api.github.com/repos/aau-giraf/weekplanner
closed
As a guardian I would like the citizens name and name of weekplan to be shown in the appbar so that I know which citizen I'm currently on
priority: highest type: refactor
Especially in choose weekplan and the overview
1.0
As a guardian I would like the citizens name and name of weekplan to be shown in the appbar so that I know which citizen I'm currently on - Especially in choose weekplan and the overview
priority
as a guardian i would like the citizens name and name of weekplan to be shown in the appbar so that i know which citizen i m currently on especially in choose weekplan and the overview
1
286,678
8,791,649,128
IssuesEvent
2018-12-21 13:30:41
projectacrn/acrn-hypervisor
https://api.github.com/repos/projectacrn/acrn-hypervisor
closed
Mediator Support for KDI Virtualization
area: hypervisor priority: high status: new type: feature
The CSME mediator shall enable virtualization support of the KDI interface to enable DAL applications on multiple Guest OS.
1.0
Mediator Support for KDI Virtualization - The CSME mediator shall enable virtualization support of the KDI interface to enable DAL applications on multiple Guest OS.
priority
mediator support for kdi virtualization the csme mediator shall enable virtualization support of the kdi interface to enable dal applications on multiple guest os
1
544,169
15,890,476,696
IssuesEvent
2021-04-10 15:35:02
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
opened
Black screen appears while loading the Production Keys for Forgerock KM
API-M 4.0.0 Priority/High React-UI Type/Bug
### Description: Black screen appears when trying to load production keys of the application under Forgerock Key Manager. ### Steps to reproduce: - Create an External KM for Forgeock - Try to navigate to Forgerock Km tab under Production Keys ### Affected Product Version: 4.0
1.0
Black screen appears while loading the Production Keys for Forgerock KM - ### Description: Black screen appears when trying to load production keys of the application under Forgerock Key Manager. ### Steps to reproduce: - Create an External KM for Forgeock - Try to navigate to Forgerock Km tab under Production Keys ### Affected Product Version: 4.0
priority
black screen appears while loading the production keys for forgerock km description black screen appears when trying to load production keys of the application under forgerock key manager steps to reproduce create an external km for forgeock try to navigate to forgerock km tab under production keys affected product version
1
162,987
6,187,151,179
IssuesEvent
2017-07-04 06:28:05
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
Implementation of ZTP based ONIE switch discovery and configuration
component:discovery component:network priority:high sprint2 type:feature
Based on the result of https://github.com/xcat2/xcat-core/issues/3060 Acceptance: 1. finish the implementation of the mandatory TODO features listed in the mini-design https://github.com/xcat2/xcat-core/wiki/mini-design-of-ZTP-based-onie-switch-discovery-and-configuration , the "nice to have" won't be covered by this task. 1.1 the implementation of autodiscover 1.2 the implementation of basic out of box configuration 2. finish function verification 3. code merge to master branch 4. finish documentation[Low priority] The customer requirement available at #2662
1.0
Implementation of ZTP based ONIE switch discovery and configuration - Based on the result of https://github.com/xcat2/xcat-core/issues/3060 Acceptance: 1. finish the implementation of the mandatory TODO features listed in the mini-design https://github.com/xcat2/xcat-core/wiki/mini-design-of-ZTP-based-onie-switch-discovery-and-configuration , the "nice to have" won't be covered by this task. 1.1 the implementation of autodiscover 1.2 the implementation of basic out of box configuration 2. finish function verification 3. code merge to master branch 4. finish documentation[Low priority] The customer requirement available at #2662
priority
implementation of ztp based onie switch discovery and configuration based on the result of acceptance finish the implementation of the mandatory todo features listed in the mini design the nice to have won t be covered by this task the implementation of autodiscover the implementation of basic out of box configuration finish function verification code merge to master branch finish documentation the customer requirement available at
1
352,495
10,542,852,780
IssuesEvent
2019-10-02 13:58:00
onaio/reveal-frontend
https://api.github.com/repos/onaio/reveal-frontend
closed
IRS Reporting map breaks when there is no data in Superset
Priority: High
To replicate, try to load: Namibia IRS Test Plan > Namibia > Oshana > Oshakati > Hanover
1.0
IRS Reporting map breaks when there is no data in Superset - To replicate, try to load: Namibia IRS Test Plan > Namibia > Oshana > Oshakati > Hanover
priority
irs reporting map breaks when there is no data in superset to replicate try to load namibia irs test plan namibia oshana oshakati hanover
1
347,341
10,428,697,832
IssuesEvent
2019-09-16 23:36:05
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
DataTable StocktakePage fixes
Bug: development DataTable Docs: not needed Effort: small Priority: High Refactor
## Describe the bug Please see PR comments in #1190 ### To reproduce N/A ### Expected behaviour N/A ### Additional context Issue is to make the changes from requested changes in #1190
1.0
DataTable StocktakePage fixes - ## Describe the bug Please see PR comments in #1190 ### To reproduce N/A ### Expected behaviour N/A ### Additional context Issue is to make the changes from requested changes in #1190
priority
datatable stocktakepage fixes describe the bug please see pr comments in to reproduce n a expected behaviour n a additional context issue is to make the changes from requested changes in
1
155,573
5,957,113,982
IssuesEvent
2017-05-28 23:09:42
jhpoelen/fb-osmose-bridge
https://api.github.com/repos/jhpoelen/fb-osmose-bridge
opened
UI Step 2 - Issue no. 15
UI High priority
@FIN-casey One of the information button provides the following information: "To remove all the species belonging to the same class from a given functional group, deselect the class from the column "Class" using the dedicated radio button; this action will deselect all the species belonging to the class from the column "Species" of the functional group." However, it is currently not possible for the user to implement this. Could you please add radio buttons in the "Class" column and make sure that, if the user deselects the class from the column "Class" using the dedicated radio button, this results in all the species belonging to the class from the column "Species" of the functional group" being deselected? Many thanks!
1.0
UI Step 2 - Issue no. 15 - @FIN-casey One of the information button provides the following information: "To remove all the species belonging to the same class from a given functional group, deselect the class from the column "Class" using the dedicated radio button; this action will deselect all the species belonging to the class from the column "Species" of the functional group." However, it is currently not possible for the user to implement this. Could you please add radio buttons in the "Class" column and make sure that, if the user deselects the class from the column "Class" using the dedicated radio button, this results in all the species belonging to the class from the column "Species" of the functional group" being deselected? Many thanks!
priority
ui step issue no fin casey one of the information button provides the following information to remove all the species belonging to the same class from a given functional group deselect the class from the column class using the dedicated radio button this action will deselect all the species belonging to the class from the column species of the functional group however it is currently not possible for the user to implement this could you please add radio buttons in the class column and make sure that if the user deselects the class from the column class using the dedicated radio button this results in all the species belonging to the class from the column species of the functional group being deselected many thanks
1
758,773
26,568,454,715
IssuesEvent
2023-01-20 23:10:11
OpenLiberty/liberty-language-server
https://api.github.com/repos/OpenLiberty/liberty-language-server
closed
Exception when trying to use completion function from a beta runtime artifact (JAXB-related)
bug high priority should-fix
Recreate - 1. Install Liberty Tools Eclipse (Early Release) 2. Configure LMP to use beta runtime artifact ```xml <configuration> <runtimeArtifact> <groupId>io.openliberty.beta</groupId> <artifactId>openliberty-runtime</artifactId> <version>LATEST</version> <type>zip</type> </runtimeArtifact> </configuration> ``` 3. Do a `mvn liberty:create` (otherwise the runtime will be irrelevant and we'll go down the default feature path) 4. Open a server.xml, position the cursor with the contents of a `<feature>xyz-1.0</feature>` element and type `Ctrl+Space`. In the debugger I see NullPointerException thrown here: ``` Injector.inject(String, byte[]) line: 294 Injector.inject(ClassLoader, String, byte[]) line: 66 AccessorInjector.prepare(Class, String, String, String...) line: 57 OptimizedAccessorFactory.get(Field) line: 157 Accessor$FieldReflection<BeanT,ValueT>.optimize(JAXBContextImpl) line: 255 SingleElementNodeProperty<BeanT,ValueT>.<init>(JAXBContextImpl, RuntimeElementPropertyInfo) line: 64 NativeConstructorAccessorImpl.newInstance0(Constructor<?>, Object[]) line: not available [native method] NativeConstructorAccessorImpl.newInstance(Object[]) line: 77 DelegatingConstructorAccessorImpl.newInstance(Object[]) line: 45 Constructor<T>.newInstanceWithCaller(Object[], boolean, Class<?>) line: 499 Constructor<T>.newInstance(Object...) line: 480 PropertyFactory.create(JAXBContextImpl, RuntimePropertyInfo) line: 99 ClassBeanInfoImpl<BeanT>.<init>(JAXBContextImpl, RuntimeClassInfo) line: 150 JAXBContextImpl.getOrCreate(RuntimeClassInfo) line: 493 JAXBContextImpl.getOrCreate(RuntimeTypeInfo) line: 512 ArrayElementNodeProperty<BeanT,ListT,ItemT>(ArrayElementProperty<BeanT,ListT,ItemT>).<init>(JAXBContextImpl, RuntimeElementPropertyInfo) line: 82 ArrayElementNodeProperty<BeanT,ListT,ItemT>.<init>(JAXBContextImpl, RuntimeElementPropertyInfo) line: 32 NativeConstructorAccessorImpl.newInstance0(Constructor<?>, Object[]) line: not available [native method] NativeConstructorAccessorImpl.newInstance(Object[]) line: 77 DelegatingConstructorAccessorImpl.newInstance(Object[]) line: 45 Constructor<T>.newInstanceWithCaller(Object[], boolean, Class<?>) line: 499 Constructor<T>.newInstance(Object...) line: 480 PropertyFactory.create(JAXBContextImpl, RuntimePropertyInfo) line: 99 ClassBeanInfoImpl<BeanT>.<init>(JAXBContextImpl, RuntimeClassInfo) line: 150 JAXBContextImpl.getOrCreate(RuntimeClassInfo) line: 493 JAXBContextImpl.<init>(JAXBContextImpl$JAXBContextBuilder) line: 310 JAXBContextImpl.<init>(JAXBContextImpl$JAXBContextBuilder, JAXBContextImpl$1) line: 109 JAXBContextImpl$JAXBContextBuilder.build() line: 1142 ContextFactory.createContext(Class[], Map<String,Object>) line: 141 NativeMethodAccessorImpl.invoke0(Method, Object, Object[]) line: not available [native method] NativeMethodAccessorImpl.invoke(Object, Object[]) line: 77 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43 Method.invoke(Object, Object...) line: 568 ContextFinder.newInstance(Class[], Map, String) line: 211 ContextFinder.find(Class[], Map) line: 372 JAXBContext.newInstance(Class[], Map<String,?>) line: 574 JAXBContext.newInstance(Class...) line: 522 FeatureService.getInstalledFeaturesList(String, String, String) line: 251 FeatureService.getFeatures(String, String, int, String) line: 183 FeatureService.getFeature(String, String, String, int, String) line: 194 LibertyHoverParticipant.getHoverFeatureDescription(String, DOMDocument) line: 70 LibertyHoverParticipant.onText(IHoverRequest) line: 57 XMLHover.getTextHover(HoverRequest, Range) line: 207 XMLHover.doHover(DOMDocument, Position, SharedSettings, CancelChecker) line: 99 XMLLanguageService.doHover(DOMDocument, Position, SharedSettings, CancelChecker) line: 167 XMLTextDocumentService.lambda$hover$5(HoverParams, DOMDocument, CancelChecker) line: 260 0x000000080047e230.apply(Object, Object) line: not available ModelTextDocuments<T>.lambda$computeModelAsync$0(TextDocumentIdentifier, BiFunction, CancelChecker) line: 116 0x000000080016c590.apply(Object) line: not available CompletableFuture$UniApply<T,V>.tryFire(int) line: 646 CompletableFuture$UniApply<T,V>(CompletableFuture$Completion).exec() line: 483 CompletableFuture$UniApply<T,V>(ForkJoinTask<V>).doExec() line: 373 ForkJoinPool$WorkQueue.topLevelExec(ForkJoinTask<?>, WorkQueue) line: 1182 ForkJoinPool.scan(ForkJoinPool$WorkQueue, int, int) line: 1655 ForkJoinPool.runWorker(ForkJoinPool$WorkQueue) line: 1622 ForkJoinWorkerThread.run() line: 165 ``` which gets caught in stack > java.lang.NullPointerException: Cannot invoke "java.lang.reflect.Method.invoke(Object, Object[])" because "com.sun.xml.bind.v2.runtime.reflect.opt.Injector.defineClass" is null caught in stack: ``` XMLHover.getTextHover(HoverRequest, Range) line: 212 XMLHover.doHover(DOMDocument, Position, SharedSettings, CancelChecker) line: 99 XMLLanguageService.doHover(DOMDocument, Position, SharedSettings, CancelChecker) line: 167 XMLTextDocumentService.lambda$hover$5(HoverParams, DOMDocument, CancelChecker) line: 260 0x000000080047e230.apply(Object, Object) line: not available ModelTextDocuments<T>.lambda$computeModelAsync$0(TextDocumentIdentifier, BiFunction, CancelChecker) line: 116 0x000000080016c590.apply(Object) line: not available CompletableFuture$UniApply<T,V>.tryFire(int) line: 646 CompletableFuture$UniApply<T,V>(CompletableFuture$Completion).exec() line: 483 CompletableFuture$UniApply<T,V>(ForkJoinTask<V>).doExec() line: 373 ForkJoinPool$WorkQueue.topLevelExec(ForkJoinTask<?>, WorkQueue) line: 1182 ForkJoinPool.scan(ForkJoinPool$WorkQueue, int, int) line: 1655 ForkJoinPool.runWorker(ForkJoinPool$WorkQueue) line: 1622 ForkJoinWorkerThread.run() line: 165 ``` Note: we are running the LCLS with Java 17 in Liberty Tools Eclipse.
1.0
Exception when trying to use completion function from a beta runtime artifact (JAXB-related) - Recreate - 1. Install Liberty Tools Eclipse (Early Release) 2. Configure LMP to use beta runtime artifact ```xml <configuration> <runtimeArtifact> <groupId>io.openliberty.beta</groupId> <artifactId>openliberty-runtime</artifactId> <version>LATEST</version> <type>zip</type> </runtimeArtifact> </configuration> ``` 3. Do a `mvn liberty:create` (otherwise the runtime will be irrelevant and we'll go down the default feature path) 4. Open a server.xml, position the cursor with the contents of a `<feature>xyz-1.0</feature>` element and type `Ctrl+Space`. In the debugger I see NullPointerException thrown here: ``` Injector.inject(String, byte[]) line: 294 Injector.inject(ClassLoader, String, byte[]) line: 66 AccessorInjector.prepare(Class, String, String, String...) line: 57 OptimizedAccessorFactory.get(Field) line: 157 Accessor$FieldReflection<BeanT,ValueT>.optimize(JAXBContextImpl) line: 255 SingleElementNodeProperty<BeanT,ValueT>.<init>(JAXBContextImpl, RuntimeElementPropertyInfo) line: 64 NativeConstructorAccessorImpl.newInstance0(Constructor<?>, Object[]) line: not available [native method] NativeConstructorAccessorImpl.newInstance(Object[]) line: 77 DelegatingConstructorAccessorImpl.newInstance(Object[]) line: 45 Constructor<T>.newInstanceWithCaller(Object[], boolean, Class<?>) line: 499 Constructor<T>.newInstance(Object...) line: 480 PropertyFactory.create(JAXBContextImpl, RuntimePropertyInfo) line: 99 ClassBeanInfoImpl<BeanT>.<init>(JAXBContextImpl, RuntimeClassInfo) line: 150 JAXBContextImpl.getOrCreate(RuntimeClassInfo) line: 493 JAXBContextImpl.getOrCreate(RuntimeTypeInfo) line: 512 ArrayElementNodeProperty<BeanT,ListT,ItemT>(ArrayElementProperty<BeanT,ListT,ItemT>).<init>(JAXBContextImpl, RuntimeElementPropertyInfo) line: 82 ArrayElementNodeProperty<BeanT,ListT,ItemT>.<init>(JAXBContextImpl, RuntimeElementPropertyInfo) line: 32 NativeConstructorAccessorImpl.newInstance0(Constructor<?>, Object[]) line: not available [native method] NativeConstructorAccessorImpl.newInstance(Object[]) line: 77 DelegatingConstructorAccessorImpl.newInstance(Object[]) line: 45 Constructor<T>.newInstanceWithCaller(Object[], boolean, Class<?>) line: 499 Constructor<T>.newInstance(Object...) line: 480 PropertyFactory.create(JAXBContextImpl, RuntimePropertyInfo) line: 99 ClassBeanInfoImpl<BeanT>.<init>(JAXBContextImpl, RuntimeClassInfo) line: 150 JAXBContextImpl.getOrCreate(RuntimeClassInfo) line: 493 JAXBContextImpl.<init>(JAXBContextImpl$JAXBContextBuilder) line: 310 JAXBContextImpl.<init>(JAXBContextImpl$JAXBContextBuilder, JAXBContextImpl$1) line: 109 JAXBContextImpl$JAXBContextBuilder.build() line: 1142 ContextFactory.createContext(Class[], Map<String,Object>) line: 141 NativeMethodAccessorImpl.invoke0(Method, Object, Object[]) line: not available [native method] NativeMethodAccessorImpl.invoke(Object, Object[]) line: 77 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43 Method.invoke(Object, Object...) line: 568 ContextFinder.newInstance(Class[], Map, String) line: 211 ContextFinder.find(Class[], Map) line: 372 JAXBContext.newInstance(Class[], Map<String,?>) line: 574 JAXBContext.newInstance(Class...) line: 522 FeatureService.getInstalledFeaturesList(String, String, String) line: 251 FeatureService.getFeatures(String, String, int, String) line: 183 FeatureService.getFeature(String, String, String, int, String) line: 194 LibertyHoverParticipant.getHoverFeatureDescription(String, DOMDocument) line: 70 LibertyHoverParticipant.onText(IHoverRequest) line: 57 XMLHover.getTextHover(HoverRequest, Range) line: 207 XMLHover.doHover(DOMDocument, Position, SharedSettings, CancelChecker) line: 99 XMLLanguageService.doHover(DOMDocument, Position, SharedSettings, CancelChecker) line: 167 XMLTextDocumentService.lambda$hover$5(HoverParams, DOMDocument, CancelChecker) line: 260 0x000000080047e230.apply(Object, Object) line: not available ModelTextDocuments<T>.lambda$computeModelAsync$0(TextDocumentIdentifier, BiFunction, CancelChecker) line: 116 0x000000080016c590.apply(Object) line: not available CompletableFuture$UniApply<T,V>.tryFire(int) line: 646 CompletableFuture$UniApply<T,V>(CompletableFuture$Completion).exec() line: 483 CompletableFuture$UniApply<T,V>(ForkJoinTask<V>).doExec() line: 373 ForkJoinPool$WorkQueue.topLevelExec(ForkJoinTask<?>, WorkQueue) line: 1182 ForkJoinPool.scan(ForkJoinPool$WorkQueue, int, int) line: 1655 ForkJoinPool.runWorker(ForkJoinPool$WorkQueue) line: 1622 ForkJoinWorkerThread.run() line: 165 ``` which gets caught in stack > java.lang.NullPointerException: Cannot invoke "java.lang.reflect.Method.invoke(Object, Object[])" because "com.sun.xml.bind.v2.runtime.reflect.opt.Injector.defineClass" is null caught in stack: ``` XMLHover.getTextHover(HoverRequest, Range) line: 212 XMLHover.doHover(DOMDocument, Position, SharedSettings, CancelChecker) line: 99 XMLLanguageService.doHover(DOMDocument, Position, SharedSettings, CancelChecker) line: 167 XMLTextDocumentService.lambda$hover$5(HoverParams, DOMDocument, CancelChecker) line: 260 0x000000080047e230.apply(Object, Object) line: not available ModelTextDocuments<T>.lambda$computeModelAsync$0(TextDocumentIdentifier, BiFunction, CancelChecker) line: 116 0x000000080016c590.apply(Object) line: not available CompletableFuture$UniApply<T,V>.tryFire(int) line: 646 CompletableFuture$UniApply<T,V>(CompletableFuture$Completion).exec() line: 483 CompletableFuture$UniApply<T,V>(ForkJoinTask<V>).doExec() line: 373 ForkJoinPool$WorkQueue.topLevelExec(ForkJoinTask<?>, WorkQueue) line: 1182 ForkJoinPool.scan(ForkJoinPool$WorkQueue, int, int) line: 1655 ForkJoinPool.runWorker(ForkJoinPool$WorkQueue) line: 1622 ForkJoinWorkerThread.run() line: 165 ``` Note: we are running the LCLS with Java 17 in Liberty Tools Eclipse.
priority
exception when trying to use completion function from a beta runtime artifact jaxb related recreate install liberty tools eclipse early release configure lmp to use beta runtime artifact xml io openliberty beta openliberty runtime latest zip do a mvn liberty create otherwise the runtime will be irrelevant and we ll go down the default feature path open a server xml position the cursor with the contents of a xyz element and type ctrl space in the debugger i see nullpointerexception thrown here injector inject string byte line injector inject classloader string byte line accessorinjector prepare class string string string line optimizedaccessorfactory get field line accessor fieldreflection optimize jaxbcontextimpl line singleelementnodeproperty jaxbcontextimpl runtimeelementpropertyinfo line nativeconstructoraccessorimpl constructor object line not available nativeconstructoraccessorimpl newinstance object line delegatingconstructoraccessorimpl newinstance object line constructor newinstancewithcaller object boolean class line constructor newinstance object line propertyfactory create jaxbcontextimpl runtimepropertyinfo line classbeaninfoimpl jaxbcontextimpl runtimeclassinfo line jaxbcontextimpl getorcreate runtimeclassinfo line jaxbcontextimpl getorcreate runtimetypeinfo line arrayelementnodeproperty arrayelementproperty jaxbcontextimpl runtimeelementpropertyinfo line arrayelementnodeproperty jaxbcontextimpl runtimeelementpropertyinfo line nativeconstructoraccessorimpl constructor object line not available nativeconstructoraccessorimpl newinstance object line delegatingconstructoraccessorimpl newinstance object line constructor newinstancewithcaller object boolean class line constructor newinstance object line propertyfactory create jaxbcontextimpl runtimepropertyinfo line classbeaninfoimpl jaxbcontextimpl runtimeclassinfo line jaxbcontextimpl getorcreate runtimeclassinfo line jaxbcontextimpl jaxbcontextimpl jaxbcontextbuilder line jaxbcontextimpl jaxbcontextimpl jaxbcontextbuilder jaxbcontextimpl line jaxbcontextimpl jaxbcontextbuilder build line contextfactory createcontext class map line nativemethodaccessorimpl method object object line not available nativemethodaccessorimpl invoke object object line delegatingmethodaccessorimpl invoke object object line method invoke object object line contextfinder newinstance class map string line contextfinder find class map line jaxbcontext newinstance class map line jaxbcontext newinstance class line featureservice getinstalledfeatureslist string string string line featureservice getfeatures string string int string line featureservice getfeature string string string int string line libertyhoverparticipant gethoverfeaturedescription string domdocument line libertyhoverparticipant ontext ihoverrequest line xmlhover gettexthover hoverrequest range line xmlhover dohover domdocument position sharedsettings cancelchecker line xmllanguageservice dohover domdocument position sharedsettings cancelchecker line xmltextdocumentservice lambda hover hoverparams domdocument cancelchecker line apply object object line not available modeltextdocuments lambda computemodelasync textdocumentidentifier bifunction cancelchecker line apply object line not available completablefuture uniapply tryfire int line completablefuture uniapply completablefuture completion exec line completablefuture uniapply forkjointask doexec line forkjoinpool workqueue toplevelexec forkjointask workqueue line forkjoinpool scan forkjoinpool workqueue int int line forkjoinpool runworker forkjoinpool workqueue line forkjoinworkerthread run line which gets caught in stack java lang nullpointerexception cannot invoke java lang reflect method invoke object object because com sun xml bind runtime reflect opt injector defineclass is null caught in stack xmlhover gettexthover hoverrequest range line xmlhover dohover domdocument position sharedsettings cancelchecker line xmllanguageservice dohover domdocument position sharedsettings cancelchecker line xmltextdocumentservice lambda hover hoverparams domdocument cancelchecker line apply object object line not available modeltextdocuments lambda computemodelasync textdocumentidentifier bifunction cancelchecker line apply object line not available completablefuture uniapply tryfire int line completablefuture uniapply completablefuture completion exec line completablefuture uniapply forkjointask doexec line forkjoinpool workqueue toplevelexec forkjointask workqueue line forkjoinpool scan forkjoinpool workqueue int int line forkjoinpool runworker forkjoinpool workqueue line forkjoinworkerthread run line note we are running the lcls with java in liberty tools eclipse
1
79,593
3,537,462,198
IssuesEvent
2016-01-18 01:11:43
keroserene/go-webrtc
https://api.github.com/repos/keroserene/go-webrtc
closed
Implement all PeerConnection callbacks
priority: high
See: https://w3c.github.io/webrtc-pc/#idl-def-RTCPeerConnection Progress: - [x] OnNegotiationNeeded - [x] OnIceCandidate - [x] OnIceCandidateError - [x] OnSignalingStateChange - [x] ~~OnIceComplete~~ - [x] OnIceGatheringStateChange - [x] OnIceConnectionStateChange - [ ] ~~OnConnectionStateChange~~ - [x] OnDataChannel
1.0
Implement all PeerConnection callbacks - See: https://w3c.github.io/webrtc-pc/#idl-def-RTCPeerConnection Progress: - [x] OnNegotiationNeeded - [x] OnIceCandidate - [x] OnIceCandidateError - [x] OnSignalingStateChange - [x] ~~OnIceComplete~~ - [x] OnIceGatheringStateChange - [x] OnIceConnectionStateChange - [ ] ~~OnConnectionStateChange~~ - [x] OnDataChannel
priority
implement all peerconnection callbacks see progress onnegotiationneeded onicecandidate onicecandidateerror onsignalingstatechange onicecomplete onicegatheringstatechange oniceconnectionstatechange onconnectionstatechange ondatachannel
1
535,179
15,683,722,592
IssuesEvent
2021-03-25 09:08:38
RasaHQ/rasa
https://api.github.com/repos/RasaHQ/rasa
closed
Rasa x not able to load all sentences in NLU inbox
area:rasa-x/ui :sparkles: priority:high type:bug :bug:
<!-- THIS INFORMATION IS MANDATORY - YOUR ISSUE WILL BE CLOSED IF IT IS MISSING. If you don't know your Rasa version, use `rasa --version`. Please format any code or console output with three ticks ``` above and below. If you are asking a usage question (e.g. "How do I do xyz") please post your question on https://forum.rasa.com instead --> **Rasa version**: **Rasa SDK version** : 2.1.2 **Rasa X version** : 0.34 **Python version**: 3.7.6 **Operating system** (windows, osx, ...): Ubuntu **Issue**: While using Rasa X the NLU inbox is filled with 122 sentences but when I scroll to view more so that I can add in Nlu file it is not able to load. I am using Azure Virtual Machine with ubuntu. I have added screenshots for the same. I waited like half an hour and still, it is not able to load data. Since, after testing we needed Nlu data to get stored in nlu.yml. So, please look at this bug as soon as possible ![image](https://user-images.githubusercontent.com/57827233/106752669-bb41ca80-6650-11eb-9586-093bc7d051cc.png) **Error (including full traceback)**: ``` No errors ``` **Command or request that led to error**: ``` rasa x --rasa-x-port 8002 ``` **Content of configuration file (config.yml)** (if relevant): ```yml ``` **Content of domain file (domain.yml)** (if relevant): ```yml ```
1.0
Rasa x not able to load all sentences in NLU inbox - <!-- THIS INFORMATION IS MANDATORY - YOUR ISSUE WILL BE CLOSED IF IT IS MISSING. If you don't know your Rasa version, use `rasa --version`. Please format any code or console output with three ticks ``` above and below. If you are asking a usage question (e.g. "How do I do xyz") please post your question on https://forum.rasa.com instead --> **Rasa version**: **Rasa SDK version** : 2.1.2 **Rasa X version** : 0.34 **Python version**: 3.7.6 **Operating system** (windows, osx, ...): Ubuntu **Issue**: While using Rasa X the NLU inbox is filled with 122 sentences but when I scroll to view more so that I can add in Nlu file it is not able to load. I am using Azure Virtual Machine with ubuntu. I have added screenshots for the same. I waited like half an hour and still, it is not able to load data. Since, after testing we needed Nlu data to get stored in nlu.yml. So, please look at this bug as soon as possible ![image](https://user-images.githubusercontent.com/57827233/106752669-bb41ca80-6650-11eb-9586-093bc7d051cc.png) **Error (including full traceback)**: ``` No errors ``` **Command or request that led to error**: ``` rasa x --rasa-x-port 8002 ``` **Content of configuration file (config.yml)** (if relevant): ```yml ``` **Content of domain file (domain.yml)** (if relevant): ```yml ```
priority
rasa x not able to load all sentences in nlu inbox this information is mandatory your issue will be closed if it is missing if you don t know your rasa version use rasa version please format any code or console output with three ticks above and below if you are asking a usage question e g how do i do xyz please post your question on instead rasa version rasa sdk version rasa x version python version operating system windows osx ubuntu issue while using rasa x the nlu inbox is filled with sentences but when i scroll to view more so that i can add in nlu file it is not able to load i am using azure virtual machine with ubuntu i have added screenshots for the same i waited like half an hour and still it is not able to load data since after testing we needed nlu data to get stored in nlu yml so please look at this bug as soon as possible error including full traceback no errors command or request that led to error rasa x rasa x port content of configuration file config yml if relevant yml content of domain file domain yml if relevant yml
1
308,188
9,435,916,997
IssuesEvent
2019-04-13 01:20:39
Seneca-CDOT/simengine
https://api.github.com/repos/Seneca-CDOT/simengine
closed
Add Temperature Sensor Templates
enginecore enhancement high-priority sensor_def
Temperature Sensors Include: - [x] CPU Temperature - [x] PSU Temperature - [x] PSU Inlet (?) - [x] Memory - [x] Ambient - [x] Systemboard - [ ] BBU (?) - [x] RAID Controller Sensor Sdr listings: ``` Ambient | 26.000 | degrees C | ok | na | 1.000 | 6.000 | 37.000 | 42.000 | na Systemboard | 50.000 | degrees C | ok | na | na | na | 75.000 | 80.000 | na VR CPU1 | 46.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR MEM AB | 35.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR MEM CD | 47.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR CPU2 | 43.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR MEM EF | 42.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR MEM GH | 36.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na CPU1 | 42.000 | degrees C | ok | na | na | na | 98.000 | 99.000 | na CPU2 | 45.000 | degrees C | ok | na | na | na | 98.000 | 99.000 | na MEM A | 35.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM B | 34.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM C | 41.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM D | 40.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM E | 39.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM F | 39.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM G | 33.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM H | 34.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na PSU1 Inlet | 41.000 | degrees C | ok | na | na | na | 57.000 | 61.000 | na PSU2 Inlet | 41.000 | degrees C | ok | na | na | na | 57.000 | 61.000 | na PSU1 | 66.000 | degrees C | ok | na | na | na | 102.000 | 107.000 | na PSU2 | 63.000 | degrees C | ok | na | na | na | 102.000 | 107.000 | na BBU | 37.000 | degrees C | ok | na | na | na | 50.000 | 55.000 | na RAID Controller | 93.000 | degrees C | ok | na | na | na | 110.000 | 115.000 | na ```
1.0
Add Temperature Sensor Templates - Temperature Sensors Include: - [x] CPU Temperature - [x] PSU Temperature - [x] PSU Inlet (?) - [x] Memory - [x] Ambient - [x] Systemboard - [ ] BBU (?) - [x] RAID Controller Sensor Sdr listings: ``` Ambient | 26.000 | degrees C | ok | na | 1.000 | 6.000 | 37.000 | 42.000 | na Systemboard | 50.000 | degrees C | ok | na | na | na | 75.000 | 80.000 | na VR CPU1 | 46.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR MEM AB | 35.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR MEM CD | 47.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR CPU2 | 43.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR MEM EF | 42.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na VR MEM GH | 36.000 | degrees C | ok | na | na | na | 120.000 | 125.000 | na CPU1 | 42.000 | degrees C | ok | na | na | na | 98.000 | 99.000 | na CPU2 | 45.000 | degrees C | ok | na | na | na | 98.000 | 99.000 | na MEM A | 35.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM B | 34.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM C | 41.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM D | 40.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM E | 39.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM F | 39.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM G | 33.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na MEM H | 34.000 | degrees C | ok | na | na | na | 78.000 | 82.000 | na PSU1 Inlet | 41.000 | degrees C | ok | na | na | na | 57.000 | 61.000 | na PSU2 Inlet | 41.000 | degrees C | ok | na | na | na | 57.000 | 61.000 | na PSU1 | 66.000 | degrees C | ok | na | na | na | 102.000 | 107.000 | na PSU2 | 63.000 | degrees C | ok | na | na | na | 102.000 | 107.000 | na BBU | 37.000 | degrees C | ok | na | na | na | 50.000 | 55.000 | na RAID Controller | 93.000 | degrees C | ok | na | na | na | 110.000 | 115.000 | na ```
priority
add temperature sensor templates temperature sensors include cpu temperature psu temperature psu inlet memory ambient systemboard bbu raid controller sensor sdr listings ambient degrees c ok na na systemboard degrees c ok na na na na vr degrees c ok na na na na vr mem ab degrees c ok na na na na vr mem cd degrees c ok na na na na vr degrees c ok na na na na vr mem ef degrees c ok na na na na vr mem gh degrees c ok na na na na degrees c ok na na na na degrees c ok na na na na mem a degrees c ok na na na na mem b degrees c ok na na na na mem c degrees c ok na na na na mem d degrees c ok na na na na mem e degrees c ok na na na na mem f degrees c ok na na na na mem g degrees c ok na na na na mem h degrees c ok na na na na inlet degrees c ok na na na na inlet degrees c ok na na na na degrees c ok na na na na degrees c ok na na na na bbu degrees c ok na na na na raid controller degrees c ok na na na na
1
479,616
13,803,877,987
IssuesEvent
2020-10-11 05:56:19
AY2021S1-CS2113T-F12-3/tp
https://api.github.com/repos/AY2021S1-CS2113T-F12-3/tp
closed
Implement Commands Class
priority.High type.Task
This class will be used to serve as a parent class for all future commands, with the necessary constructors and methods implemented so that it serves as a basis for specific command classes that extend from it.
1.0
Implement Commands Class - This class will be used to serve as a parent class for all future commands, with the necessary constructors and methods implemented so that it serves as a basis for specific command classes that extend from it.
priority
implement commands class this class will be used to serve as a parent class for all future commands with the necessary constructors and methods implemented so that it serves as a basis for specific command classes that extend from it
1
522,391
15,158,987,954
IssuesEvent
2021-02-12 02:48:21
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[9.3] New Tag Icons
Category: Art Priority: High Squad: Saguaro Type: Task
We need some new tag icons, specifically one for different 'Aquatic Carcasses'-- I'm thinking a clearly dead fish will do for this for now, it can be used on all sizes most likely since that's how the other carcasses work for terrestrial animals. We also need a Clothes tag icon. - [x] Aquatic Carcass Tag Icon - [x] Clothes Tag Icon ![image.png](https://images.zenhubusercontent.com/5ee003d5cfaa7fb4ca69f828/0146c21f-028a-4dbd-b51c-e7862f4f9dc5)
1.0
[9.3] New Tag Icons - We need some new tag icons, specifically one for different 'Aquatic Carcasses'-- I'm thinking a clearly dead fish will do for this for now, it can be used on all sizes most likely since that's how the other carcasses work for terrestrial animals. We also need a Clothes tag icon. - [x] Aquatic Carcass Tag Icon - [x] Clothes Tag Icon ![image.png](https://images.zenhubusercontent.com/5ee003d5cfaa7fb4ca69f828/0146c21f-028a-4dbd-b51c-e7862f4f9dc5)
priority
new tag icons we need some new tag icons specifically one for different aquatic carcasses i m thinking a clearly dead fish will do for this for now it can be used on all sizes most likely since that s how the other carcasses work for terrestrial animals we also need a clothes tag icon aquatic carcass tag icon clothes tag icon
1
506,312
14,662,264,719
IssuesEvent
2020-12-29 06:44:17
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.apple.com - design is broken
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox Mobile 86.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:86.0) Gecko/86.0 Firefox/86.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64548 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.apple.com/my/shop/buy-iphone/iphone-12 **Browser / Version**: Firefox Mobile 86.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Design is broken **Description**: Items are misaligned **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/cfcb4e4f-4354-4244-99d0-2ac1e99a27fb.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201225095506</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/12/49dd451b-a5bc-472c-a0ce-d91fb56fdde2) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.apple.com - design is broken - <!-- @browser: Firefox Mobile 86.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:86.0) Gecko/86.0 Firefox/86.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64548 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.apple.com/my/shop/buy-iphone/iphone-12 **Browser / Version**: Firefox Mobile 86.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Design is broken **Description**: Items are misaligned **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/cfcb4e4f-4354-4244-99d0-2ac1e99a27fb.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201225095506</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/12/49dd451b-a5bc-472c-a0ce-d91fb56fdde2) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
design is broken url browser version firefox mobile operating system android tested another browser no problem type design is broken description items are misaligned steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
445,536
12,832,295,938
IssuesEvent
2020-07-07 07:24:40
comunica/comunica-packager
https://api.github.com/repos/comunica/comunica-packager
closed
Improve lay-out
priority:high
There seem to be some issues with the buttons on smaller screens: ![Uploading Screen Shot 2020-06-22 at 15.04.36.png…]() I would suggest making the button text simpler, i.e.: `Import`, `Export`, and `Reset`. While you're at it, I would also make the logo smaller, and add the text `Packager` next to it or below, in the same or similar font. Because we should reserve as much space as possible for the form itself.
1.0
Improve lay-out - There seem to be some issues with the buttons on smaller screens: ![Uploading Screen Shot 2020-06-22 at 15.04.36.png…]() I would suggest making the button text simpler, i.e.: `Import`, `Export`, and `Reset`. While you're at it, I would also make the logo smaller, and add the text `Packager` next to it or below, in the same or similar font. Because we should reserve as much space as possible for the form itself.
priority
improve lay out there seem to be some issues with the buttons on smaller screens i would suggest making the button text simpler i e import export and reset while you re at it i would also make the logo smaller and add the text packager next to it or below in the same or similar font because we should reserve as much space as possible for the form itself
1
467,055
13,439,964,433
IssuesEvent
2020-09-07 23:04:21
ctm/mb2-doc
https://api.github.com/repos/ctm/mb2-doc
closed
couldn't identify bustor
bug easy high priority
Twice in last night's tournaments mb2 wasn't able to identify who bust whom. This was due to the game being Korean, there being no high and mb2 not looking at the low. Obviously, what mb2 is doing is not right for Korean and can and will be changed. I added some more info to my bounty corner cases issue (#418).
1.0
couldn't identify bustor - Twice in last night's tournaments mb2 wasn't able to identify who bust whom. This was due to the game being Korean, there being no high and mb2 not looking at the low. Obviously, what mb2 is doing is not right for Korean and can and will be changed. I added some more info to my bounty corner cases issue (#418).
priority
couldn t identify bustor twice in last night s tournaments wasn t able to identify who bust whom this was due to the game being korean there being no high and not looking at the low obviously what is doing is not right for korean and can and will be changed i added some more info to my bounty corner cases issue
1