Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
443,322
12,792,548,904
IssuesEvent
2020-07-02 01:40:39
lardemua/atom
https://api.github.com/repos/lardemua/atom
closed
Merge general hand eye with calibrate
High Priority enhancement help wanted
Should do this to advance in the general approach. Defined with @eupedrosa that the best way is to try to put in calibrate the stuff from hand eye, i.e. to use the calibrate as base.
1.0
Merge general hand eye with calibrate - Should do this to advance in the general approach. Defined with @eupedrosa that the best way is to try to put in calibrate the stuff from hand eye, i.e. to use the calibrate as base.
priority
merge general hand eye with calibrate should do this to advance in the general approach defined with eupedrosa that the best way is to try to put in calibrate the stuff from hand eye i e to use the calibrate as base
1
141,381
5,435,457,609
IssuesEvent
2017-03-05 17:09:33
Templarian/MaterialDesign
https://api.github.com/repos/Templarian/MaterialDesign
closed
Exponent and Root
Alias Icon High Priority Icon Request
I'm sure this was part of a request at some point but I can't seem to find these icons. An `x²` and `√x` would be really useful. Could also use a decimal point (the dot talked about in #1709 would work), and a degree symbol like `x°`.
1.0
Exponent and Root - I'm sure this was part of a request at some point but I can't seem to find these icons. An `x²` and `√x` would be really useful. Could also use a decimal point (the dot talked about in #1709 would work), and a degree symbol like `x°`.
priority
exponent and root i m sure this was part of a request at some point but i can t seem to find these icons an x² and √x would be really useful could also use a decimal point the dot talked about in would work and a degree symbol like x°
1
131,858
5,166,426,990
IssuesEvent
2017-01-17 16:12:26
snaiperskaya96/test-import-repo
https://api.github.com/repos/snaiperskaya96/test-import-repo
opened
Delete old 'auto invoice' module
Accepted Enhancement High Priority
https://trello.com/c/XWeij4td/487-delete-old-auto-invoice-module This must be done after all of our Brightpearl accounts are fully upgraded to version 4.90.
1.0
Delete old 'auto invoice' module - https://trello.com/c/XWeij4td/487-delete-old-auto-invoice-module This must be done after all of our Brightpearl accounts are fully upgraded to version 4.90.
priority
delete old auto invoice module this must be done after all of our brightpearl accounts are fully upgraded to version
1
737,245
25,507,848,224
IssuesEvent
2022-11-28 10:48:37
opensquare-network/bounties
https://api.github.com/repos/opensquare-network/bounties
closed
Award extrinsic keep broadcasting
bug priority:high
1. The award button should be disabled if work not submitted 2. The broadcasting toast keep showing after clicking award button when no work not submitted ![SbpTgxSDyX](https://user-images.githubusercontent.com/2264908/198668094-519bfdde-55de-4332-abea-0d368e9b393f.jpg)
1.0
Award extrinsic keep broadcasting - 1. The award button should be disabled if work not submitted 2. The broadcasting toast keep showing after clicking award button when no work not submitted ![SbpTgxSDyX](https://user-images.githubusercontent.com/2264908/198668094-519bfdde-55de-4332-abea-0d368e9b393f.jpg)
priority
award extrinsic keep broadcasting the award button should be disabled if work not submitted the broadcasting toast keep showing after clicking award button when no work not submitted
1
32,883
2,760,832,254
IssuesEvent
2015-04-28 14:24:24
DLR-SC/tixi
https://api.github.com/repos/DLR-SC/tixi
closed
Function for specifying output stream of error and warning messages
auto-migrated Milestone-Release2.1.2 Priority-High Type-Enhancement
``` Currently, all error and warnings are printed to stdout. This is sometimes impractical. E.g. we would like to get TiXI Message inside TiGLViewer. Therefore we have to pipe the messages into the Logging framework of TiGL. Adding a new logging framework to Tixi is overkill, but we could provide a function that allows setting a different output stream. The function could be tixiSetStandardOut(FILE * outstream) ``` Original issue reported on code.google.com by `martinsi...@gmail.com` on 14 Nov 2013 at 8:40
1.0
Function for specifying output stream of error and warning messages - ``` Currently, all error and warnings are printed to stdout. This is sometimes impractical. E.g. we would like to get TiXI Message inside TiGLViewer. Therefore we have to pipe the messages into the Logging framework of TiGL. Adding a new logging framework to Tixi is overkill, but we could provide a function that allows setting a different output stream. The function could be tixiSetStandardOut(FILE * outstream) ``` Original issue reported on code.google.com by `martinsi...@gmail.com` on 14 Nov 2013 at 8:40
priority
function for specifying output stream of error and warning messages currently all error and warnings are printed to stdout this is sometimes impractical e g we would like to get tixi message inside tiglviewer therefore we have to pipe the messages into the logging framework of tigl adding a new logging framework to tixi is overkill but we could provide a function that allows setting a different output stream the function could be tixisetstandardout file outstream original issue reported on code google com by martinsi gmail com on nov at
1
780,204
27,384,570,300
IssuesEvent
2023-02-28 12:26:12
Azure/mec-app-solution-accelerator
https://api.github.com/repos/Azure/mec-app-solution-accelerator
closed
[doc] Create an .MD explaining how to provision a new camera
issue P1 (High priority)
Add that as an .MD with a link from README.md explaining how to provision a new camera.
1.0
[doc] Create an .MD explaining how to provision a new camera - Add that as an .MD with a link from README.md explaining how to provision a new camera.
priority
create an md explaining how to provision a new camera add that as an md with a link from readme md explaining how to provision a new camera
1
576,783
17,094,640,473
IssuesEvent
2021-07-08 23:14:36
CHOMPStation2/CHOMPStation2
https://api.github.com/repos/CHOMPStation2/CHOMPStation2
closed
Mapping support needed for the change to conveyor belts
High Priority Map Edit
#### Brief description of the issue Due to https://github.com/CHOMPStation2/CHOMPStation2/pull/2441, conveyor belts set to diagonal directions are pointing in the wrong directions, commonly problematic at the mining base that relies on it to process materials. #### What you expected to happen For conveyor belts to work before the PR merge. #### What actually happened Conveyor belts are pointing in the wrong direction if they were set to diagonals. #### Steps to reproduce - Step 1 - Go to Mining on Sif - Step 2 - Check conveyor belts - Step 3 - Laugh/Cry #### Code Revision Server revision: B:-Using TGS- D:-Using TGS- Commit: 7438d0f486dcc60bafc39dd0e8266db5b55194fd TGS version: 4.11.1 DMAPI version: 5.3.0 #### Anything else you may wish to add: - This generally affects all conveyor belts mapped in, just blocks mining from their job.
1.0
Mapping support needed for the change to conveyor belts - #### Brief description of the issue Due to https://github.com/CHOMPStation2/CHOMPStation2/pull/2441, conveyor belts set to diagonal directions are pointing in the wrong directions, commonly problematic at the mining base that relies on it to process materials. #### What you expected to happen For conveyor belts to work before the PR merge. #### What actually happened Conveyor belts are pointing in the wrong direction if they were set to diagonals. #### Steps to reproduce - Step 1 - Go to Mining on Sif - Step 2 - Check conveyor belts - Step 3 - Laugh/Cry #### Code Revision Server revision: B:-Using TGS- D:-Using TGS- Commit: 7438d0f486dcc60bafc39dd0e8266db5b55194fd TGS version: 4.11.1 DMAPI version: 5.3.0 #### Anything else you may wish to add: - This generally affects all conveyor belts mapped in, just blocks mining from their job.
priority
mapping support needed for the change to conveyor belts brief description of the issue due to conveyor belts set to diagonal directions are pointing in the wrong directions commonly problematic at the mining base that relies on it to process materials what you expected to happen for conveyor belts to work before the pr merge what actually happened conveyor belts are pointing in the wrong direction if they were set to diagonals steps to reproduce step go to mining on sif step check conveyor belts step laugh cry code revision server revision b using tgs d using tgs commit tgs version dmapi version anything else you may wish to add this generally affects all conveyor belts mapped in just blocks mining from their job
1
614,495
19,184,245,408
IssuesEvent
2021-12-04 23:26:27
aaronparker/evergreen
https://api.github.com/repos/aaronparker/evergreen
closed
[Bug]: Microsoft.NET Download URL's changed?
bug priority:high
### What happened? Microsoft.NET 'LTS' Channel downloads appear to have moved from https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/.... 'Current' Channel still seems to be valid for downloads. If I go to the Microsoft .NET Download site - the new url appears to be: https://dotnet.microsoft.com/download/dotnet/thank-you/runtime-desktop-6.0.0-windows-x64-installer ### Version 2111.448 ### What PowerShell edition/s are you running Evergreen on? Windows PowerShell ### Which operating system/s are you running Evergreen on? Windows Server 2016+ ### Have you reviewed the documentation? - [X] Troubleshooting at: https://stealthpuppy.com/evergreen/troubleshoot/ - [X] Known issues at: https://stealthpuppy.com/evergreen/issues/ ### Verbose output ```shell Get-EvergreenApp -Name Microsoft.NET -Verbose VERBOSE: Get-EvergreenApp: Function exists: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1. VERBOSE: Get-EvergreenApp: Dot sourcing: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1. VERBOSE: Get-FunctionResource: read application resource strings from [C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Man ifests\Microsoft.NET.json] VERBOSE: Get-EvergreenApp: Calling: Get-Microsoft.NET. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UserAgent: Mozilla/5.0 (Windows NT; Windows NT 10.0; en-US) AppleWebKit /534.6 (KHTML, like Gecko) Chrome/7.0.500.0 Safari/534.6]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Method: Default]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [ErrorAction: Continue]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UseBasicParsing: True]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Uri: https://dotnetcli.blob.core.windows.net/dotnet/Runtime/Current/lat est.version]. VERBOSE: GET https://dotnetcli.blob.core.windows.net/dotnet/Runtime/Current/latest.version with 0-byte payload VERBOSE: received 6-byte response of content type text/plain VERBOSE: Invoke-WebRequestWrapper: Response: [200]. VERBOSE: Invoke-WebRequestWrapper: Content type: [text/plain]. VERBOSE: Invoke-WebRequestWrapper: Returning content of length: [6]. VERBOSE: Get-Microsoft.NET: found version: 5.0.12. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UserAgent: Mozilla/5.0 (Windows NT; Windows NT 10.0; en-US) AppleWebKit /534.6 (KHTML, like Gecko) Chrome/7.0.500.0 Safari/534.6]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Method: Default]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [ErrorAction: Continue]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UseBasicParsing: True]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Uri: https://dotnetcli.blob.core.windows.net/dotnet/Runtime/LTS/latest. version]. VERBOSE: GET https://dotnetcli.blob.core.windows.net/dotnet/Runtime/LTS/latest.version with 0-byte payload VERBOSE: received 5-byte response of content type text/plain VERBOSE: Invoke-WebRequestWrapper: Response: [200]. VERBOSE: Invoke-WebRequestWrapper: Content type: [text/plain]. VERBOSE: Invoke-WebRequestWrapper: Returning content of length: [5]. VERBOSE: Get-Microsoft.NET: found version: 6.0.0. VERBOSE: Get-EvergreenApp: Output result from: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1. Version Architecture Channel URI ------- ------------ ------- --- 6.0.0 x64 LTS https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/windowsdesktop-runtime-6.0.0-win-x64.exe 6.0.0 x86 LTS https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/windowsdesktop-runtime-6.0.0-win-x86.exe 5.0.12 x64 Current https://dotnetcli.blob.core.windows.net/dotnet/WindowsDesktop/5.0.12/windowsdesktop-runtime-5.0.12-win-... 5.0.12 x86 Current https://dotnetcli.blob.core.windows.net/dotnet/WindowsDesktop/5.0.12/windowsdesktop-runtime-5.0.12-win-... ```
1.0
[Bug]: Microsoft.NET Download URL's changed? - ### What happened? Microsoft.NET 'LTS' Channel downloads appear to have moved from https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/.... 'Current' Channel still seems to be valid for downloads. If I go to the Microsoft .NET Download site - the new url appears to be: https://dotnet.microsoft.com/download/dotnet/thank-you/runtime-desktop-6.0.0-windows-x64-installer ### Version 2111.448 ### What PowerShell edition/s are you running Evergreen on? Windows PowerShell ### Which operating system/s are you running Evergreen on? Windows Server 2016+ ### Have you reviewed the documentation? - [X] Troubleshooting at: https://stealthpuppy.com/evergreen/troubleshoot/ - [X] Known issues at: https://stealthpuppy.com/evergreen/issues/ ### Verbose output ```shell Get-EvergreenApp -Name Microsoft.NET -Verbose VERBOSE: Get-EvergreenApp: Function exists: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1. VERBOSE: Get-EvergreenApp: Dot sourcing: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1. VERBOSE: Get-FunctionResource: read application resource strings from [C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Man ifests\Microsoft.NET.json] VERBOSE: Get-EvergreenApp: Calling: Get-Microsoft.NET. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UserAgent: Mozilla/5.0 (Windows NT; Windows NT 10.0; en-US) AppleWebKit /534.6 (KHTML, like Gecko) Chrome/7.0.500.0 Safari/534.6]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Method: Default]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [ErrorAction: Continue]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UseBasicParsing: True]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Uri: https://dotnetcli.blob.core.windows.net/dotnet/Runtime/Current/lat est.version]. VERBOSE: GET https://dotnetcli.blob.core.windows.net/dotnet/Runtime/Current/latest.version with 0-byte payload VERBOSE: received 6-byte response of content type text/plain VERBOSE: Invoke-WebRequestWrapper: Response: [200]. VERBOSE: Invoke-WebRequestWrapper: Content type: [text/plain]. VERBOSE: Invoke-WebRequestWrapper: Returning content of length: [6]. VERBOSE: Get-Microsoft.NET: found version: 5.0.12. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UserAgent: Mozilla/5.0 (Windows NT; Windows NT 10.0; en-US) AppleWebKit /534.6 (KHTML, like Gecko) Chrome/7.0.500.0 Safari/534.6]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Method: Default]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [ErrorAction: Continue]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UseBasicParsing: True]. VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Uri: https://dotnetcli.blob.core.windows.net/dotnet/Runtime/LTS/latest. version]. VERBOSE: GET https://dotnetcli.blob.core.windows.net/dotnet/Runtime/LTS/latest.version with 0-byte payload VERBOSE: received 5-byte response of content type text/plain VERBOSE: Invoke-WebRequestWrapper: Response: [200]. VERBOSE: Invoke-WebRequestWrapper: Content type: [text/plain]. VERBOSE: Invoke-WebRequestWrapper: Returning content of length: [5]. VERBOSE: Get-Microsoft.NET: found version: 6.0.0. VERBOSE: Get-EvergreenApp: Output result from: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1. Version Architecture Channel URI ------- ------------ ------- --- 6.0.0 x64 LTS https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/windowsdesktop-runtime-6.0.0-win-x64.exe 6.0.0 x86 LTS https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/windowsdesktop-runtime-6.0.0-win-x86.exe 5.0.12 x64 Current https://dotnetcli.blob.core.windows.net/dotnet/WindowsDesktop/5.0.12/windowsdesktop-runtime-5.0.12-win-... 5.0.12 x86 Current https://dotnetcli.blob.core.windows.net/dotnet/WindowsDesktop/5.0.12/windowsdesktop-runtime-5.0.12-win-... ```
priority
microsoft net download url s changed what happened microsoft net lts channel downloads appear to have moved from current channel still seems to be valid for downloads if i go to the microsoft net download site the new url appears to be version what powershell edition s are you running evergreen on windows powershell which operating system s are you running evergreen on windows server have you reviewed the documentation troubleshooting at known issues at verbose output shell get evergreenapp name microsoft net verbose verbose get evergreenapp function exists c program files windowspowershell modules evergreen apps get microsoft net verbose get evergreenapp dot sourcing c program files windowspowershell modules evergreen apps get microsoft net verbose get functionresource read application resource strings from c program files windowspowershell modules evergreen man ifests microsoft net json verbose get evergreenapp calling get microsoft net verbose invoke webrequestwrapper invoke webrequest parameter useragent mozilla windows nt windows nt en us applewebkit khtml like gecko chrome safari verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter uri est version verbose get with byte payload verbose received byte response of content type text plain verbose invoke webrequestwrapper response verbose invoke webrequestwrapper content type verbose invoke webrequestwrapper returning content of length verbose get microsoft net found version verbose invoke webrequestwrapper invoke webrequest parameter useragent mozilla windows nt windows nt en us applewebkit khtml like gecko chrome safari verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter uri version verbose get with byte payload verbose received byte response of content type text plain verbose invoke webrequestwrapper response verbose invoke webrequestwrapper content type verbose invoke webrequestwrapper returning content of length verbose get microsoft net found version verbose get evergreenapp output result from c program files windowspowershell modules evergreen apps get microsoft net version architecture channel uri lts lts current current
1
282,606
8,708,460,740
IssuesEvent
2018-12-06 10:58:09
pablotabares/decide
https://api.github.com/repos/pablotabares/decide
closed
Add poll creation functionality
bot enhancement priority: high
Telegram bot must provide commands to create a poll, including questions and answers.
1.0
Add poll creation functionality - Telegram bot must provide commands to create a poll, including questions and answers.
priority
add poll creation functionality telegram bot must provide commands to create a poll including questions and answers
1
635,471
20,403,353,866
IssuesEvent
2022-02-23 00:26:01
CoEDL/nyingarn-workspace
https://api.github.com/repos/CoEDL/nyingarn-workspace
closed
Users need to be able to download specific files and delete specific files
enhancement priority-high
Implement ability to see resource files, download them (e.g. digivol csv) and delete specific files.
1.0
Users need to be able to download specific files and delete specific files - Implement ability to see resource files, download them (e.g. digivol csv) and delete specific files.
priority
users need to be able to download specific files and delete specific files implement ability to see resource files download them e g digivol csv and delete specific files
1
373,325
11,042,216,416
IssuesEvent
2019-12-09 08:40:39
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Issue in inferRecordFieldType method
Area/Language Component/Compiler Points/1 Priority/High Type/Improvement
**Description:** In the Types class, we can infer the type of a record field using `inferRecordFieldType` method. `anydata` is not considered in this method, `any` will be returned in such cases.
1.0
Issue in inferRecordFieldType method - **Description:** In the Types class, we can infer the type of a record field using `inferRecordFieldType` method. `anydata` is not considered in this method, `any` will be returned in such cases.
priority
issue in inferrecordfieldtype method description in the types class we can infer the type of a record field using inferrecordfieldtype method anydata is not considered in this method any will be returned in such cases
1
441,258
12,710,053,444
IssuesEvent
2020-06-23 13:20:29
RonAsis/Wsep202
https://api.github.com/repos/RonAsis/Wsep202
opened
bug- not responsive enough - can't appoint all users as owner
High priority bug
and doesn't tell why- there is no error message
1.0
bug- not responsive enough - can't appoint all users as owner - and doesn't tell why- there is no error message
priority
bug not responsive enough can t appoint all users as owner and doesn t tell why there is no error message
1
243,593
7,859,496,551
IssuesEvent
2018-06-21 16:46:50
minio/minio-go
https://api.github.com/repos/minio/minio-go
closed
Error while running Azure tests on Mint
priority: high
When I was running Azure gateway tests on Mint, I got the following error: ``` { "args": { "bucketName": "minio-go-test-mj6c6n45dw0bpa4e", "objectName": "test-object", "opts": "", "size": -1 }, "duration": 289, "function": "PutObject(bucketName, objectName, reader, size, opts)", "message": "Expected content-language 'en-US' doesn't match with StatObject return value", "name": "minio-go: testPutObjectWithContentLanguage", "status": "FAIL" } ```
1.0
Error while running Azure tests on Mint - When I was running Azure gateway tests on Mint, I got the following error: ``` { "args": { "bucketName": "minio-go-test-mj6c6n45dw0bpa4e", "objectName": "test-object", "opts": "", "size": -1 }, "duration": 289, "function": "PutObject(bucketName, objectName, reader, size, opts)", "message": "Expected content-language 'en-US' doesn't match with StatObject return value", "name": "minio-go: testPutObjectWithContentLanguage", "status": "FAIL" } ```
priority
error while running azure tests on mint when i was running azure gateway tests on mint i got the following error args bucketname minio go test objectname test object opts size duration function putobject bucketname objectname reader size opts message expected content language en us doesn t match with statobject return value name minio go testputobjectwithcontentlanguage status fail
1
797,763
28,154,687,278
IssuesEvent
2023-04-03 06:18:02
AY2223S2-CS2113-T15-4/tp
https://api.github.com/repos/AY2223S2-CS2113-T15-4/tp
closed
[PE-D][Tester A] There is some logging that is disrupting the user experience when using commands
type.Bug priority.High severity.High
![image.png](https://raw.githubusercontent.com/SSzeWen/ped/main/files/f1fd6e78-4c94-451c-80c2-328c407dec64.png) <!--session: 1680252479707-92304da9-1923-4959-b973-1d1a8a61eabd--> <!--Version: Web v3.4.7--> ------------- Labels: `type.FunctionalityBug` `severity.Medium` original: SSzeWen/ped#1
1.0
[PE-D][Tester A] There is some logging that is disrupting the user experience when using commands - ![image.png](https://raw.githubusercontent.com/SSzeWen/ped/main/files/f1fd6e78-4c94-451c-80c2-328c407dec64.png) <!--session: 1680252479707-92304da9-1923-4959-b973-1d1a8a61eabd--> <!--Version: Web v3.4.7--> ------------- Labels: `type.FunctionalityBug` `severity.Medium` original: SSzeWen/ped#1
priority
there is some logging that is disrupting the user experience when using commands labels type functionalitybug severity medium original sszewen ped
1
581,808
17,332,278,413
IssuesEvent
2021-07-28 05:12:54
sacloud/terraform-provider-sakuracloud
https://api.github.com/repos/sacloud/terraform-provider-sakuracloud
closed
外部ツール経由でプラン変更された場合のID変更の追跡
area/resources priority/high v2
### 概要 さくらのクラウドではプラン変更時にリソースのIDが変更される。 対象リソース: - サーバ - ルータ - ELB Terraformからプラン変更を行なった場合、IDが変更されることへの対応が実装されているが、 外部ツール(AutoScalerなど)からプラン変更が行われた場合にリソースが追跡できなくなる。 これは、各リソースのReadでIDを元にさくらのクラウドAPIを用いてリソースの情報を参照しているため。 (IDで検索し404が返ってきた場合はリソースが削除されたとみなす) 追跡できなくなってもimportすることで復帰させることが出来るが運用的に煩雑。 このため、何らかのルールにしたがって変更前のIDをメタデータとしてリソースに保持しておき、IDでの検索が404になった場合にはメタデータを利用して検索するようにフォールバックする。 ### 実装案 `@previous-id=123456789012`のようなタグで変更前のIDを表す。 Terraform側はまずIDでの検索を試し、404になった場合は同一ゾーンのリソースから`@previous-id=<現在保持しているID>`というタグを条件に検索する。 API呼び出し例: `GET /server?{"Filter":{"Tags.Name":"@previous-id=123456789012"}}`
1.0
外部ツール経由でプラン変更された場合のID変更の追跡 - ### 概要 さくらのクラウドではプラン変更時にリソースのIDが変更される。 対象リソース: - サーバ - ルータ - ELB Terraformからプラン変更を行なった場合、IDが変更されることへの対応が実装されているが、 外部ツール(AutoScalerなど)からプラン変更が行われた場合にリソースが追跡できなくなる。 これは、各リソースのReadでIDを元にさくらのクラウドAPIを用いてリソースの情報を参照しているため。 (IDで検索し404が返ってきた場合はリソースが削除されたとみなす) 追跡できなくなってもimportすることで復帰させることが出来るが運用的に煩雑。 このため、何らかのルールにしたがって変更前のIDをメタデータとしてリソースに保持しておき、IDでの検索が404になった場合にはメタデータを利用して検索するようにフォールバックする。 ### 実装案 `@previous-id=123456789012`のようなタグで変更前のIDを表す。 Terraform側はまずIDでの検索を試し、404になった場合は同一ゾーンのリソースから`@previous-id=<現在保持しているID>`というタグを条件に検索する。 API呼び出し例: `GET /server?{"Filter":{"Tags.Name":"@previous-id=123456789012"}}`
priority
外部ツール経由でプラン変更された場合のid変更の追跡 概要 さくらのクラウドではプラン変更時にリソースのidが変更される。 対象リソース サーバ ルータ elb terraformからプラン変更を行なった場合、idが変更されることへの対応が実装されているが、 外部ツール autoscalerなど からプラン変更が行われた場合にリソースが追跡できなくなる。 これは、各リソースのreadでidを元にさくらのクラウドapiを用いてリソースの情報を参照しているため。 追跡できなくなってもimportすることで復帰させることが出来るが運用的に煩雑。 このため、何らかのルールにしたがって変更前のidをメタデータとしてリソースに保持しておき、 。 実装案 previous id のようなタグで変更前のidを表す。 terraform側はまずidでの検索を試し、 previous id というタグを条件に検索する。 api呼び出し例 get server filter tags name previous id
1
448,881
12,959,526,980
IssuesEvent
2020-07-20 13:08:14
wso2/micro-integrator
https://api.github.com/repos/wso2/micro-integrator
closed
Inconsistency in printing logs while using the payloadFactory mediator
Priority/High Severity/Major
**Description:** 1. Please find the following proxy service. ``` <?xml version="1.0" encoding="UTF-8"?> <proxy xmlns="http://ws.apache.org/ns/synapse" name="test" startOnLoad="true" statistics="disable" trace="disable" transports="http,https"> <target> <inSequence> <payloadFactory media-type="xml"> <format> <values>Test123</values> </format> <args/> </payloadFactory> <log level="full"/> <log> <property expression="//*[local-name()='values']" name="objects4"/> </log> <respond/> </inSequence> </target> <description/> </proxy> ``` 2. Deploy it inside wso2mi-1.1.0 and invoke it. Expected result ` <values>Test123</values>` Actual result `Test123` **Suggested Labels:** wso2mi-1.1.0, payloadFactory **Affected Product Version:** wso2mi-1.1.0 **OS, DB, other environment details and versions:** Linnux **Related Issues:** https://github.com/wso2/product-ei/issues/2092
1.0
Inconsistency in printing logs while using the payloadFactory mediator - **Description:** 1. Please find the following proxy service. ``` <?xml version="1.0" encoding="UTF-8"?> <proxy xmlns="http://ws.apache.org/ns/synapse" name="test" startOnLoad="true" statistics="disable" trace="disable" transports="http,https"> <target> <inSequence> <payloadFactory media-type="xml"> <format> <values>Test123</values> </format> <args/> </payloadFactory> <log level="full"/> <log> <property expression="//*[local-name()='values']" name="objects4"/> </log> <respond/> </inSequence> </target> <description/> </proxy> ``` 2. Deploy it inside wso2mi-1.1.0 and invoke it. Expected result ` <values>Test123</values>` Actual result `Test123` **Suggested Labels:** wso2mi-1.1.0, payloadFactory **Affected Product Version:** wso2mi-1.1.0 **OS, DB, other environment details and versions:** Linnux **Related Issues:** https://github.com/wso2/product-ei/issues/2092
priority
inconsistency in printing logs while using the payloadfactory mediator description please find the following proxy service deploy it inside and invoke it expected result actual result suggested labels payloadfactory affected product version os db other environment details and versions linnux related issues
1
185,292
6,720,769,436
IssuesEvent
2017-10-16 09:07:03
kedgeproject/kedge
https://api.github.com/repos/kedgeproject/kedge
closed
replicas set to 0 if not specified
kind/bug kind/task priority/high
If replicas is not set in Kedge file, the generated output shouldn't have it either. Now it defaults to 0, which leads to a confusing situation where your containers are not started when DC id deployed. ``` name: foo controller: deploymentconfig containers: - image: quay.io/tomkral/sleeper ``` ``` ▶ ./kedge generate -f test.yaml --- apiVersion: v1 kind: DeploymentConfig metadata: creationTimestamp: null name: foo spec: replicas: 0 strategy: resources: {} template: metadata: creationTimestamp: null spec: containers: - image: quay.io/tomkral/sleeper name: foo resources: {} test: false triggers: null status: availableReplicas: 0 latestVersion: 0 observedGeneration: 0 replicas: 0 unavailableReplicas: 0 updatedReplicas: 0 ``` replicas shouldn't be set if it's not set in Kedge file, OpenShift default will be used on the cluster side.
1.0
replicas set to 0 if not specified - If replicas is not set in Kedge file, the generated output shouldn't have it either. Now it defaults to 0, which leads to a confusing situation where your containers are not started when DC id deployed. ``` name: foo controller: deploymentconfig containers: - image: quay.io/tomkral/sleeper ``` ``` ▶ ./kedge generate -f test.yaml --- apiVersion: v1 kind: DeploymentConfig metadata: creationTimestamp: null name: foo spec: replicas: 0 strategy: resources: {} template: metadata: creationTimestamp: null spec: containers: - image: quay.io/tomkral/sleeper name: foo resources: {} test: false triggers: null status: availableReplicas: 0 latestVersion: 0 observedGeneration: 0 replicas: 0 unavailableReplicas: 0 updatedReplicas: 0 ``` replicas shouldn't be set if it's not set in Kedge file, OpenShift default will be used on the cluster side.
priority
replicas set to if not specified if replicas is not set in kedge file the generated output shouldn t have it either now it defaults to which leads to a confusing situation where your containers are not started when dc id deployed name foo controller deploymentconfig containers image quay io tomkral sleeper ▶ kedge generate f test yaml apiversion kind deploymentconfig metadata creationtimestamp null name foo spec replicas strategy resources template metadata creationtimestamp null spec containers image quay io tomkral sleeper name foo resources test false triggers null status availablereplicas latestversion observedgeneration replicas unavailablereplicas updatedreplicas replicas shouldn t be set if it s not set in kedge file openshift default will be used on the cluster side
1
517,364
15,007,623,422
IssuesEvent
2021-01-31 05:52:59
Left-on-Read/app
https://api.github.com/repos/Left-on-Read/app
closed
Implement one filter
high priority (p1) setup
Implement filtering - we can start with just a single filter, such as by date. ~This should be redux-driven.~
1.0
Implement one filter - Implement filtering - we can start with just a single filter, such as by date. ~This should be redux-driven.~
priority
implement one filter implement filtering we can start with just a single filter such as by date this should be redux driven
1
473,405
13,641,942,594
IssuesEvent
2020-09-25 14:50:47
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
CPU memory leak when using torch.no_grad()
high priority module: autograd topic: memory usage triage review triaged
## 🐛 Bug If use `torch.no_grad()` block, the cpu memory will continually increase untill OOM kill happens. But once remove the `no_grad`, everything would be all right. I tried del loss or put the validation step into a function, but the memory leak still happens. Is my code wrong? or a BUG? ## To Reproduce Steps to reproduce the behavior: 1. Validation step using `torch.no_grad()`. 2. Using `CrossEntropyLoss()` as criterion. 3. CPU RAM continually increasing occurs a OOM kill by system(Ubuntu 18.04). PyTorch 1.6 ``` for epoch in range(10): net.train() # Good training. for data in trainloader: inputs, labels = data['images'], data['masks'] for idx in range(0, len(inputs), 7): optimizer.zero_grad() outputs = net(inputs[idx:idx + 7]) loss = criterion(outputs, labels[idx:idx + 7]) loss.backward() optimizer.step() # Bad validation. net.eval() test_loss = 0.0 test_times = 0 for data in testloader: # !!!!!!!!👇 with torch.no_grad(): inputs, labels = data['images'], data['masks'] for idx in range(0, len(inputs), 7): # or put no_grad here, leaking still happens. outputs = net(inputs[idx:idx + 7]) loss = criterion(outputs, labels[idx:idx + 7]) test_loss += loss.item() test_times += 1 test_loss /= test_times ``` ## Expected behavior Normally valid without increasing CPU RAM. ## Environment - PyTorch Version (e.g., 1.0): 1.6 - OS (e.g., Linux): Ubuntu 18.04 - How you installed PyTorch (`conda`, `pip`, source): conda - Build command you used (if compiling from source): NaN - Python version: 3.8.5 - CUDA/cuDNN version: 10.2 - GPU models and configuration: Tesla V100 16G - Any other relevant information: with system RAM 128G ## Additional context Thanks for all your excellent work! cc @ezyang @gchanan @zou3519 @albanD @gqchen @pearu @nikitaved
1.0
CPU memory leak when using torch.no_grad() - ## 🐛 Bug If use `torch.no_grad()` block, the cpu memory will continually increase untill OOM kill happens. But once remove the `no_grad`, everything would be all right. I tried del loss or put the validation step into a function, but the memory leak still happens. Is my code wrong? or a BUG? ## To Reproduce Steps to reproduce the behavior: 1. Validation step using `torch.no_grad()`. 2. Using `CrossEntropyLoss()` as criterion. 3. CPU RAM continually increasing occurs a OOM kill by system(Ubuntu 18.04). PyTorch 1.6 ``` for epoch in range(10): net.train() # Good training. for data in trainloader: inputs, labels = data['images'], data['masks'] for idx in range(0, len(inputs), 7): optimizer.zero_grad() outputs = net(inputs[idx:idx + 7]) loss = criterion(outputs, labels[idx:idx + 7]) loss.backward() optimizer.step() # Bad validation. net.eval() test_loss = 0.0 test_times = 0 for data in testloader: # !!!!!!!!👇 with torch.no_grad(): inputs, labels = data['images'], data['masks'] for idx in range(0, len(inputs), 7): # or put no_grad here, leaking still happens. outputs = net(inputs[idx:idx + 7]) loss = criterion(outputs, labels[idx:idx + 7]) test_loss += loss.item() test_times += 1 test_loss /= test_times ``` ## Expected behavior Normally valid without increasing CPU RAM. ## Environment - PyTorch Version (e.g., 1.0): 1.6 - OS (e.g., Linux): Ubuntu 18.04 - How you installed PyTorch (`conda`, `pip`, source): conda - Build command you used (if compiling from source): NaN - Python version: 3.8.5 - CUDA/cuDNN version: 10.2 - GPU models and configuration: Tesla V100 16G - Any other relevant information: with system RAM 128G ## Additional context Thanks for all your excellent work! cc @ezyang @gchanan @zou3519 @albanD @gqchen @pearu @nikitaved
priority
cpu memory leak when using torch no grad 🐛 bug if use torch no grad block the cpu memory will continually increase untill oom kill happens but once remove the no grad everything would be all right i tried del loss or put the validation step into a function but the memory leak still happens is my code wrong or a bug to reproduce steps to reproduce the behavior validation step using torch no grad using crossentropyloss as criterion cpu ram continually increasing occurs a oom kill by system ubuntu pytorch for epoch in range net train good training for data in trainloader inputs labels data data for idx in range len inputs optimizer zero grad outputs net inputs loss criterion outputs labels loss backward optimizer step bad validation net eval test loss test times for data in testloader 👇 with torch no grad inputs labels data data for idx in range len inputs or put no grad here leaking still happens outputs net inputs loss criterion outputs labels test loss loss item test times test loss test times expected behavior normally valid without increasing cpu ram environment pytorch version e g os e g linux ubuntu how you installed pytorch conda pip source conda build command you used if compiling from source nan python version cuda cudnn version gpu models and configuration tesla any other relevant information with system ram additional context thanks for all your excellent work cc ezyang gchanan alband gqchen pearu nikitaved
1
617,097
19,342,641,181
IssuesEvent
2021-12-15 07:18:14
ballerina-platform/ballerina-dev-website
https://api.github.com/repos/ballerina-platform/ballerina-dev-website
closed
Add Content on How to Write a Connector in Bio
Priority/Highest Area/Docs Type/Task Points/1
**Description:** Need to update [1] according to the latest Swan Lake changes and add the content to Bio. [1] https://medium.com/ballerina-techblog/how-to-write-a-client-endpoint-in-ballerina-3c24c185ffaf **Suggested Labels:** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees:** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees--> **Affected Product Version:** **OS, Browser, other environment details and versions:** **Steps to reproduce:** **Related Issues:** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
1.0
Add Content on How to Write a Connector in Bio - **Description:** Need to update [1] according to the latest Swan Lake changes and add the content to Bio. [1] https://medium.com/ballerina-techblog/how-to-write-a-client-endpoint-in-ballerina-3c24c185ffaf **Suggested Labels:** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees:** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees--> **Affected Product Version:** **OS, Browser, other environment details and versions:** **Steps to reproduce:** **Related Issues:** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
priority
add content on how to write a connector in bio description need to update according to the latest swan lake changes and add the content to bio suggested labels suggested assignees affected product version os browser other environment details and versions steps to reproduce related issues
1
711,306
24,457,641,549
IssuesEvent
2022-10-07 08:17:32
AY2223S1-CS2113-F11-4/tp
https://api.github.com/repos/AY2223S1-CS2113-F11-4/tp
closed
Adding New Prescription Record: add
type.Story priority.High
As a doctor / user, I want to be able to add a a new prescription for a patient so that I can retrieve it for their future visits and records <img width="499" alt="image" src="https://user-images.githubusercontent.com/31297758/193443112-74662d47-2998-405d-9315-1da60dec36d5.png">
1.0
Adding New Prescription Record: add - As a doctor / user, I want to be able to add a a new prescription for a patient so that I can retrieve it for their future visits and records <img width="499" alt="image" src="https://user-images.githubusercontent.com/31297758/193443112-74662d47-2998-405d-9315-1da60dec36d5.png">
priority
adding new prescription record add as a doctor user i want to be able to add a a new prescription for a patient so that i can retrieve it for their future visits and records img width alt image src
1
312,047
9,542,320,880
IssuesEvent
2019-05-01 03:08:45
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
opened
Customer requisition finalisation crash
Bug Effort small Ivory Coast (phase 1) Priority: High
Build Number: 2.3.0-rc0 dev+apk Description: ![image](https://user-images.githubusercontent.com/7684221/57004385-c0274e00-6c22-11e9-91c9-9e62ef160db7.png) Reproducible: yes Reproduction Steps: 1. receive a customer requisition 2. enter a value 3. finalise it 4. some loading spinner and a RSoD (in dev, crash in apk) Comments: There were items in my customer requisition that my store didn't actually have visible in their store.
1.0
Customer requisition finalisation crash - Build Number: 2.3.0-rc0 dev+apk Description: ![image](https://user-images.githubusercontent.com/7684221/57004385-c0274e00-6c22-11e9-91c9-9e62ef160db7.png) Reproducible: yes Reproduction Steps: 1. receive a customer requisition 2. enter a value 3. finalise it 4. some loading spinner and a RSoD (in dev, crash in apk) Comments: There were items in my customer requisition that my store didn't actually have visible in their store.
priority
customer requisition finalisation crash build number dev apk description reproducible yes reproduction steps receive a customer requisition enter a value finalise it some loading spinner and a rsod in dev crash in apk comments there were items in my customer requisition that my store didn t actually have visible in their store
1
460,117
13,205,116,210
IssuesEvent
2020-08-14 17:15:56
googleinterns/bazel-rules-fuzzing
https://api.github.com/repos/googleinterns/bazel-rules-fuzzing
closed
Enable regression support in the launcher script
high priority
## Expected Behavior Launcher needs to support running regression test without continuous fuzzing test. To achieve this, a string_flag `engine` will be added to decide the launcher's behavior. After the launcher is modified, a new rule `regression_launcher` will be needed to start the launcher in the regression mode. ## Actual Behavior Only continuous fuzzing test mode is supported
1.0
Enable regression support in the launcher script - ## Expected Behavior Launcher needs to support running regression test without continuous fuzzing test. To achieve this, a string_flag `engine` will be added to decide the launcher's behavior. After the launcher is modified, a new rule `regression_launcher` will be needed to start the launcher in the regression mode. ## Actual Behavior Only continuous fuzzing test mode is supported
priority
enable regression support in the launcher script expected behavior launcher needs to support running regression test without continuous fuzzing test to achieve this a string flag engine will be added to decide the launcher s behavior after the launcher is modified a new rule regression launcher will be needed to start the launcher in the regression mode actual behavior only continuous fuzzing test mode is supported
1
36,066
2,795,249,985
IssuesEvent
2015-05-11 20:59:18
Arabidopsis-Information-Portal/adama
https://api.github.com/repos/Arabidopsis-Information-Portal/adama
closed
Restrict names of namespaces to current allowed characters
bug high priority
In Adama 0.3 the namespace together with the service name correspond to a container image. The only characters allowed for a container image are [a-z0-9_.-]. This is an implementation detail and it **should not** leak to the user (see #45 and #3). While we fix those issues, Adama should refuse to create a namespace with forbidden characters. Right now, Adama accepts namespaces with forbidden characters, and it refuses service names with those characters. However, even if the service name is compliant, being under a non-compliant namespace produces an invalid image name.
1.0
Restrict names of namespaces to current allowed characters - In Adama 0.3 the namespace together with the service name correspond to a container image. The only characters allowed for a container image are [a-z0-9_.-]. This is an implementation detail and it **should not** leak to the user (see #45 and #3). While we fix those issues, Adama should refuse to create a namespace with forbidden characters. Right now, Adama accepts namespaces with forbidden characters, and it refuses service names with those characters. However, even if the service name is compliant, being under a non-compliant namespace produces an invalid image name.
priority
restrict names of namespaces to current allowed characters in adama the namespace together with the service name correspond to a container image the only characters allowed for a container image are this is an implementation detail and it should not leak to the user see and while we fix those issues adama should refuse to create a namespace with forbidden characters right now adama accepts namespaces with forbidden characters and it refuses service names with those characters however even if the service name is compliant being under a non compliant namespace produces an invalid image name
1
744,886
25,959,460,957
IssuesEvent
2022-12-18 17:54:16
zigtools/zls
https://api.github.com/repos/zigtools/zls
closed
Integer overflow when parsing incomplete function
bug priority:high fuzzing result
### Zig Version 0.11.0-dev.782+0b4461d97 ### Zig Language Server Version 3526f5fb84b89b6327fc21c6b836b842bf4db90d ### Steps to Reproduce Open the following file: ```zig AtEnd() void { _ = @import("std"); const expect = try expect(result[1]); var buf = [_:0]u8{ 'a', 'b'}, 42.0f}; }; }); if (builtin.zig_backend == .stage2_wasm) return error.SkipZigTest; // TODO try expect(@as(u32, 0xFE00BE00), structPtr.*); } pub fn set_checksum_size = 1000; v.func_field(0) == 2); try expect(ctz(@as(u121, 3)), const tmp = ref.*; \\#define foo "a string"; .is_var_args); try testOneCtzVector(u128, 64, @splat(4, @as(u24, 0x6a2c48))); const X = struct { \\const std = @import("std"); pub fn main() void { var bad: f128 = 0.0_; _ = @Type(@typeInfo(Foo).Union.tag_type == f32); try expect(length == std.mem.eql(u8, messages here is a VarDecl in scope } \\0: first arg \\ return error.SkipZigTest; // TODO comptime try expect(G3.Fn.return_trace; return Foo {.x = 13}; } var some_struct_param_type() void { _ = v; }; try testTruncWithVectors() !void { var target: [*c]u8 = @ptrCast([*]u8, &l.array); try expect(a == -1); try S.entry(true, false }; _ = c.printf("0.000000000FFFFFFFF000000000000p+3 .a = .{ S.declaration of label 'blk' // :8:12: error: union declared here // :24:20: error: loop in named 'c' in enum 'tmp.Letter' // :1:7: error: C import block // :5:10: error: parameters cannot cast into pointer" { const X = struct field" { var a: c_int = @ptrCast(*u32, &bytes); try expect(result == 1) catch unreachable; } // error // backend=stage2 // target=native // // :1:8: error: this is a longer message, "integer overflow = @addWithOverflow(u64, a, b, &res)); \\} try expectEqual = std.testing.expectEqualStrings = std.fs; pub const B = enum(u8) { const x: f32 = 0.0; var default value stored to trigger the bug //a second vector initialization syntax // :2:1: error: expectedBitSize = 80; comptime { return error union by return" { const S = extern struct { const std = @import("behavior = .Pipe; \\} }); const i = g + h; // 100 return 1234; var array = array[i]; fn foo(a: bool, b: bool = true, } } // error // backend=stage2 // target=native // // :2:5: note: control flow inside runtime isNan(nan_times_zero)); .build_modes = true, }; _ = c; b1_6: u1, resume frame; try expectError(error.FailedToCreateEntry; defer stdout.print("All your code here, or load and store" { if (comptime { const msg = @ptrToInt("Hello, World!\n")), ); } fn b() void { if (true) {} else |err| { .{ .name = "unsigned int choose[1][1] == 'a'); } test "double nested unpacked", \\source.zig:11:5: [address] in foo (test) } }; var array1: [4]u8 = "aoeu"; try std.testing.expect; const expr" { if (builtin.zig_backend == .stage2_x86_64) return false; comptime try expect(12 == default initialization syntax // :2:12: error: expected { 0, 0 }, @as(u64, 0x3ff000000000000000000000000000000000000000000000FFFFFFFFFFFFFFFFFFFFFFFFFF); try expect(comptime { _ = Foo; } // error // backend=stage2 // target=native // // :4:30: error: expected type 'anyframe->i32); try consume_tuple(t2 ++ .{0}, 2); } }; .step = b.step("test", write_src = b.addObjectFromWriteFile(src_basename).?); \\4: " .fnPtr = bar, \\pub export fn entry() void { _ = ignore; \\int foo(); } const A = struct { \\ mov $14, %rdx fn foo() void { try expect(mem.eql(u8, message, "integer = u32; \\pub export fn entry3() callconv(.Async) void { var array = [_]u8{ .y = 2 }; const zig_args.append(list); \\ if (builtin.zig_backend == .stage2_sparc64) return x; _ = stack_trace: ?*std.builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO var b: i8 = -18; try foo(); try std.testing.ex ``` ### Expected Behavior It doesn't integer underflow. ### Actual Behavior ```log thread 16460 panic: integer overflow C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:2192:61: 0x7ff6a4baee34 in fullCall (zls.exe.obj) const maybe_async_token = tree.firstToken(info.fn_expr) - 1; ^ C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:1883:24: 0x7ff6a4b3f2af in callOne (zls.exe.obj) .fn_expr = data.lhs, ^ C:\Programming\Zig\buzz\repos\zls\src\ast.zig:1086:24: 0x7ff6a4a7810a in callFull (zls.exe.obj) => tree.callOne(buf, node), ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2964:38: 0x7ff6a49d7864 in makeScopeInternal (zls.exe.obj) const call = ast.callFull(tree, node_idx, &buf).?; ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3025:64: 0x7ff6a49d8166 in makeScopeInternal (zls.exe.obj) try makeScopeInternal(allocator, context, field.ast.type_expr); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2560:30: 0x7ff6a4a76a93 in makeInnerScope (zls.exe.obj) try makeScopeInternal(allocator, context, decl); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2631:31: 0x7ff6a49d4df6 in makeScopeInternal (zls.exe.obj) try makeInnerScope(allocator, context, node_idx); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2496:33: 0x7ff6a49d41cd in makeDocumentScope (zls.exe.obj) .enums = &document_scope.enum_completions, ^ C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:612:65: 0x7ff6a4a6493f in createDocument (zls.exe.obj) var document_scope = try analysis.makeDocumentScope(self.allocator, tree); ^ C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:158:39: 0x7ff6a49d3268 in openDocument (zls.exe.obj) handle.* = try self.createDocument(duped_uri, duped_text, true); ^ C:\Programming\Zig\buzz\repos\zls\src\Server.zig:1867:111: 0x7ff6a49d2e2f in openDocumentHandler__anon_12308 (zls.exe.obj) const handle = try server.document_store.openDocument(req.params.textDocument.uri, req.params.textDocument.text); ^ C:\Programming\Zig\buzz\repos\zls\src\Server.zig:2964:35: 0x7ff6a4a36475 in processJsonRpc__anon_10459 (zls.exe.obj) method_info[2](server, writer, id, request_obj) catch |err| { ^ C:\Programming\Zig\buzz\repos\zls\src\main.zig:51:34: 0x7ff6a4a3d754 in loop (zls.exe.obj) try server.processJsonRpc(writer, buffer); ^ C:\Programming\Zig\buzz\repos\zls\src\main.zig:281:13: 0x7ff6a4a3dbd2 in main (zls.exe.obj) try loop(&server); ^ C:\Programming\Zig\zig-from-the-website\lib\std\start.zig:385:41: 0x7ff6a4a3e077 in WinStartup (zls.exe.obj) std.debug.maybeEnableSegfaultHandler(); ^ ???:?:?: 0x7ffab1d9559f in ??? (???) ???:?:?: 0x7ffab2c0485a in ??? (???) ``` (By the way this *might be* a Zig issue - shouldn't the parser always error instead of underflowing, or is this just super cursed?)
1.0
Integer overflow when parsing incomplete function - ### Zig Version 0.11.0-dev.782+0b4461d97 ### Zig Language Server Version 3526f5fb84b89b6327fc21c6b836b842bf4db90d ### Steps to Reproduce Open the following file: ```zig AtEnd() void { _ = @import("std"); const expect = try expect(result[1]); var buf = [_:0]u8{ 'a', 'b'}, 42.0f}; }; }); if (builtin.zig_backend == .stage2_wasm) return error.SkipZigTest; // TODO try expect(@as(u32, 0xFE00BE00), structPtr.*); } pub fn set_checksum_size = 1000; v.func_field(0) == 2); try expect(ctz(@as(u121, 3)), const tmp = ref.*; \\#define foo "a string"; .is_var_args); try testOneCtzVector(u128, 64, @splat(4, @as(u24, 0x6a2c48))); const X = struct { \\const std = @import("std"); pub fn main() void { var bad: f128 = 0.0_; _ = @Type(@typeInfo(Foo).Union.tag_type == f32); try expect(length == std.mem.eql(u8, messages here is a VarDecl in scope } \\0: first arg \\ return error.SkipZigTest; // TODO comptime try expect(G3.Fn.return_trace; return Foo {.x = 13}; } var some_struct_param_type() void { _ = v; }; try testTruncWithVectors() !void { var target: [*c]u8 = @ptrCast([*]u8, &l.array); try expect(a == -1); try S.entry(true, false }; _ = c.printf("0.000000000FFFFFFFF000000000000p+3 .a = .{ S.declaration of label 'blk' // :8:12: error: union declared here // :24:20: error: loop in named 'c' in enum 'tmp.Letter' // :1:7: error: C import block // :5:10: error: parameters cannot cast into pointer" { const X = struct field" { var a: c_int = @ptrCast(*u32, &bytes); try expect(result == 1) catch unreachable; } // error // backend=stage2 // target=native // // :1:8: error: this is a longer message, "integer overflow = @addWithOverflow(u64, a, b, &res)); \\} try expectEqual = std.testing.expectEqualStrings = std.fs; pub const B = enum(u8) { const x: f32 = 0.0; var default value stored to trigger the bug //a second vector initialization syntax // :2:1: error: expectedBitSize = 80; comptime { return error union by return" { const S = extern struct { const std = @import("behavior = .Pipe; \\} }); const i = g + h; // 100 return 1234; var array = array[i]; fn foo(a: bool, b: bool = true, } } // error // backend=stage2 // target=native // // :2:5: note: control flow inside runtime isNan(nan_times_zero)); .build_modes = true, }; _ = c; b1_6: u1, resume frame; try expectError(error.FailedToCreateEntry; defer stdout.print("All your code here, or load and store" { if (comptime { const msg = @ptrToInt("Hello, World!\n")), ); } fn b() void { if (true) {} else |err| { .{ .name = "unsigned int choose[1][1] == 'a'); } test "double nested unpacked", \\source.zig:11:5: [address] in foo (test) } }; var array1: [4]u8 = "aoeu"; try std.testing.expect; const expr" { if (builtin.zig_backend == .stage2_x86_64) return false; comptime try expect(12 == default initialization syntax // :2:12: error: expected { 0, 0 }, @as(u64, 0x3ff000000000000000000000000000000000000000000000FFFFFFFFFFFFFFFFFFFFFFFFFF); try expect(comptime { _ = Foo; } // error // backend=stage2 // target=native // // :4:30: error: expected type 'anyframe->i32); try consume_tuple(t2 ++ .{0}, 2); } }; .step = b.step("test", write_src = b.addObjectFromWriteFile(src_basename).?); \\4: " .fnPtr = bar, \\pub export fn entry() void { _ = ignore; \\int foo(); } const A = struct { \\ mov $14, %rdx fn foo() void { try expect(mem.eql(u8, message, "integer = u32; \\pub export fn entry3() callconv(.Async) void { var array = [_]u8{ .y = 2 }; const zig_args.append(list); \\ if (builtin.zig_backend == .stage2_sparc64) return x; _ = stack_trace: ?*std.builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO var b: i8 = -18; try foo(); try std.testing.ex ``` ### Expected Behavior It doesn't integer underflow. ### Actual Behavior ```log thread 16460 panic: integer overflow C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:2192:61: 0x7ff6a4baee34 in fullCall (zls.exe.obj) const maybe_async_token = tree.firstToken(info.fn_expr) - 1; ^ C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:1883:24: 0x7ff6a4b3f2af in callOne (zls.exe.obj) .fn_expr = data.lhs, ^ C:\Programming\Zig\buzz\repos\zls\src\ast.zig:1086:24: 0x7ff6a4a7810a in callFull (zls.exe.obj) => tree.callOne(buf, node), ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2964:38: 0x7ff6a49d7864 in makeScopeInternal (zls.exe.obj) const call = ast.callFull(tree, node_idx, &buf).?; ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3025:64: 0x7ff6a49d8166 in makeScopeInternal (zls.exe.obj) try makeScopeInternal(allocator, context, field.ast.type_expr); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2560:30: 0x7ff6a4a76a93 in makeInnerScope (zls.exe.obj) try makeScopeInternal(allocator, context, decl); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2631:31: 0x7ff6a49d4df6 in makeScopeInternal (zls.exe.obj) try makeInnerScope(allocator, context, node_idx); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2496:33: 0x7ff6a49d41cd in makeDocumentScope (zls.exe.obj) .enums = &document_scope.enum_completions, ^ C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:612:65: 0x7ff6a4a6493f in createDocument (zls.exe.obj) var document_scope = try analysis.makeDocumentScope(self.allocator, tree); ^ C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:158:39: 0x7ff6a49d3268 in openDocument (zls.exe.obj) handle.* = try self.createDocument(duped_uri, duped_text, true); ^ C:\Programming\Zig\buzz\repos\zls\src\Server.zig:1867:111: 0x7ff6a49d2e2f in openDocumentHandler__anon_12308 (zls.exe.obj) const handle = try server.document_store.openDocument(req.params.textDocument.uri, req.params.textDocument.text); ^ C:\Programming\Zig\buzz\repos\zls\src\Server.zig:2964:35: 0x7ff6a4a36475 in processJsonRpc__anon_10459 (zls.exe.obj) method_info[2](server, writer, id, request_obj) catch |err| { ^ C:\Programming\Zig\buzz\repos\zls\src\main.zig:51:34: 0x7ff6a4a3d754 in loop (zls.exe.obj) try server.processJsonRpc(writer, buffer); ^ C:\Programming\Zig\buzz\repos\zls\src\main.zig:281:13: 0x7ff6a4a3dbd2 in main (zls.exe.obj) try loop(&server); ^ C:\Programming\Zig\zig-from-the-website\lib\std\start.zig:385:41: 0x7ff6a4a3e077 in WinStartup (zls.exe.obj) std.debug.maybeEnableSegfaultHandler(); ^ ???:?:?: 0x7ffab1d9559f in ??? (???) ???:?:?: 0x7ffab2c0485a in ??? (???) ``` (By the way this *might be* a Zig issue - shouldn't the parser always error instead of underflowing, or is this just super cursed?)
priority
integer overflow when parsing incomplete function zig version dev zig language server version steps to reproduce open the following file zig atend void import std const expect try expect result var buf a b if builtin zig backend wasm return error skipzigtest todo try expect as structptr pub fn set checksum size v func field try expect ctz as const tmp ref define foo a string is var args try testonectzvector splat as const x struct const std import std pub fn main void var bad type typeinfo foo union tag type try expect length std mem eql messages here is a vardecl in scope first arg return error skipzigtest todo comptime try expect fn return trace return foo x var some struct param type void v try testtruncwithvectors void var target ptrcast l array try expect a try s entry true false c printf a s declaration of label blk error union declared here error loop in named c in enum tmp letter error c import block error parameters cannot cast into pointer const x struct field var a c int ptrcast bytes try expect result catch unreachable error backend target native error this is a longer message integer overflow addwithoverflow a b res try expectequal std testing expectequalstrings std fs pub const b enum const x var default value stored to trigger the bug a second vector initialization syntax error expectedbitsize comptime return error union by return const s extern struct const std import behavior pipe const i g h return var array array fn foo a bool b bool true error backend target native note control flow inside runtime isnan nan times zero build modes true c resume frame try expecterror error failedtocreateentry defer stdout print all your code here or load and store if comptime const msg ptrtoint hello world n fn b void if true else err name unsigned int choose a test double nested unpacked source zig in foo test var aoeu try std testing expect const expr if builtin zig backend return false comptime try expect default initialization syntax error expected as try expect comptime foo error backend target native error expected type anyframe try consume tuple step b step test write src b addobjectfromwritefile src basename fnptr bar pub export fn entry void ignore int foo const a struct mov rdx fn foo void try expect mem eql message integer pub export fn callconv async void var array y const zig args append list if builtin zig backend return x stack trace std builtin zig backend return error skipzigtest todo var b try foo try std testing ex expected behavior it doesn t integer underflow actual behavior log thread panic integer overflow c programming zig zig from the website lib std zig ast zig in fullcall zls exe obj const maybe async token tree firsttoken info fn expr c programming zig zig from the website lib std zig ast zig in callone zls exe obj fn expr data lhs c programming zig buzz repos zls src ast zig in callfull zls exe obj tree callone buf node c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj const call ast callfull tree node idx buf c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj try makescopeinternal allocator context field ast type expr c programming zig buzz repos zls src analysis zig in makeinnerscope zls exe obj try makescopeinternal allocator context decl c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj try makeinnerscope allocator context node idx c programming zig buzz repos zls src analysis zig in makedocumentscope zls exe obj enums document scope enum completions c programming zig buzz repos zls src documentstore zig in createdocument zls exe obj var document scope try analysis makedocumentscope self allocator tree c programming zig buzz repos zls src documentstore zig in opendocument zls exe obj handle try self createdocument duped uri duped text true c programming zig buzz repos zls src server zig in opendocumenthandler anon zls exe obj const handle try server document store opendocument req params textdocument uri req params textdocument text c programming zig buzz repos zls src server zig in processjsonrpc anon zls exe obj method info server writer id request obj catch err c programming zig buzz repos zls src main zig in loop zls exe obj try server processjsonrpc writer buffer c programming zig buzz repos zls src main zig in main zls exe obj try loop server c programming zig zig from the website lib std start zig in winstartup zls exe obj std debug maybeenablesegfaulthandler in in by the way this might be a zig issue shouldn t the parser always error instead of underflowing or is this just super cursed
1
823,311
30,989,685,286
IssuesEvent
2023-08-09 02:47:47
Karooobar/Voyager
https://api.github.com/repos/Karooobar/Voyager
closed
Need to pass storeId, owner ID, from the user who is logged in
High Priority
Right now in helpers like ItemHelper, CategoryHelper, and storeHelper, we are hardcoding the store id that is 200. We need to modify this so that we use the store id of the user that is logged in.
1.0
Need to pass storeId, owner ID, from the user who is logged in - Right now in helpers like ItemHelper, CategoryHelper, and storeHelper, we are hardcoding the store id that is 200. We need to modify this so that we use the store id of the user that is logged in.
priority
need to pass storeid owner id from the user who is logged in right now in helpers like itemhelper categoryhelper and storehelper we are hardcoding the store id that is we need to modify this so that we use the store id of the user that is logged in
1
629,928
20,071,470,861
IssuesEvent
2022-02-04 07:33:07
debops/debops
https://api.github.com/repos/debops/debops
closed
bootstrap-sssd and bootstrap-ldap don't seem to be idempotent
bug priority: high tag: LDAP
Both ``nslcd`` and ``sssd`` contain configuration file generation tasks which include directives like: ``` - name: Generate nslcd configuration ... when: nslcd__ldap_base_dn|d() ``` and: ``` - name: Generate sssd configuration ... when: sssd__ldap_base_dn|d() ``` The ``nslcd__ldap_base_dn`` variables and ``sssd__ldap_base_dn`` variables are both defined as: `'{{ ansible_local.ldap.base_dn|d([]) }}'` In the ``bootstrap-*`` case, ``ansible_local.ldap.base_dn`` is initially undefined, and later set by the ``debops.ldap`` role. However, the ``sssd__ldap_base_dn`` and ``nslcd__ldap_base_dn`` variables aren't recalculated, so the configuration files ``/etc/sssd/sssd.conf`` and ``/etc/nslcd.conf`` files aren't generated (in the latter case, a default file is generated by the DPKG package, so the fact the file is present is misleading). I tried copying the trick from ``debops.ldap``: ``` - name: Take note of the current LDAP configuration set_fact: # Track the changes in the configuration state # between role executions in the same play. ldap__fact_configured: '{{ ldap__configured }}' # Re-instantiate dependent variables to evaluate variables that use them. # Without this, dependent variables may contain outdated configuration. ldap__fact_dependent_tasks: '{{ ldap__dependent_tasks }}' tags: [ 'role::ldap:tasks', 'skip::ldap:tasks' ] ``` To update ``sssd__ldap_base_dn`` and ``nslcd__ldap_base_dn``, but that doesn't seem to actually recalculate them. A simple workaround is to run the ``bootstrap-*`` scripts twice for a host...where the second invocation will show: ``` PLAY RECAP *********************************************************************************** example : ok=200 changed=2 unreachable=0 failed=0 skipped=138 rescued=0 ignored=0 ``` (Note the ``changed`` value) Would be happy to do a PR, but I'm kind of stumped here (tried on ``2.10.7+merged+base+2.10.8+dfsg-1`` from Debian unstable).
1.0
bootstrap-sssd and bootstrap-ldap don't seem to be idempotent - Both ``nslcd`` and ``sssd`` contain configuration file generation tasks which include directives like: ``` - name: Generate nslcd configuration ... when: nslcd__ldap_base_dn|d() ``` and: ``` - name: Generate sssd configuration ... when: sssd__ldap_base_dn|d() ``` The ``nslcd__ldap_base_dn`` variables and ``sssd__ldap_base_dn`` variables are both defined as: `'{{ ansible_local.ldap.base_dn|d([]) }}'` In the ``bootstrap-*`` case, ``ansible_local.ldap.base_dn`` is initially undefined, and later set by the ``debops.ldap`` role. However, the ``sssd__ldap_base_dn`` and ``nslcd__ldap_base_dn`` variables aren't recalculated, so the configuration files ``/etc/sssd/sssd.conf`` and ``/etc/nslcd.conf`` files aren't generated (in the latter case, a default file is generated by the DPKG package, so the fact the file is present is misleading). I tried copying the trick from ``debops.ldap``: ``` - name: Take note of the current LDAP configuration set_fact: # Track the changes in the configuration state # between role executions in the same play. ldap__fact_configured: '{{ ldap__configured }}' # Re-instantiate dependent variables to evaluate variables that use them. # Without this, dependent variables may contain outdated configuration. ldap__fact_dependent_tasks: '{{ ldap__dependent_tasks }}' tags: [ 'role::ldap:tasks', 'skip::ldap:tasks' ] ``` To update ``sssd__ldap_base_dn`` and ``nslcd__ldap_base_dn``, but that doesn't seem to actually recalculate them. A simple workaround is to run the ``bootstrap-*`` scripts twice for a host...where the second invocation will show: ``` PLAY RECAP *********************************************************************************** example : ok=200 changed=2 unreachable=0 failed=0 skipped=138 rescued=0 ignored=0 ``` (Note the ``changed`` value) Would be happy to do a PR, but I'm kind of stumped here (tried on ``2.10.7+merged+base+2.10.8+dfsg-1`` from Debian unstable).
priority
bootstrap sssd and bootstrap ldap don t seem to be idempotent both nslcd and sssd contain configuration file generation tasks which include directives like name generate nslcd configuration when nslcd ldap base dn d and name generate sssd configuration when sssd ldap base dn d the nslcd ldap base dn variables and sssd ldap base dn variables are both defined as ansible local ldap base dn d in the bootstrap case ansible local ldap base dn is initially undefined and later set by the debops ldap role however the sssd ldap base dn and nslcd ldap base dn variables aren t recalculated so the configuration files etc sssd sssd conf and etc nslcd conf files aren t generated in the latter case a default file is generated by the dpkg package so the fact the file is present is misleading i tried copying the trick from debops ldap name take note of the current ldap configuration set fact track the changes in the configuration state between role executions in the same play ldap fact configured ldap configured re instantiate dependent variables to evaluate variables that use them without this dependent variables may contain outdated configuration ldap fact dependent tasks ldap dependent tasks tags to update sssd ldap base dn and nslcd ldap base dn but that doesn t seem to actually recalculate them a simple workaround is to run the bootstrap scripts twice for a host where the second invocation will show play recap example ok changed unreachable failed skipped rescued ignored note the changed value would be happy to do a pr but i m kind of stumped here tried on merged base dfsg from debian unstable
1
343,656
10,334,532,770
IssuesEvent
2019-09-03 08:34:34
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
[master-preview] Broken Treasuary with new Government system
High Priority
1. You can place Treasury and change Tax without rights. 2. Elected Title with access to Treasury doesn't give you access to Treasury ![unknown (2)](https://user-images.githubusercontent.com/45708377/64156280-21fbfc80-ce3d-11e9-8479-ce8ef0228ab4.png) 3. It's in another table, but It's directly connected to Treasury: The description in the elected title that you can set the tax. But I think it’s need to describe not only the tax but also Treasury itself. ![image](https://user-images.githubusercontent.com/45708377/64156949-689e2680-ce3e-11e9-8c4d-c2c025b0f95a.png) 4. Same trouble at 8.3.0: You can set tax more then 100%. Need to limit it. ![unknown (3)](https://user-images.githubusercontent.com/45708377/64156984-7ce22380-ce3e-11e9-9f07-1fbce00c9f94.png)
1.0
[master-preview] Broken Treasuary with new Government system - 1. You can place Treasury and change Tax without rights. 2. Elected Title with access to Treasury doesn't give you access to Treasury ![unknown (2)](https://user-images.githubusercontent.com/45708377/64156280-21fbfc80-ce3d-11e9-8479-ce8ef0228ab4.png) 3. It's in another table, but It's directly connected to Treasury: The description in the elected title that you can set the tax. But I think it’s need to describe not only the tax but also Treasury itself. ![image](https://user-images.githubusercontent.com/45708377/64156949-689e2680-ce3e-11e9-8c4d-c2c025b0f95a.png) 4. Same trouble at 8.3.0: You can set tax more then 100%. Need to limit it. ![unknown (3)](https://user-images.githubusercontent.com/45708377/64156984-7ce22380-ce3e-11e9-9f07-1fbce00c9f94.png)
priority
broken treasuary with new government system you can place treasury and change tax without rights elected title with access to treasury doesn t give you access to treasury it s in another table but it s directly connected to treasury the description in the elected title that you can set the tax but i think it’s need to describe not only the tax but also treasury itself same trouble at you can set tax more then need to limit it
1
471,168
13,561,468,572
IssuesEvent
2020-09-18 04:38:40
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
IDP, Add authentication configuration fail
Affected/5.11.0-m31 Component/Identity Apps Priority/High Severity/Critical bug console dev-exp
**Describe the issue:** Following check box at IDP , authentication configuration wizard check box are not clickable 1. passive STS authentication configuration wizard check boxes 2. SAML authentication configuration wizard check boxes console error: ``` vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 Uncaught TypeError: Cannot read property 'toString' of undefined at Object.onChange (main.babc4f7b.chunk.js?ef057d352ba2713628dc:1) at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 at t.handleChange (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 at Object.s (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at d (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 ``` **How to reproduce:** 1. login to console 2. Navigate to the developer page and identity provider 3. Click to add and and new IDP 4. Click to edit IDP and select Authentication tab 5. Select passive STS the Authenticator and click next 6. click check boxes of these , -Passive STS user ID location -Enable SAML assertion signature validation - Enable SAML assertion audience validation **Expected behavior:** Check box not select as expected **Environment information** - Product Version: [e.g., IS 5.11.0-m31]
1.0
IDP, Add authentication configuration fail - **Describe the issue:** Following check box at IDP , authentication configuration wizard check box are not clickable 1. passive STS authentication configuration wizard check boxes 2. SAML authentication configuration wizard check boxes console error: ``` vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 Uncaught TypeError: Cannot read property 'toString' of undefined at Object.onChange (main.babc4f7b.chunk.js?ef057d352ba2713628dc:1) at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 at t.handleChange (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 at Object.s (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at d (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2) at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 ``` **How to reproduce:** 1. login to console 2. Navigate to the developer page and identity provider 3. Click to add and and new IDP 4. Click to edit IDP and select Authentication tab 5. Select passive STS the Authenticator and click next 6. click check boxes of these , -Passive STS user ID location -Enable SAML assertion signature validation - Enable SAML assertion audience validation **Expected behavior:** Check box not select as expected **Environment information** - Product Version: [e.g., IS 5.11.0-m31]
priority
idp add authentication configuration fail describe the issue following check box at idp authentication configuration wizard check box are not clickable passive sts authentication configuration wizard check boxes saml authentication configuration wizard check boxes console error vendors main chunk js uncaught typeerror cannot read property tostring of undefined at object onchange main chunk js at e exports vendors main chunk js at e exports vendors main chunk js at e exports vendors main chunk js at vendors main chunk js at t handlechange vendors main chunk js at vendors main chunk js at object s vendors main chunk js at d vendors main chunk js at vendors main chunk js how to reproduce login to console navigate to the developer page and identity provider click to add and and new idp click to edit idp and select authentication tab select passive sts the authenticator and click next click check boxes of these passive sts user id location enable saml assertion signature validation enable saml assertion audience validation expected behavior check box not select as expected environment information product version
1
405,339
11,871,802,750
IssuesEvent
2020-03-26 14:54:58
kenodressel/quarantine-hero
https://api.github.com/repos/kenodressel/quarantine-hero
closed
Reactivate Google GeoLocation API
enhancement inprogress priority-high
- [ ] Migrate Database (plz --> geolocations) - [ ] Verify that the website still works when we flip the feature flag - [ ] Verify that the firebase functions still works when we flip the feature flag - [ ] make the repo backwards compatible so that we can always go back to plz with out migration The backwards compatibility is enforce by only allowing users to input a plz in the "askForHelp" form and the "offer-help" form.
1.0
Reactivate Google GeoLocation API - - [ ] Migrate Database (plz --> geolocations) - [ ] Verify that the website still works when we flip the feature flag - [ ] Verify that the firebase functions still works when we flip the feature flag - [ ] make the repo backwards compatible so that we can always go back to plz with out migration The backwards compatibility is enforce by only allowing users to input a plz in the "askForHelp" form and the "offer-help" form.
priority
reactivate google geolocation api migrate database plz geolocations verify that the website still works when we flip the feature flag verify that the firebase functions still works when we flip the feature flag make the repo backwards compatible so that we can always go back to plz with out migration the backwards compatibility is enforce by only allowing users to input a plz in the askforhelp form and the offer help form
1
576,032
17,069,359,417
IssuesEvent
2021-07-07 11:23:28
ParamagicDev/mrujs
https://api.github.com/repos/ParamagicDev/mrujs
closed
Feature: Turbo integration
Priority: High Priority: Normal Project Size: Medium Type: Feature
#### Describe Problem currently Turbo + Mrujs have quite a bit of intersection. I'm not even sure if this is a good idea tbh. This is more so a "future possibility" there will be a lot of trial and error to make this happen. So this issue will serve as a future reference of possible changes that need to be made. #### Suggest Solution first change: modify querySelectors. https://github.com/rails/rails/commit/8de181dac638457564677e02c71a9d04fd3e7fde
2.0
Feature: Turbo integration - #### Describe Problem currently Turbo + Mrujs have quite a bit of intersection. I'm not even sure if this is a good idea tbh. This is more so a "future possibility" there will be a lot of trial and error to make this happen. So this issue will serve as a future reference of possible changes that need to be made. #### Suggest Solution first change: modify querySelectors. https://github.com/rails/rails/commit/8de181dac638457564677e02c71a9d04fd3e7fde
priority
feature turbo integration describe problem currently turbo mrujs have quite a bit of intersection i m not even sure if this is a good idea tbh this is more so a future possibility there will be a lot of trial and error to make this happen so this issue will serve as a future reference of possible changes that need to be made suggest solution first change modify queryselectors
1
594,469
18,046,341,463
IssuesEvent
2021-09-19 00:32:13
practice-enterprise/api
https://api.github.com/repos/practice-enterprise/api
opened
All requests to DB needs to lowered
Priority: High Type: Enhancement
- [ ] Adjust polling - [ ] Lower requests to DB drastically - [ ] Announcements, user calls - [ ] ...
1.0
All requests to DB needs to lowered - - [ ] Adjust polling - [ ] Lower requests to DB drastically - [ ] Announcements, user calls - [ ] ...
priority
all requests to db needs to lowered adjust polling lower requests to db drastically announcements user calls
1
343,070
10,325,021,752
IssuesEvent
2019-09-01 14:09:33
Eve-ning/vsrg
https://api.github.com/repos/Eve-ning/vsrg
closed
Use directories instead of filters
aesthetics high-priority
Getting messy with game specific implementations, it's either I file them properly or use another repo
1.0
Use directories instead of filters - Getting messy with game specific implementations, it's either I file them properly or use another repo
priority
use directories instead of filters getting messy with game specific implementations it s either i file them properly or use another repo
1
480,030
13,822,378,502
IssuesEvent
2020-10-13 04:56:03
Kedyn/fusliez-notes
https://api.github.com/repos/Kedyn/fusliez-notes
closed
Tree shake the font awesome library
Priority: High Status: Accepted Type: Maintenance
@fortawesome/free-solid-svg-icons/index.es.js is a very big file. Tree shake it down to the handful of icons we use in the project.
1.0
Tree shake the font awesome library - @fortawesome/free-solid-svg-icons/index.es.js is a very big file. Tree shake it down to the handful of icons we use in the project.
priority
tree shake the font awesome library fortawesome free solid svg icons index es js is a very big file tree shake it down to the handful of icons we use in the project
1
44,786
2,914,446,389
IssuesEvent
2015-06-23 05:12:38
CameronProbert/VUW-SWEN303-Assignment4
https://api.github.com/repos/CameronProbert/VUW-SWEN303-Assignment4
closed
Add stats for courts page
TODO High Priority
Add stats as text for overall/per season and with a team selected and without
1.0
Add stats for courts page - Add stats as text for overall/per season and with a team selected and without
priority
add stats for courts page add stats as text for overall per season and with a team selected and without
1
428,102
12,403,100,943
IssuesEvent
2020-05-21 13:19:47
MoonchildProductions/UXP
https://api.github.com/repos/MoonchildProductions/UXP
closed
Fieldset disabled no longer works
Assigned C: DOM C: Layout High Priority Regression
Regression 28.8.4 -> 28.9.0, still broken in 28.9.3 Potentially a regression from the html5 parser work done? Marking as parser component for now. http://jsbin.com/bibiqi/1/edit?html,css,output -- test input fields should all have a red background in the disabled fieldset.
1.0
Fieldset disabled no longer works - Regression 28.8.4 -> 28.9.0, still broken in 28.9.3 Potentially a regression from the html5 parser work done? Marking as parser component for now. http://jsbin.com/bibiqi/1/edit?html,css,output -- test input fields should all have a red background in the disabled fieldset.
priority
fieldset disabled no longer works regression still broken in potentially a regression from the parser work done marking as parser component for now test input fields should all have a red background in the disabled fieldset
1
388,565
11,489,266,531
IssuesEvent
2020-02-11 15:14:55
level73/membernet
https://api.github.com/repos/level73/membernet
closed
MANAGE USERS tab not appearing when hovering over
Platform: Membernet Priority: High Status: Wontfix Type: Bug
Frequently, I am unable to access the manage users tab, it does not load when hovering.
1.0
MANAGE USERS tab not appearing when hovering over - Frequently, I am unable to access the manage users tab, it does not load when hovering.
priority
manage users tab not appearing when hovering over frequently i am unable to access the manage users tab it does not load when hovering
1
110,576
4,434,766,224
IssuesEvent
2016-08-18 04:59:29
OpenSRP/opensrp-client
https://api.github.com/repos/OpenSRP/opensrp-client
opened
Migrate Bidan app drishti codebase to OpenSRP codebase
High Priority INDONESIA
Currently Bidan app using old dristhi codebase.
1.0
Migrate Bidan app drishti codebase to OpenSRP codebase - Currently Bidan app using old dristhi codebase.
priority
migrate bidan app drishti codebase to opensrp codebase currently bidan app using old dristhi codebase
1
238,141
7,775,278,206
IssuesEvent
2018-06-05 01:50:52
ESGF/esgf-compute-wps
https://api.github.com/repos/ESGF/esgf-compute-wps
opened
Expose regrid and domain as a per operation setting in workflow
feature priority/high
**Is your feature request related to a problem? Please describe.** You cannot set the regrid or domain for individual operations in a workflow. **Describe the solution you'd like** When you configure an operation you should be able to set the domain or regrid. **Describe alternatives you've considered** n/a **Additional context** n/a
1.0
Expose regrid and domain as a per operation setting in workflow - **Is your feature request related to a problem? Please describe.** You cannot set the regrid or domain for individual operations in a workflow. **Describe the solution you'd like** When you configure an operation you should be able to set the domain or regrid. **Describe alternatives you've considered** n/a **Additional context** n/a
priority
expose regrid and domain as a per operation setting in workflow is your feature request related to a problem please describe you cannot set the regrid or domain for individual operations in a workflow describe the solution you d like when you configure an operation you should be able to set the domain or regrid describe alternatives you ve considered n a additional context n a
1
336,209
10,173,198,972
IssuesEvent
2019-08-08 12:36:04
ecwinc/pact-server
https://api.github.com/repos/ecwinc/pact-server
closed
Easy way to upload data
high priority
Idea: type in some admin password into browser page and get redirected to "upload" page where one can upload geoJSON files, sending their content to the database. Will be necessary once we have all the new data from Jae (uploading the raw JSON from a script will take a long time)
1.0
Easy way to upload data - Idea: type in some admin password into browser page and get redirected to "upload" page where one can upload geoJSON files, sending their content to the database. Will be necessary once we have all the new data from Jae (uploading the raw JSON from a script will take a long time)
priority
easy way to upload data idea type in some admin password into browser page and get redirected to upload page where one can upload geojson files sending their content to the database will be necessary once we have all the new data from jae uploading the raw json from a script will take a long time
1
175,592
6,552,460,010
IssuesEvent
2017-09-05 18:26:38
vtyulb/BSA-Analytics
https://api.github.com/repos/vtyulb/BSA-Analytics
closed
не работает precise search
High priority
После обновления программы precise search выдает ошибку на любых файлах как на сервере, так и на локальном компьютере
1.0
не работает precise search - После обновления программы precise search выдает ошибку на любых файлах как на сервере, так и на локальном компьютере
priority
не работает precise search после обновления программы precise search выдает ошибку на любых файлах как на сервере так и на локальном компьютере
1
281,839
8,700,292,993
IssuesEvent
2018-12-05 08:15:16
AICrowd/ai-crowd-3
https://api.github.com/repos/AICrowd/ai-crowd-3
closed
500 after middle button clicking on a video in leaderboard
June 2018 bug high priority
_From @AdamStelmaszczyk on July 06, 2017 13:35_ After middle button clicking (which opens a link in a new tab) on any video image on [leaderboard page](https://www.crowdai.org/challenges/nips-2017-learning-to-run/leaderboards), 500 "We're sorry, but something went wrong" page always comes up, [example](https://www.crowdai.org/challenges/nips-2017-learning-to-run/leaderboards/video_modal?submission_id=493). Left button clicking works (because it only opens the modal window, no URL change). _Copied from original issue: crowdAI/crowdai#243_
1.0
500 after middle button clicking on a video in leaderboard - _From @AdamStelmaszczyk on July 06, 2017 13:35_ After middle button clicking (which opens a link in a new tab) on any video image on [leaderboard page](https://www.crowdai.org/challenges/nips-2017-learning-to-run/leaderboards), 500 "We're sorry, but something went wrong" page always comes up, [example](https://www.crowdai.org/challenges/nips-2017-learning-to-run/leaderboards/video_modal?submission_id=493). Left button clicking works (because it only opens the modal window, no URL change). _Copied from original issue: crowdAI/crowdai#243_
priority
after middle button clicking on a video in leaderboard from adamstelmaszczyk on july after middle button clicking which opens a link in a new tab on any video image on we re sorry but something went wrong page always comes up left button clicking works because it only opens the modal window no url change copied from original issue crowdai crowdai
1
745,456
25,985,017,621
IssuesEvent
2022-12-19 22:52:09
gmlc-dispatches/dispatches
https://api.github.com/repos/gmlc-dispatches/dispatches
closed
Test failures in RE DoubleLoopOptimization.ipynb: ValueError: Error evaluating Param value (UnitOnT0State[309_WIND_1])
Priority:High
This occurred with #164, possibly as a consequence of updating to IDAES 2.0.0b2: ``` ValueError: Error evaluating Param value (UnitOnT0State[309_WIND_1]): The Param value is currently set to an invalid value. This is typically from a scalar Param or mutable Indexed Param without an initial or default value. ``` See https://github.com/gmlc-dispatches/dispatches/actions/runs/3678468960/jobs/6221778238#step:5:385 for more details.
1.0
Test failures in RE DoubleLoopOptimization.ipynb: ValueError: Error evaluating Param value (UnitOnT0State[309_WIND_1]) - This occurred with #164, possibly as a consequence of updating to IDAES 2.0.0b2: ``` ValueError: Error evaluating Param value (UnitOnT0State[309_WIND_1]): The Param value is currently set to an invalid value. This is typically from a scalar Param or mutable Indexed Param without an initial or default value. ``` See https://github.com/gmlc-dispatches/dispatches/actions/runs/3678468960/jobs/6221778238#step:5:385 for more details.
priority
test failures in re doubleloopoptimization ipynb valueerror error evaluating param value this occurred with possibly as a consequence of updating to idaes valueerror error evaluating param value the param value is currently set to an invalid value this is typically from a scalar param or mutable indexed param without an initial or default value see for more details
1
701,193
24,090,101,591
IssuesEvent
2022-09-19 14:08:46
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Modal Component: CSS Blur filter causing performance issues
[Type] Bug [Priority] High General Interface [Type] Performance [Feature] Component System
### Description When trying to use any modal with the new blurred design using Gutenberg 14.0.2 and WordPress 6.0.2, there's noticeable lag. This happens in the Site Editor and Post Editor when doing things like: opening pattern explorer, opening template part inserter, and opening preferences. I can't tell if this is related to the new design or something else entirely so let me know if the title needs to be updated :) ### Step-by-step reproduction instructions 1. Open the Site Editor. 2. Open the Inserter, go to the patterns tab, and hit "explore". Notice lag. 3. Close out of that experience. 4. Open preferences from the three dot menu. Notice lag. 5. Close out of that experience. 6. Add a template part block and choose the option to select from a current template. 7. See modal pop up and notice lag. ### Screenshots, screen recording, code snippet https://user-images.githubusercontent.com/26996883/188512881-4e06f49a-be65-4f3c-b8cf-690c612e670c.mov ### Environment info - WordPress 6.0.2 - TT2 and Tove (two block themes) - Gutenberg 14.0.2 ### Please confirm that you have searched existing issues in the repo. Yes ### Please confirm that you have tested with all plugins deactivated except Gutenberg. Yes
1.0
Modal Component: CSS Blur filter causing performance issues - ### Description When trying to use any modal with the new blurred design using Gutenberg 14.0.2 and WordPress 6.0.2, there's noticeable lag. This happens in the Site Editor and Post Editor when doing things like: opening pattern explorer, opening template part inserter, and opening preferences. I can't tell if this is related to the new design or something else entirely so let me know if the title needs to be updated :) ### Step-by-step reproduction instructions 1. Open the Site Editor. 2. Open the Inserter, go to the patterns tab, and hit "explore". Notice lag. 3. Close out of that experience. 4. Open preferences from the three dot menu. Notice lag. 5. Close out of that experience. 6. Add a template part block and choose the option to select from a current template. 7. See modal pop up and notice lag. ### Screenshots, screen recording, code snippet https://user-images.githubusercontent.com/26996883/188512881-4e06f49a-be65-4f3c-b8cf-690c612e670c.mov ### Environment info - WordPress 6.0.2 - TT2 and Tove (two block themes) - Gutenberg 14.0.2 ### Please confirm that you have searched existing issues in the repo. Yes ### Please confirm that you have tested with all plugins deactivated except Gutenberg. Yes
priority
modal component css blur filter causing performance issues description when trying to use any modal with the new blurred design using gutenberg and wordpress there s noticeable lag this happens in the site editor and post editor when doing things like opening pattern explorer opening template part inserter and opening preferences i can t tell if this is related to the new design or something else entirely so let me know if the title needs to be updated step by step reproduction instructions open the site editor open the inserter go to the patterns tab and hit explore notice lag close out of that experience open preferences from the three dot menu notice lag close out of that experience add a template part block and choose the option to select from a current template see modal pop up and notice lag screenshots screen recording code snippet environment info wordpress and tove two block themes gutenberg please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg yes
1
356,856
10,598,338,286
IssuesEvent
2019-10-10 04:26:26
harmony-one/harmony
https://api.github.com/repos/harmony-one/harmony
closed
Joining testnet fails
bug high priority
**Describe the bug** Running a node on AWS to join testnet fails **To Reproduce** Log on to the AWS Insance and run ``` curl -O https://raw.githubusercontent.com/harmony-one/harmony/master/scripts/node.sh chmod +x node.sh sudo ./node.sh ``` **Expected behavior** Node joins succesfully **Environment (please complete the following information):** - AWS EC2 instance **Additional context** Below is the error log ``` INFO [04-03|17:20:36.685] [SYNC] Current Block port=9000 ip=34.222.46.182 blockHex=0x811737e250857a0faaacefcd94e066ef01a8adede7873d8cd3a8104edd291a56 ERROR[04-03|17:20:36.690] cannot create group sig mask: key not found port=9000 ip=34.222.46.182 DEBUG[04-03|17:20:36.697] Trie cache stats after commit misses=0 unloads=0 DEBUG[04-03|17:20:36.697] Inserted new block number=1 hash=6b1e9e…3c63d4 uncles=0 txs=3 gas=1454996 elapsed=11.641ms INFO [04-03|17:20:36.697] Imported new chain segment blocks=1 txs=3 mgas=1.455 elapsed=11.893ms mgasps=122.336 number=1 hash=6b1e9e…3c63d4 age=1d12h51m cache=13.37kB INFO [04-03|17:20:36.697] [SYNC] new block added to blockchain port=9000 ip=34.222.46.182 blockHeight=1 blockHex=0x6b1e9e06ce095a10acfd44fd51accaa7b1d10489c3e30890f8404a30093c63d4 INFO [04-03|17:20:36.698] [SYNC] Current Block port=9000 ip=34.222.46.182 blockHex=0x6b1e9e06ce095a10acfd44fd51accaa7b1d10489c3e30890f8404a30093c63d4 ERROR[04-03|17:20:36.699] cannot create group sig mask: key not found port=9000 ip=34.222.46.182 ERROR[04-03|17:20:36.700] ########## BAD BLOCK ######### Chain config: {ChainID: 0 Homestead: 0 DAO: <nil> DAOSupport: false EIP150: 0 EIP155: 0 EIP158: 0 Byzantium: 0 Constantinople: 0 ConstantinopleFix: 0 Engine: ethash} Number: 2 Hash: 0x10e4c3caa8562e8867adab7bb46c48fcc34b069dc94e98de7466aff188d3e4be &{[] 1 77048 [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0] [] [129 146 152 85 228 159 49 19 1 173 88 245 109 97 14 71 99 244 250 58 146 204 164 231 41 159 233 211 35 78 3 228] [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0] 77048} Error: invalid merkle root (remote: 237abd2e63ae0cf8a2bc1801d3b183d83d599c8d57da3522a7f45339f21e8d46 local: 25dd7137dc20eb62813dc812fb7a900f4eb6f616c174b184a7a2cb6af53dc930) ############################## DEBUG[04-03|17:20:36.700] Error adding new block to blockchain port=9000 ip=34.222.46.182 Error="invalid merkle root (remote: 237abd2e63ae0cf8a2bc1801d3b183d83d599c8d57da3522a7f45339f21e8d46 local: 25dd7137dc20eb62813dc812fb7a900f4eb6f616c174b184a7a2cb6af53dc930)" INFO [04-03|17:20:36.704] [SYNC] clean up inactive peers port=9000 ip=34.222.46.182 activeNumber=7 ```
1.0
Joining testnet fails - **Describe the bug** Running a node on AWS to join testnet fails **To Reproduce** Log on to the AWS Insance and run ``` curl -O https://raw.githubusercontent.com/harmony-one/harmony/master/scripts/node.sh chmod +x node.sh sudo ./node.sh ``` **Expected behavior** Node joins succesfully **Environment (please complete the following information):** - AWS EC2 instance **Additional context** Below is the error log ``` INFO [04-03|17:20:36.685] [SYNC] Current Block port=9000 ip=34.222.46.182 blockHex=0x811737e250857a0faaacefcd94e066ef01a8adede7873d8cd3a8104edd291a56 ERROR[04-03|17:20:36.690] cannot create group sig mask: key not found port=9000 ip=34.222.46.182 DEBUG[04-03|17:20:36.697] Trie cache stats after commit misses=0 unloads=0 DEBUG[04-03|17:20:36.697] Inserted new block number=1 hash=6b1e9e…3c63d4 uncles=0 txs=3 gas=1454996 elapsed=11.641ms INFO [04-03|17:20:36.697] Imported new chain segment blocks=1 txs=3 mgas=1.455 elapsed=11.893ms mgasps=122.336 number=1 hash=6b1e9e…3c63d4 age=1d12h51m cache=13.37kB INFO [04-03|17:20:36.697] [SYNC] new block added to blockchain port=9000 ip=34.222.46.182 blockHeight=1 blockHex=0x6b1e9e06ce095a10acfd44fd51accaa7b1d10489c3e30890f8404a30093c63d4 INFO [04-03|17:20:36.698] [SYNC] Current Block port=9000 ip=34.222.46.182 blockHex=0x6b1e9e06ce095a10acfd44fd51accaa7b1d10489c3e30890f8404a30093c63d4 ERROR[04-03|17:20:36.699] cannot create group sig mask: key not found port=9000 ip=34.222.46.182 ERROR[04-03|17:20:36.700] ########## BAD BLOCK ######### Chain config: {ChainID: 0 Homestead: 0 DAO: <nil> DAOSupport: false EIP150: 0 EIP155: 0 EIP158: 0 Byzantium: 0 Constantinople: 0 ConstantinopleFix: 0 Engine: ethash} Number: 2 Hash: 0x10e4c3caa8562e8867adab7bb46c48fcc34b069dc94e98de7466aff188d3e4be &{[] 1 77048 [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0] [] [129 146 152 85 228 159 49 19 1 173 88 245 109 97 14 71 99 244 250 58 146 204 164 231 41 159 233 211 35 78 3 228] [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0] 77048} Error: invalid merkle root (remote: 237abd2e63ae0cf8a2bc1801d3b183d83d599c8d57da3522a7f45339f21e8d46 local: 25dd7137dc20eb62813dc812fb7a900f4eb6f616c174b184a7a2cb6af53dc930) ############################## DEBUG[04-03|17:20:36.700] Error adding new block to blockchain port=9000 ip=34.222.46.182 Error="invalid merkle root (remote: 237abd2e63ae0cf8a2bc1801d3b183d83d599c8d57da3522a7f45339f21e8d46 local: 25dd7137dc20eb62813dc812fb7a900f4eb6f616c174b184a7a2cb6af53dc930)" INFO [04-03|17:20:36.704] [SYNC] clean up inactive peers port=9000 ip=34.222.46.182 activeNumber=7 ```
priority
joining testnet fails describe the bug running a node on aws to join testnet fails to reproduce log on to the aws insance and run curl o chmod x node sh sudo node sh expected behavior node joins succesfully environment please complete the following information aws instance additional context below is the error log info current block port ip blockhex error cannot create group sig mask key not found port ip debug trie cache stats after commit misses unloads debug inserted new block number hash … uncles txs gas elapsed info imported new chain segment blocks txs mgas elapsed mgasps number hash … age cache info new block added to blockchain port ip blockheight blockhex info current block port ip blockhex error cannot create group sig mask key not found port ip error bad block chain config chainid homestead dao daosupport false byzantium constantinople constantinoplefix engine ethash number hash error invalid merkle root remote local debug error adding new block to blockchain port ip error invalid merkle root remote local info clean up inactive peers port ip activenumber
1
519,040
15,039,098,965
IssuesEvent
2021-02-02 18:13:46
qutebrowser/qutebrowser
https://api.github.com/repos/qutebrowser/qutebrowser
closed
content.cookies.store=false stopped working with 2.0
priority: 0 - high status: needs triage
**Version info**: qutebrowser v2.0.1 Git commit: Backend: QtWebEngine (Chromium 83.0.4103.122) Qt: 5.15.2 **Does the bug happen if you start with `--temp-basedir`?**: Needs a session. **Description** Cookies are stored with `content.cookies.store` set to `false`. **How to reproduce** Set `content.cookies.store` to false, login some, e.g. GitHub. Restart. Still logged in.
1.0
content.cookies.store=false stopped working with 2.0 - **Version info**: qutebrowser v2.0.1 Git commit: Backend: QtWebEngine (Chromium 83.0.4103.122) Qt: 5.15.2 **Does the bug happen if you start with `--temp-basedir`?**: Needs a session. **Description** Cookies are stored with `content.cookies.store` set to `false`. **How to reproduce** Set `content.cookies.store` to false, login some, e.g. GitHub. Restart. Still logged in.
priority
content cookies store false stopped working with version info qutebrowser git commit backend qtwebengine chromium qt does the bug happen if you start with temp basedir needs a session description cookies are stored with content cookies store set to false how to reproduce set content cookies store to false login some e g github restart still logged in
1
381,802
11,288,245,649
IssuesEvent
2020-01-16 07:21:26
cvfadmin/COVE
https://api.github.com/repos/cvfadmin/COVE
closed
Pagination breaks when user zooms in our out of page.
bug high priority
cmd +/cmd - **Proposed Solution: Switch to per page pagination instead of infinite scroll**
1.0
Pagination breaks when user zooms in our out of page. - cmd +/cmd - **Proposed Solution: Switch to per page pagination instead of infinite scroll**
priority
pagination breaks when user zooms in our out of page cmd cmd proposed solution switch to per page pagination instead of infinite scroll
1
161,153
6,110,016,318
IssuesEvent
2017-06-21 14:15:42
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] Delete of a published page not working
bug Priority: Highest!
Using Editorial bp: 1. Create a new temp landing page (e.g. Health 2). 2. Publish the new page. 3. Delete the page. Result: the page is never deleted. Same thing happens if you edit and existing page, publish and then try to delete it.
1.0
[studio] Delete of a published page not working - Using Editorial bp: 1. Create a new temp landing page (e.g. Health 2). 2. Publish the new page. 3. Delete the page. Result: the page is never deleted. Same thing happens if you edit and existing page, publish and then try to delete it.
priority
delete of a published page not working using editorial bp create a new temp landing page e g health publish the new page delete the page result the page is never deleted same thing happens if you edit and existing page publish and then try to delete it
1
364,656
10,771,904,273
IssuesEvent
2019-11-02 11:14:08
AY1920S1-CS2103T-T11-1/main
https://api.github.com/repos/AY1920S1-CS2103T-T11-1/main
closed
Change how Price field is parsed
priority.High
Make Price field not accept 0 Improve error message for negative price values Block more than 2d.p. for price inputs Related to #212 #170
1.0
Change how Price field is parsed - Make Price field not accept 0 Improve error message for negative price values Block more than 2d.p. for price inputs Related to #212 #170
priority
change how price field is parsed make price field not accept improve error message for negative price values block more than p for price inputs related to
1
522,141
15,158,052,413
IssuesEvent
2021-02-12 00:13:26
NOAA-GSL/MATS
https://api.github.com/repos/NOAA-GSL/MATS
opened
Add plot timeline button/capability to MATS timeseries
Priority: High Project: MATS Type: Feature
--- Author Name: **molly.b.smith** (@mollybsmith-noaa) Original Redmine Issue: 56360, https://vlab.ncep.noaa.gov/redmine/issues/56360 Original Date: 2018-10-16 Original Assignee: randy.pierce --- We would like to add the capability to show what times have data for a particular set of parameters in MATS.
1.0
Add plot timeline button/capability to MATS timeseries - --- Author Name: **molly.b.smith** (@mollybsmith-noaa) Original Redmine Issue: 56360, https://vlab.ncep.noaa.gov/redmine/issues/56360 Original Date: 2018-10-16 Original Assignee: randy.pierce --- We would like to add the capability to show what times have data for a particular set of parameters in MATS.
priority
add plot timeline button capability to mats timeseries author name molly b smith mollybsmith noaa original redmine issue original date original assignee randy pierce we would like to add the capability to show what times have data for a particular set of parameters in mats
1
627,920
19,956,823,230
IssuesEvent
2022-01-28 00:54:24
BlockySurvival/issue-tracker
https://api.github.com/repos/BlockySurvival/issue-tracker
closed
players can create new accounts with empty passwords
bug high priority security
empty passwords are a disaster. the setting is correct, so this is one more thing multicraft is ignoring for unknown reasons.
1.0
players can create new accounts with empty passwords - empty passwords are a disaster. the setting is correct, so this is one more thing multicraft is ignoring for unknown reasons.
priority
players can create new accounts with empty passwords empty passwords are a disaster the setting is correct so this is one more thing multicraft is ignoring for unknown reasons
1
416,717
12,150,817,387
IssuesEvent
2020-04-24 18:42:00
woocommerce/woocommerce-android
https://api.github.com/repos/woocommerce/woocommerce-android
closed
4.1-rc-1: Crash in login discovery view
Priority: High [Type] Crash
Sentry Issue: [WOOCOMMERCE-ANDROID-GY](https://sentry.io/organizations/a8c/issues/1627191543/?referrer=github_integration) ``` ClassCastException: com.google.android.material.appbar.AppBarLayout cannot be cast to androidx.appcompat.widget.Toolbar at com.woocommerce.android.ui.login.LoginDiscoveryErrorFragment.onViewCreated(LoginDiscoveryErrorFragment.kt:90) ... (14 additional frame(s) were not displayed) com.google.android.material.appbar.AppBarLayout cannot be cast to androidx.appcompat.widget.Toolbar ```
1.0
4.1-rc-1: Crash in login discovery view - Sentry Issue: [WOOCOMMERCE-ANDROID-GY](https://sentry.io/organizations/a8c/issues/1627191543/?referrer=github_integration) ``` ClassCastException: com.google.android.material.appbar.AppBarLayout cannot be cast to androidx.appcompat.widget.Toolbar at com.woocommerce.android.ui.login.LoginDiscoveryErrorFragment.onViewCreated(LoginDiscoveryErrorFragment.kt:90) ... (14 additional frame(s) were not displayed) com.google.android.material.appbar.AppBarLayout cannot be cast to androidx.appcompat.widget.Toolbar ```
priority
rc crash in login discovery view sentry issue classcastexception com google android material appbar appbarlayout cannot be cast to androidx appcompat widget toolbar at com woocommerce android ui login logindiscoveryerrorfragment onviewcreated logindiscoveryerrorfragment kt additional frame s were not displayed com google android material appbar appbarlayout cannot be cast to androidx appcompat widget toolbar
1
291,516
8,926,230,710
IssuesEvent
2019-01-22 03:08:34
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Master: "plan research" tutorial broken
High Priority
1) checking "show all" not counts, maybe refers to https://github.com/StrangeLoopGames/EcoIssues/issues/10702 2) "mouse over" - no tooltip, no complete :( ![default](https://user-images.githubusercontent.com/4980243/51320395-b2cdc580-1a70-11e9-9410-8851d1e60554.png) ![default](https://user-images.githubusercontent.com/4980243/51320511-266fd280-1a71-11e9-9e97-5db7642de520.png)
1.0
Master: "plan research" tutorial broken - 1) checking "show all" not counts, maybe refers to https://github.com/StrangeLoopGames/EcoIssues/issues/10702 2) "mouse over" - no tooltip, no complete :( ![default](https://user-images.githubusercontent.com/4980243/51320395-b2cdc580-1a70-11e9-9410-8851d1e60554.png) ![default](https://user-images.githubusercontent.com/4980243/51320511-266fd280-1a71-11e9-9e97-5db7642de520.png)
priority
master plan research tutorial broken checking show all not counts maybe refers to mouse over no tooltip no complete
1
145,340
5,565,128,486
IssuesEvent
2017-03-26 11:25:32
metasfresh/metasfresh
https://api.github.com/repos/metasfresh/metasfresh
opened
Manufacturing Order UI
enhancement high priority
### Is this a bug or feature request? Feature Request ### What is the current behavior? Currently, there is no Manufacturing Order UI Window in metasfresh. ### What is the expected or desired behavior? Create a generic window w/ Grid View that lists all nonprocessed Manufacturing Orders. This Window is the starting point for https://github.com/metasfresh/metasfresh-webui-frontend/issues/580.
1.0
Manufacturing Order UI - ### Is this a bug or feature request? Feature Request ### What is the current behavior? Currently, there is no Manufacturing Order UI Window in metasfresh. ### What is the expected or desired behavior? Create a generic window w/ Grid View that lists all nonprocessed Manufacturing Orders. This Window is the starting point for https://github.com/metasfresh/metasfresh-webui-frontend/issues/580.
priority
manufacturing order ui is this a bug or feature request feature request what is the current behavior currently there is no manufacturing order ui window in metasfresh what is the expected or desired behavior create a generic window w grid view that lists all nonprocessed manufacturing orders this window is the starting point for
1
258,032
8,150,301,950
IssuesEvent
2018-08-22 12:34:51
metasfresh/metasfresh
https://api.github.com/repos/metasfresh/metasfresh
closed
Support Doctype Text templates also on purchase order
branch:master priority:high type:enhancement
### Is this a bug or feature request? FR ### What is the current behavior? metasfresh supports putting header and footer text tempates into sales order and invoice http://docs.metasfresh.org/webui_collection/EN/Print_text_on_documents-doctype.html if you try this for purchase order the text template will be reapplied on completion. case: 1. configure header and footer text for doctype purchase order 1. create a new purchase order with lines 1. OK: text is copied to fields in purchase order 1. edit the text and put something else 1. OK: text is displayed 1. complete the order 1. NOK: text is set back to default ### What is the expected or desired behavior? we do want to support that also in purchase order todo: - change that behaviour - create an automatic test
1.0
Support Doctype Text templates also on purchase order - ### Is this a bug or feature request? FR ### What is the current behavior? metasfresh supports putting header and footer text tempates into sales order and invoice http://docs.metasfresh.org/webui_collection/EN/Print_text_on_documents-doctype.html if you try this for purchase order the text template will be reapplied on completion. case: 1. configure header and footer text for doctype purchase order 1. create a new purchase order with lines 1. OK: text is copied to fields in purchase order 1. edit the text and put something else 1. OK: text is displayed 1. complete the order 1. NOK: text is set back to default ### What is the expected or desired behavior? we do want to support that also in purchase order todo: - change that behaviour - create an automatic test
priority
support doctype text templates also on purchase order is this a bug or feature request fr what is the current behavior metasfresh supports putting header and footer text tempates into sales order and invoice if you try this for purchase order the text template will be reapplied on completion case configure header and footer text for doctype purchase order create a new purchase order with lines ok text is copied to fields in purchase order edit the text and put something else ok text is displayed complete the order nok text is set back to default what is the expected or desired behavior we do want to support that also in purchase order todo change that behaviour create an automatic test
1
194,985
6,901,413,083
IssuesEvent
2017-11-25 06:55:30
metasfresh/metasfresh-webui-frontend
https://api.github.com/repos/metasfresh/metasfresh-webui-frontend
closed
Navigation Menu Favorite stars wrong
branch:master priority:high type:bug
### Is this a bug or feature request? Bug ### What is the current behavior? The Bookmark Favorites in Navigation menu are behaving weirdly. #### Which are the steps to reproduce? 1. Open the Navigation Menu and hover over an entry. Move to grey star to bookmark it. Star disappears. NOK 2. Move over an entry until you see a dark grey star and move out to the right. A light grey star remains. NOK 3. Open a Window, then alt+1. Next to window name try to bookmark the window there. Does not work. NOK. ### What is the expected or desired behavior? Please fix. ![image](https://user-images.githubusercontent.com/13365687/33131930-623e13be-cf98-11e7-8fba-2117f9fcb630.png) ![image](https://user-images.githubusercontent.com/13365687/33131955-75f9a864-cf98-11e7-891e-bf028b1c2c82.png)
1.0
Navigation Menu Favorite stars wrong - ### Is this a bug or feature request? Bug ### What is the current behavior? The Bookmark Favorites in Navigation menu are behaving weirdly. #### Which are the steps to reproduce? 1. Open the Navigation Menu and hover over an entry. Move to grey star to bookmark it. Star disappears. NOK 2. Move over an entry until you see a dark grey star and move out to the right. A light grey star remains. NOK 3. Open a Window, then alt+1. Next to window name try to bookmark the window there. Does not work. NOK. ### What is the expected or desired behavior? Please fix. ![image](https://user-images.githubusercontent.com/13365687/33131930-623e13be-cf98-11e7-8fba-2117f9fcb630.png) ![image](https://user-images.githubusercontent.com/13365687/33131955-75f9a864-cf98-11e7-891e-bf028b1c2c82.png)
priority
navigation menu favorite stars wrong is this a bug or feature request bug what is the current behavior the bookmark favorites in navigation menu are behaving weirdly which are the steps to reproduce open the navigation menu and hover over an entry move to grey star to bookmark it star disappears nok move over an entry until you see a dark grey star and move out to the right a light grey star remains nok open a window then alt next to window name try to bookmark the window there does not work nok what is the expected or desired behavior please fix
1
417,814
12,179,634,853
IssuesEvent
2020-04-28 11:02:06
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
closed
OG troops on religion change
:beetle: bug :beetle: :exclamation: priority high
<!-- **DO NOT REMOVE PRE-EXISTING LINES** ------------------------------------------------------------------------------------------------------------ --> **Your mod version is:** 78a3f6f458e30148cfec06e599d14843de3f69aa **What expansions do you have installed?** All **Are you using any submods/mods? If so, which?** No **Please explain your issue in as much detail as possible:** Reported by @KoiKarp > you should lose armies and generals requested from old gods if you convert, atm you keep them after coverting. **Steps to reproduce the issue:** Get army from old gods, convert to other religion. **Upload an attachment below: .zip of your save, or screenshots:**
1.0
OG troops on religion change - <!-- **DO NOT REMOVE PRE-EXISTING LINES** ------------------------------------------------------------------------------------------------------------ --> **Your mod version is:** 78a3f6f458e30148cfec06e599d14843de3f69aa **What expansions do you have installed?** All **Are you using any submods/mods? If so, which?** No **Please explain your issue in as much detail as possible:** Reported by @KoiKarp > you should lose armies and generals requested from old gods if you convert, atm you keep them after coverting. **Steps to reproduce the issue:** Get army from old gods, convert to other religion. **Upload an attachment below: .zip of your save, or screenshots:**
priority
og troops on religion change do not remove pre existing lines your mod version is what expansions do you have installed all are you using any submods mods if so which no please explain your issue in as much detail as possible reported by koikarp you should lose armies and generals requested from old gods if you convert atm you keep them after coverting steps to reproduce the issue get army from old gods convert to other religion upload an attachment below zip of your save or screenshots
1
154,621
5,923,178,999
IssuesEvent
2017-05-23 07:09:32
openxc/openxc-ios-framework
https://api.github.com/repos/openxc/openxc-ios-framework
closed
Command support
Priority:High Status:In-progress Type:Enhancement
framework currently only supports "device_id" and "version" commands. Need to add the remaining from https://github.com/openxc/openxc-message-format/blob/master/JSON.mkd#commands. Some of these may come along with protobuf support.
1.0
Command support - framework currently only supports "device_id" and "version" commands. Need to add the remaining from https://github.com/openxc/openxc-message-format/blob/master/JSON.mkd#commands. Some of these may come along with protobuf support.
priority
command support framework currently only supports device id and version commands need to add the remaining from some of these may come along with protobuf support
1
504,812
14,621,309,770
IssuesEvent
2020-12-22 21:23:25
nfdi4plants/Swate
https://api.github.com/repos/nfdi4plants/Swate
closed
Provide validation information via XML metadata
Area: Excel Priority: High Type: Enhancement Type: Interop
It is possible to set metadata of cells via office.js. Especially for users that create templates for others to use, information on the proposed cell contents is helpful. This would be set via a JSON string ~~in the XML tag~~ with the key "validationParams" and should have the following fields: - **importance** - an integer between 0 and 100 - **content type** - one of : ontologyTerm, text, number, int, decimal, url, boolean - if the content type is ontologyTerm, asdd the isA restriction if set: term[restriction] - [x] Create a JSON schema for this - [x] Create a form to generate this for a column _Edit: Office.js only supports custom tags, not custom attributes on existing tags. Therefore, we need to investigate alternative places to save this information. For now we can add the validation table to the app model and later use that data to generate the file/tags/etc we need._ _Edit2(by @Freymaurer): As of the latest decision all swate validation data will now be saved in the excel custom xml, as desribed [here](https://github.com/nfdi4plants/Swate/issues/45#issuecomment-733794910)._ - [x] Add validation data to the app model. - [x] Swate can write validation data into excel custom xml.
1.0
Provide validation information via XML metadata - It is possible to set metadata of cells via office.js. Especially for users that create templates for others to use, information on the proposed cell contents is helpful. This would be set via a JSON string ~~in the XML tag~~ with the key "validationParams" and should have the following fields: - **importance** - an integer between 0 and 100 - **content type** - one of : ontologyTerm, text, number, int, decimal, url, boolean - if the content type is ontologyTerm, asdd the isA restriction if set: term[restriction] - [x] Create a JSON schema for this - [x] Create a form to generate this for a column _Edit: Office.js only supports custom tags, not custom attributes on existing tags. Therefore, we need to investigate alternative places to save this information. For now we can add the validation table to the app model and later use that data to generate the file/tags/etc we need._ _Edit2(by @Freymaurer): As of the latest decision all swate validation data will now be saved in the excel custom xml, as desribed [here](https://github.com/nfdi4plants/Swate/issues/45#issuecomment-733794910)._ - [x] Add validation data to the app model. - [x] Swate can write validation data into excel custom xml.
priority
provide validation information via xml metadata it is possible to set metadata of cells via office js especially for users that create templates for others to use information on the proposed cell contents is helpful this would be set via a json string in the xml tag with the key validationparams and should have the following fields importance an integer between and content type one of ontologyterm text number int decimal url boolean if the content type is ontologyterm asdd the isa restriction if set term create a json schema for this create a form to generate this for a column edit office js only supports custom tags not custom attributes on existing tags therefore we need to investigate alternative places to save this information for now we can add the validation table to the app model and later use that data to generate the file tags etc we need by freymaurer as of the latest decision all swate validation data will now be saved in the excel custom xml as desribed add validation data to the app model swate can write validation data into excel custom xml
1
439,950
12,690,975,678
IssuesEvent
2020-06-21 14:50:10
sodafoundation/SIM
https://api.github.com/repos/sodafoundation/SIM
closed
Create Virtual environment for deploying Dolphin
Feature High Priority
<!-- This form is for bug reports and feature requests! --> **Is this a BUG REPORT or FEATURE REQUEST?**: / feature **What happened**: Need to run dolphin in virtual environment as to to keep dependencies required by project by creating isolated python virtual environments. **What you expected to happen**: Dolphin should work independtly even if there are other pyhton projects which uses ddifferent modules. **How to reproduce it (as minimally and precisely as possible)**: NA **Anything else we need to know?**: NA **Environment**: - Dolphin(release/branch) version: - OS (e.g. from /etc/os-release): - Kernel (e.g. `uname -a`): - Install tools: - Others:
1.0
Create Virtual environment for deploying Dolphin - <!-- This form is for bug reports and feature requests! --> **Is this a BUG REPORT or FEATURE REQUEST?**: / feature **What happened**: Need to run dolphin in virtual environment as to to keep dependencies required by project by creating isolated python virtual environments. **What you expected to happen**: Dolphin should work independtly even if there are other pyhton projects which uses ddifferent modules. **How to reproduce it (as minimally and precisely as possible)**: NA **Anything else we need to know?**: NA **Environment**: - Dolphin(release/branch) version: - OS (e.g. from /etc/os-release): - Kernel (e.g. `uname -a`): - Install tools: - Others:
priority
create virtual environment for deploying dolphin is this a bug report or feature request feature what happened need to run dolphin in virtual environment as to to keep dependencies required by project by creating isolated python virtual environments what you expected to happen dolphin should work independtly even if there are other pyhton projects which uses ddifferent modules how to reproduce it as minimally and precisely as possible na anything else we need to know na environment dolphin release branch version os e g from etc os release kernel e g uname a install tools others
1
551,234
16,165,221,570
IssuesEvent
2021-05-01 10:51:06
sopra-fs21-group-01/server
https://api.github.com/repos/sopra-fs21-group-01/server
closed
Implement Exceptions
high priority
## acceptance criteria: -no default exceptions occurs during login/registration ## estimated time: 1h
1.0
Implement Exceptions - ## acceptance criteria: -no default exceptions occurs during login/registration ## estimated time: 1h
priority
implement exceptions acceptance criteria no default exceptions occurs during login registration estimated time
1
148,448
5,682,304,805
IssuesEvent
2017-04-13 09:20:14
openxpki/openxpki
https://api.github.com/repos/openxpki/openxpki
closed
Add useful alias name to chain certificates in Java KeyStore
feature request high priority
At least some versions of keytool add the ISSUER name as alias to a chain certificate which leads to confusion. Proposed solution is to add the subject (CN or full DN) to all certificates in the store.
1.0
Add useful alias name to chain certificates in Java KeyStore - At least some versions of keytool add the ISSUER name as alias to a chain certificate which leads to confusion. Proposed solution is to add the subject (CN or full DN) to all certificates in the store.
priority
add useful alias name to chain certificates in java keystore at least some versions of keytool add the issuer name as alias to a chain certificate which leads to confusion proposed solution is to add the subject cn or full dn to all certificates in the store
1
633,644
20,261,129,441
IssuesEvent
2022-02-15 07:33:46
rtCamp/rtMedia
https://api.github.com/repos/rtCamp/rtMedia
closed
Blank media page is visible in Twenty Twenty-Two theme
bug Priority/High
``` Environment : Tested on pre-release Branch WordPress Version: 5.9 BuddyPress Version: 10.0.0 rtMedia Version: 4.6.9 theme - Twenty Twenty-Two ``` The blank media page is displayed in the new theme. But on others theme page is rendered properly. Note: no console error and warning are displayed. Issue observed on both legacy and Nouveau template https://user-images.githubusercontent.com/44057535/151926551-1beac4ed-61f5-4adc-9054-960985334c92.mp4
1.0
Blank media page is visible in Twenty Twenty-Two theme - ``` Environment : Tested on pre-release Branch WordPress Version: 5.9 BuddyPress Version: 10.0.0 rtMedia Version: 4.6.9 theme - Twenty Twenty-Two ``` The blank media page is displayed in the new theme. But on others theme page is rendered properly. Note: no console error and warning are displayed. Issue observed on both legacy and Nouveau template https://user-images.githubusercontent.com/44057535/151926551-1beac4ed-61f5-4adc-9054-960985334c92.mp4
priority
blank media page is visible in twenty twenty two theme environment tested on pre release branch wordpress version buddypress version rtmedia version theme twenty twenty two the blank media page is displayed in the new theme but on others theme page is rendered properly note no console error and warning are displayed issue observed on both legacy and nouveau template
1
719,686
24,767,036,250
IssuesEvent
2022-10-22 17:05:48
ufosc/Club_Website_2
https://api.github.com/repos/ufosc/Club_Website_2
closed
Set up EJS linting
good first issue help wanted high priority difficulty: easy
**Todo:** Set up an EJS linter. This is an easy way to automatically style all our html/EJS code and provide more meaningful error messages for EJS (the ejs module is designed to be minimal so they don't serve detailed errors without a linter). Here is the module: https://www.npmjs.com/package/ejs-lint You'll then need to add the linting and testing commands to `package.json`: ```json "scripts": { "start": "nodemon app.js", "test": "standard", "lint": "standard && YOUR_TEST_CMD_HERE" , "fix": "standard --fix && YOUR_AUTO_LINT_CMD_HERE" }, ```
1.0
Set up EJS linting - **Todo:** Set up an EJS linter. This is an easy way to automatically style all our html/EJS code and provide more meaningful error messages for EJS (the ejs module is designed to be minimal so they don't serve detailed errors without a linter). Here is the module: https://www.npmjs.com/package/ejs-lint You'll then need to add the linting and testing commands to `package.json`: ```json "scripts": { "start": "nodemon app.js", "test": "standard", "lint": "standard && YOUR_TEST_CMD_HERE" , "fix": "standard --fix && YOUR_AUTO_LINT_CMD_HERE" }, ```
priority
set up ejs linting todo set up an ejs linter this is an easy way to automatically style all our html ejs code and provide more meaningful error messages for ejs the ejs module is designed to be minimal so they don t serve detailed errors without a linter here is the module you ll then need to add the linting and testing commands to package json json scripts start nodemon app js test standard lint standard your test cmd here fix standard fix your auto lint cmd here
1
564,275
16,722,372,859
IssuesEvent
2021-06-10 08:52:50
GluuFederation/community-edition-setup
https://api.github.com/repos/GluuFederation/community-edition-setup
closed
Import o_metric.ldif after o=backend is created
High Priority bug exclude from release notes
The template specify ou=statistic,o=metric but there’s no parent o=metric entry there. The import sequence is what matters here **o_metric.ldif** (contains o=metric) -> **base.ldif** Reported by @iromli
1.0
Import o_metric.ldif after o=backend is created - The template specify ou=statistic,o=metric but there’s no parent o=metric entry there. The import sequence is what matters here **o_metric.ldif** (contains o=metric) -> **base.ldif** Reported by @iromli
priority
import o metric ldif after o backend is created the template specify ou statistic o metric but there’s no parent o metric entry there the import sequence is what matters here o metric ldif contains o metric base ldif reported by iromli
1
83,359
3,634,000,962
IssuesEvent
2016-02-11 16:29:23
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
TDS translations incorrectly assign the "divider" tag for divided roads.
Category: Core Priority: High Status: In Progress Type: Bug
Hoot uses the "divider=yes" tag to decide if a way needs to be split into two separate ways prior to conflation. The SEP attribute in MGCP and TDS is used to indicate that a road is divided. With MGCP, a single piece of geometry can be tagged with SEP = True to indicate that it is a divided road. In TDS, the extraction spec states that ''' The opposing travelled ways of a divided highway shall be extracted as separate Road features. Horizontal separation is what classifies a Road as a divided highway. There is no minimum separation criterion for a divided highway. Each of the Road delineations shall be attributed as 'Divided' is 'True' (SEP = 1001) [i.e., "Divided"]. If a divided highway also shows a vertical barrier, each of the road delineations shall be attributed as 'Median Present' is 'True' (MES = 1001) [i.e., "With Median"]. ''' For TDS, the SEP attribute should not be translated to the "divider" tag as this will cause Hoot to split an already separated road.
1.0
TDS translations incorrectly assign the "divider" tag for divided roads. - Hoot uses the "divider=yes" tag to decide if a way needs to be split into two separate ways prior to conflation. The SEP attribute in MGCP and TDS is used to indicate that a road is divided. With MGCP, a single piece of geometry can be tagged with SEP = True to indicate that it is a divided road. In TDS, the extraction spec states that ''' The opposing travelled ways of a divided highway shall be extracted as separate Road features. Horizontal separation is what classifies a Road as a divided highway. There is no minimum separation criterion for a divided highway. Each of the Road delineations shall be attributed as 'Divided' is 'True' (SEP = 1001) [i.e., "Divided"]. If a divided highway also shows a vertical barrier, each of the road delineations shall be attributed as 'Median Present' is 'True' (MES = 1001) [i.e., "With Median"]. ''' For TDS, the SEP attribute should not be translated to the "divider" tag as this will cause Hoot to split an already separated road.
priority
tds translations incorrectly assign the divider tag for divided roads hoot uses the divider yes tag to decide if a way needs to be split into two separate ways prior to conflation the sep attribute in mgcp and tds is used to indicate that a road is divided with mgcp a single piece of geometry can be tagged with sep true to indicate that it is a divided road in tds the extraction spec states that the opposing travelled ways of a divided highway shall be extracted as separate road features horizontal separation is what classifies a road as a divided highway there is no minimum separation criterion for a divided highway each of the road delineations shall be attributed as divided is true sep if a divided highway also shows a vertical barrier each of the road delineations shall be attributed as median present is true mes for tds the sep attribute should not be translated to the divider tag as this will cause hoot to split an already separated road
1
195,242
6,905,757,759
IssuesEvent
2017-11-27 08:44:01
sevin7676/SurvivalArenaTracking
https://api.github.com/repos/sevin7676/SurvivalArenaTracking
opened
Tasks are often far too difficult to complete
Priority#1: High Status#3: In Progress Type#3: Bug
Tasks are often far too difficult to complete or require an unreasonable amount of time for the reward provided. (reported by many league 5 players) Reported 10/16/2017 Dev response: (multiple responses - said they are looking into it)
1.0
Tasks are often far too difficult to complete - Tasks are often far too difficult to complete or require an unreasonable amount of time for the reward provided. (reported by many league 5 players) Reported 10/16/2017 Dev response: (multiple responses - said they are looking into it)
priority
tasks are often far too difficult to complete tasks are often far too difficult to complete or require an unreasonable amount of time for the reward provided reported by many league players reported dev response multiple responses said they are looking into it
1
268,766
8,411,452,218
IssuesEvent
2018-10-12 13:56:40
CS2113-AY1819S1-W12-2/main
https://api.github.com/repos/CS2113-AY1819S1-W12-2/main
opened
As a user I can download the notes of a specific module in one command
priority.high type.enhancement
This is to avoid downloading unnecessary notes.
1.0
As a user I can download the notes of a specific module in one command - This is to avoid downloading unnecessary notes.
priority
as a user i can download the notes of a specific module in one command this is to avoid downloading unnecessary notes
1
315,696
9,631,087,979
IssuesEvent
2019-05-15 13:32:48
code4romania/monitorizare-vot-android
https://api.github.com/repos/code4romania/monitorizare-vot-android
closed
Update questions data model - add form id
android bug enhancement good first issue help wanted high priority
Questions should store form id & question id. Swagger for API can be found [here](https://mv-mobile-prod.azurewebsites.net/swagger/ui/index.html). Best practices - codebase language should be English - we :two_hearts: unit tests
1.0
Update questions data model - add form id - Questions should store form id & question id. Swagger for API can be found [here](https://mv-mobile-prod.azurewebsites.net/swagger/ui/index.html). Best practices - codebase language should be English - we :two_hearts: unit tests
priority
update questions data model add form id questions should store form id question id swagger for api can be found best practices codebase language should be english we two hearts unit tests
1
719,564
24,764,109,189
IssuesEvent
2022-10-22 09:27:20
devvsakib/hacktoberfest-react-project
https://api.github.com/repos/devvsakib/hacktoberfest-react-project
closed
[BRANCH - "convert-react"] DELETE PROJECTS FOLDER
enhancement hacktoberfest hacktoberfest-accepted [priority: high]
**Navigate to folder** Check 2nd Branch, name- convert-react Clone this, navigate to main-website folder,this is our main website we are converting into react website, ## Feature? No ## Bug? No Task: Delete PROJECTS folder. we will add projects later, though we have "main" branch with "PROJECTS" folder so we don't need to have it in our "convert-react" branch. Please be aware, and make pr to convert-react branch You can join our discord server - let's build together [Join](https://discord.gg/xwRbDwnbMr)
1.0
[BRANCH - "convert-react"] DELETE PROJECTS FOLDER - **Navigate to folder** Check 2nd Branch, name- convert-react Clone this, navigate to main-website folder,this is our main website we are converting into react website, ## Feature? No ## Bug? No Task: Delete PROJECTS folder. we will add projects later, though we have "main" branch with "PROJECTS" folder so we don't need to have it in our "convert-react" branch. Please be aware, and make pr to convert-react branch You can join our discord server - let's build together [Join](https://discord.gg/xwRbDwnbMr)
priority
delete projects folder navigate to folder check branch name convert react clone this navigate to main website folder this is our main website we are converting into react website feature no bug no task delete projects folder we will add projects later though we have main branch with projects folder so we don t need to have it in our convert react branch please be aware and make pr to convert react branch you can join our discord server let s build together
1
63,369
3,194,713,668
IssuesEvent
2015-09-30 13:38:04
fusioninventory/fusioninventory-for-glpi
https://api.github.com/repos/fusioninventory/fusioninventory-for-glpi
closed
Missing update process to split Tasks with multiple TaskJobs
Category: Tasks Component: For junior contributor Component: Found in version Priority: High Status: Closed Tracker: Bug
--- Author Name: **Kevin Roy** (@kiniou) Original Redmine Issue: 2018, http://forge.fusioninventory.org/issues/2018 Original Date: 2013-03-20 Original Assignee: David Durieux --- Since multijobs task are abandonned, taskjobs should be splitted into one task per job.
1.0
Missing update process to split Tasks with multiple TaskJobs - --- Author Name: **Kevin Roy** (@kiniou) Original Redmine Issue: 2018, http://forge.fusioninventory.org/issues/2018 Original Date: 2013-03-20 Original Assignee: David Durieux --- Since multijobs task are abandonned, taskjobs should be splitted into one task per job.
priority
missing update process to split tasks with multiple taskjobs author name kevin roy kiniou original redmine issue original date original assignee david durieux since multijobs task are abandonned taskjobs should be splitted into one task per job
1
354,725
10,571,432,535
IssuesEvent
2019-10-07 07:06:43
AY1920S1-CS2103-T11-2/main
https://api.github.com/repos/AY1920S1-CS2103-T11-2/main
closed
Initial code refactor
priority.High status.Ongoing type.Task
Refactor code to prepare for MAMS implementation, Rename all relevant classes.
1.0
Initial code refactor - Refactor code to prepare for MAMS implementation, Rename all relevant classes.
priority
initial code refactor refactor code to prepare for mams implementation rename all relevant classes
1
116,482
4,702,775,375
IssuesEvent
2016-10-13 04:17:45
CS2103AUG2016-T14-C4/main
https://api.github.com/repos/CS2103AUG2016-T14-C4/main
opened
As a user, I can type commands without a certain order in arguments
priority.high type.task
so that I only need to remember the flags for each command and not the order
1.0
As a user, I can type commands without a certain order in arguments - so that I only need to remember the flags for each command and not the order
priority
as a user i can type commands without a certain order in arguments so that i only need to remember the flags for each command and not the order
1
506,566
14,667,971,182
IssuesEvent
2020-12-29 20:00:23
mike-eason/solaris
https://api.github.com/repos/mike-eason/solaris
closed
As Solaris I would like to upgrade the server infrastructure
enhancement estimate: 13 priority: high status: done
- Upgrade Ubuntu - Upgrade MongoDB Might as well spin up a new server afresh. This may resolve the issue with the messages not being marked as read.
1.0
As Solaris I would like to upgrade the server infrastructure - - Upgrade Ubuntu - Upgrade MongoDB Might as well spin up a new server afresh. This may resolve the issue with the messages not being marked as read.
priority
as solaris i would like to upgrade the server infrastructure upgrade ubuntu upgrade mongodb might as well spin up a new server afresh this may resolve the issue with the messages not being marked as read
1
115,286
4,662,351,494
IssuesEvent
2016-10-05 03:14:00
communityshare/communityshare
https://api.github.com/repos/communityshare/communityshare
closed
Alert text shown briefly on page load
[Effort] Good first issues [Priority] High [Skills] Angular [Type] Bug [Type] UX
When loading the app home page, there is an alert that briefly shows before the rest of the HTML is rendered. It's empty and looks like a bug when viewing the page. This alert should not appear. Likely resolved inside of `static/index.html` ![alerttext](https://cloud.githubusercontent.com/assets/5431237/17349160/78dce4b4-58d0-11e6-8157-8d387012ce1d.gif) This may have been introduced in in [this commit](https://github.com/communityshare/communityshare/commit/d085fb091cdfeaeffd682c73906f9bc9f8e4a68e).
1.0
Alert text shown briefly on page load - When loading the app home page, there is an alert that briefly shows before the rest of the HTML is rendered. It's empty and looks like a bug when viewing the page. This alert should not appear. Likely resolved inside of `static/index.html` ![alerttext](https://cloud.githubusercontent.com/assets/5431237/17349160/78dce4b4-58d0-11e6-8157-8d387012ce1d.gif) This may have been introduced in in [this commit](https://github.com/communityshare/communityshare/commit/d085fb091cdfeaeffd682c73906f9bc9f8e4a68e).
priority
alert text shown briefly on page load when loading the app home page there is an alert that briefly shows before the rest of the html is rendered it s empty and looks like a bug when viewing the page this alert should not appear likely resolved inside of static index html this may have been introduced in in
1
98,845
4,031,973,513
IssuesEvent
2016-05-18 18:59:16
neviko/Diff-sign-project
https://api.github.com/repos/neviko/Diff-sign-project
closed
style and design of the learning mode
4 - Done point: 5 priority: high
design an construct the style of all categories of learning mode learning house learning playground learning general learning letters
1.0
style and design of the learning mode - design an construct the style of all categories of learning mode learning house learning playground learning general learning letters
priority
style and design of the learning mode design an construct the style of all categories of learning mode learning house learning playground learning general learning letters
1
514,477
14,939,910,888
IssuesEvent
2021-01-25 17:33:08
mlee97/SOEN-390-Team5
https://api.github.com/repos/mlee97/SOEN-390-Team5
opened
[USER STORY]: As an ERP user, I want to login, so that I can access the appropriate interface.
priority: high risk: high story point: 5 user story
# State the Related Minimal Requirement: The system should allow only registered users to access the information, possible with different accesses for different roles.
1.0
[USER STORY]: As an ERP user, I want to login, so that I can access the appropriate interface. - # State the Related Minimal Requirement: The system should allow only registered users to access the information, possible with different accesses for different roles.
priority
as an erp user i want to login so that i can access the appropriate interface state the related minimal requirement the system should allow only registered users to access the information possible with different accesses for different roles
1
312,012
9,541,888,911
IssuesEvent
2019-05-01 00:21:11
certbot/certbot
https://api.github.com/repos/certbot/certbot
closed
Fedora apachectl 2.4.39 no longer supports passing arguments to httpd
area: apache current sprint priority: high
## My operating system is (include version): Fedora 29 ## I installed Certbot with (certbot-auto, OS package manager, pip, etc): OS package manager Additionally: httpd-2.4.39 ## I ran this command and it produced this output: certbot renew --apache apachectl: The "-v" option is not supported. ## Certbot's behavior differed from what I expected because: certbot gets invalid output from apachectl ## Here is a Certbot log showing the issue (if available): /usr/sbin/apachectl doesn't support -v, nor -t -D DUMP_RUN_CFG, DUMP_INCLUDES, DUMP_MODULES Hacking apachectl (or a different one in the path) to specifically allow these options makes it work. Running httpd -v or -t -D DUMP_RUN_CFG does work, apachectl doesn't. ## Here is the relevant nginx server block or Apache virtualhost for the domain I am configuring: not relevant, change to /usr/sbin/apachectl, probably in 2.4.39. httpd (and config) is fine, apachectl is not. not sure if I should file a bug here, or with Fedora Issue #5439 seems very similar, but appears to be for restart of apache only. This issue occurs before the certificate -- certbot is still verifying if apache is installed.
1.0
Fedora apachectl 2.4.39 no longer supports passing arguments to httpd - ## My operating system is (include version): Fedora 29 ## I installed Certbot with (certbot-auto, OS package manager, pip, etc): OS package manager Additionally: httpd-2.4.39 ## I ran this command and it produced this output: certbot renew --apache apachectl: The "-v" option is not supported. ## Certbot's behavior differed from what I expected because: certbot gets invalid output from apachectl ## Here is a Certbot log showing the issue (if available): /usr/sbin/apachectl doesn't support -v, nor -t -D DUMP_RUN_CFG, DUMP_INCLUDES, DUMP_MODULES Hacking apachectl (or a different one in the path) to specifically allow these options makes it work. Running httpd -v or -t -D DUMP_RUN_CFG does work, apachectl doesn't. ## Here is the relevant nginx server block or Apache virtualhost for the domain I am configuring: not relevant, change to /usr/sbin/apachectl, probably in 2.4.39. httpd (and config) is fine, apachectl is not. not sure if I should file a bug here, or with Fedora Issue #5439 seems very similar, but appears to be for restart of apache only. This issue occurs before the certificate -- certbot is still verifying if apache is installed.
priority
fedora apachectl no longer supports passing arguments to httpd my operating system is include version fedora i installed certbot with certbot auto os package manager pip etc os package manager additionally httpd i ran this command and it produced this output certbot renew apache apachectl the v option is not supported certbot s behavior differed from what i expected because certbot gets invalid output from apachectl here is a certbot log showing the issue if available usr sbin apachectl doesn t support v nor t d dump run cfg dump includes dump modules hacking apachectl or a different one in the path to specifically allow these options makes it work running httpd v or t d dump run cfg does work apachectl doesn t here is the relevant nginx server block or apache virtualhost for the domain i am configuring not relevant change to usr sbin apachectl probably in httpd and config is fine apachectl is not not sure if i should file a bug here or with fedora issue seems very similar but appears to be for restart of apache only this issue occurs before the certificate certbot is still verifying if apache is installed
1
561,813
16,624,605,863
IssuesEvent
2021-06-03 07:59:46
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
[LS] No documentation provided on hover for Record Fields
Area/Hover Priority/High SwanLakeDump Team/LanguageServer Type/Bug
**Description:** $subject **Steps to reproduce:** ```ballerina public type NodeCredential record {| # IP Address string ip; int port; string username; constants:NodeType nodetype; # Optfield string optField ?; |}; ```
1.0
[LS] No documentation provided on hover for Record Fields - **Description:** $subject **Steps to reproduce:** ```ballerina public type NodeCredential record {| # IP Address string ip; int port; string username; constants:NodeType nodetype; # Optfield string optField ?; |}; ```
priority
no documentation provided on hover for record fields description subject steps to reproduce ballerina public type nodecredential record ip address string ip int port string username constants nodetype nodetype optfield string optfield
1
76,754
3,491,912,585
IssuesEvent
2016-01-04 17:48:18
ceylon/ceylon
https://api.github.com/repos/ceylon/ceylon
closed
Broken if-then-else expression in JavaScript compilation
bug c-compiler-js high priority
The following code, which perfectly works when compiling to JVM breaks in the JavaScript version of the program: ``` "Run the module `ifelse`." shared void run() { Integer x = 1; Integer y = 2; String text = (if (x > 0) then x.string + "a " else "") + y.string + "b"; assert (text == "1a 2b"); } ``` When extracting the expression `(if (x > 0) then x.string + "a " else "")` to a new local variable, everything is fine in JavaScript, too.
1.0
Broken if-then-else expression in JavaScript compilation - The following code, which perfectly works when compiling to JVM breaks in the JavaScript version of the program: ``` "Run the module `ifelse`." shared void run() { Integer x = 1; Integer y = 2; String text = (if (x > 0) then x.string + "a " else "") + y.string + "b"; assert (text == "1a 2b"); } ``` When extracting the expression `(if (x > 0) then x.string + "a " else "")` to a new local variable, everything is fine in JavaScript, too.
priority
broken if then else expression in javascript compilation the following code which perfectly works when compiling to jvm breaks in the javascript version of the program run the module ifelse shared void run integer x integer y string text if x then x string a else y string b assert text when extracting the expression if x then x string a else to a new local variable everything is fine in javascript too
1
752,981
26,335,998,190
IssuesEvent
2023-01-10 14:23:33
vaticle/typedb
https://api.github.com/repos/vaticle/typedb
closed
Improve error message for MacOS M1 chip users
type: bug priority: high
## Problem to Solve MacBook M1 users get the following unhelpful error message on server startup: ``` ================================================================ ________ __ __ _____ _______ _____ _____ |__ __|\ \ / /| _ \ | _ || _ \ | _ \ | | \ \/ / | | | || | |__|| | | || | | | | | \ / | |/ / | |___ | | | || |/ / | | | | | __/ | ___| | | | || _ \ | | | | | | | | __ | | | || | | | | | | | | | | |_| || |/ / | |/ / |__| |__| |__| |_______||_____/ |_____/ A STRONGLY-TYPED DATABASE ================================================================ Exception in thread "main" java.lang.UnsatisfiedLinkError: /private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: dlopen(/private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib, 1): no suitable image found. Did find: /private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: mach-o, but wrong architecture /private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: mach-o, but wrong architecture at java.base/jdk.internal.loader.NativeLibraries.load(Native Method) at java.base/jdk.internal.loader.NativeLibraries$NativeLibraryImpl.open(NativeLibraries.java:384) at java.base/jdk.internal.loader.NativeLibraries.loadLibrary(NativeLibraries.java:228) at java.base/jdk.internal.loader.NativeLibraries.loadLibrary(NativeLibraries.java:170) at java.base/java.lang.ClassLoader.loadLibrary(ClassLoader.java:2389) at java.base/java.lang.Runtime.load0(Runtime.java:755) at java.base/java.lang.System.load(System.java:1953) at org.rocksdb.NativeLibraryLoader.loadLibraryFromJar(NativeLibraryLoader.java:79) at org.rocksdb.NativeLibraryLoader.loadLibrary(NativeLibraryLoader.java:57) at org.rocksdb.RocksDB.loadLibrary(RocksDB.java:69) at org.rocksdb.RocksDB.<clinit>(RocksDB.java:38) at com.vaticle.typedb.core.rocks.RocksTypeDB.<clinit>(RocksTypeDB.java:47) at com.vaticle.typedb.core.rocks.RocksFactory.typedb(RocksFactory.java:34) at com.vaticle.typedb.core.server.TypeDBServer.<init>(TypeDBServer.java:88) at com.vaticle.typedb.core.server.TypeDBServer.<init>(TypeDBServer.java:72) at com.vaticle.typedb.core.server.TypeDBServer.runServer(TypeDBServer.java:244) at com.vaticle.typedb.core.server.TypeDBServer.main(TypeDBServer.java:225) ``` ## Proposed Solution We should detect MacOS M1 architecture running on a non-x64 JDK and give a better error message, something like: "The installed JDK is not compatible with TypeDB, please use a JDK build that targets x64 architecture. See https://forum.vaticle.com/t/running-typedb-on-an-m1-macbook/53 for details"
1.0
Improve error message for MacOS M1 chip users - ## Problem to Solve MacBook M1 users get the following unhelpful error message on server startup: ``` ================================================================ ________ __ __ _____ _______ _____ _____ |__ __|\ \ / /| _ \ | _ || _ \ | _ \ | | \ \/ / | | | || | |__|| | | || | | | | | \ / | |/ / | |___ | | | || |/ / | | | | | __/ | ___| | | | || _ \ | | | | | | | | __ | | | || | | | | | | | | | | |_| || |/ / | |/ / |__| |__| |__| |_______||_____/ |_____/ A STRONGLY-TYPED DATABASE ================================================================ Exception in thread "main" java.lang.UnsatisfiedLinkError: /private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: dlopen(/private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib, 1): no suitable image found. Did find: /private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: mach-o, but wrong architecture /private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: mach-o, but wrong architecture at java.base/jdk.internal.loader.NativeLibraries.load(Native Method) at java.base/jdk.internal.loader.NativeLibraries$NativeLibraryImpl.open(NativeLibraries.java:384) at java.base/jdk.internal.loader.NativeLibraries.loadLibrary(NativeLibraries.java:228) at java.base/jdk.internal.loader.NativeLibraries.loadLibrary(NativeLibraries.java:170) at java.base/java.lang.ClassLoader.loadLibrary(ClassLoader.java:2389) at java.base/java.lang.Runtime.load0(Runtime.java:755) at java.base/java.lang.System.load(System.java:1953) at org.rocksdb.NativeLibraryLoader.loadLibraryFromJar(NativeLibraryLoader.java:79) at org.rocksdb.NativeLibraryLoader.loadLibrary(NativeLibraryLoader.java:57) at org.rocksdb.RocksDB.loadLibrary(RocksDB.java:69) at org.rocksdb.RocksDB.<clinit>(RocksDB.java:38) at com.vaticle.typedb.core.rocks.RocksTypeDB.<clinit>(RocksTypeDB.java:47) at com.vaticle.typedb.core.rocks.RocksFactory.typedb(RocksFactory.java:34) at com.vaticle.typedb.core.server.TypeDBServer.<init>(TypeDBServer.java:88) at com.vaticle.typedb.core.server.TypeDBServer.<init>(TypeDBServer.java:72) at com.vaticle.typedb.core.server.TypeDBServer.runServer(TypeDBServer.java:244) at com.vaticle.typedb.core.server.TypeDBServer.main(TypeDBServer.java:225) ``` ## Proposed Solution We should detect MacOS M1 architecture running on a non-x64 JDK and give a better error message, something like: "The installed JDK is not compatible with TypeDB, please use a JDK build that targets x64 architecture. See https://forum.vaticle.com/t/running-typedb-on-an-m1-macbook/53 for details"
priority
improve error message for macos chip users problem to solve macbook users get the following unhelpful error message on server startup a strongly typed database exception in thread main java lang unsatisfiedlinkerror private var folders ry t jnilib dlopen private var folders ry t jnilib no suitable image found did find private var folders ry t jnilib mach o but wrong architecture private var folders ry t jnilib mach o but wrong architecture at java base jdk internal loader nativelibraries load native method at java base jdk internal loader nativelibraries nativelibraryimpl open nativelibraries java at java base jdk internal loader nativelibraries loadlibrary nativelibraries java at java base jdk internal loader nativelibraries loadlibrary nativelibraries java at java base java lang classloader loadlibrary classloader java at java base java lang runtime runtime java at java base java lang system load system java at org rocksdb nativelibraryloader loadlibraryfromjar nativelibraryloader java at org rocksdb nativelibraryloader loadlibrary nativelibraryloader java at org rocksdb rocksdb loadlibrary rocksdb java at org rocksdb rocksdb rocksdb java at com vaticle typedb core rocks rockstypedb rockstypedb java at com vaticle typedb core rocks rocksfactory typedb rocksfactory java at com vaticle typedb core server typedbserver typedbserver java at com vaticle typedb core server typedbserver typedbserver java at com vaticle typedb core server typedbserver runserver typedbserver java at com vaticle typedb core server typedbserver main typedbserver java proposed solution we should detect macos architecture running on a non jdk and give a better error message something like the installed jdk is not compatible with typedb please use a jdk build that targets architecture see for details
1
62,794
3,193,234,367
IssuesEvent
2015-09-30 03:02:18
DarkstarProject/darkstar
https://api.github.com/repos/DarkstarProject/darkstar
closed
Maat's Cap SMN fight
High Priority
If the pet keeps auto attacking Maat and kills him while the ending CS plays (cant move) it does not count as a win for the Maat's cap quest.
1.0
Maat's Cap SMN fight - If the pet keeps auto attacking Maat and kills him while the ending CS plays (cant move) it does not count as a win for the Maat's cap quest.
priority
maat s cap smn fight if the pet keeps auto attacking maat and kills him while the ending cs plays cant move it does not count as a win for the maat s cap quest
1
305,012
9,358,442,940
IssuesEvent
2019-04-02 02:26:41
CS2103-AY1819S2-W17-4/main
https://api.github.com/repos/CS2103-AY1819S2-W17-4/main
closed
Find Expense
priority.High status.Ongoing type.Epic type.Story
To complete Task - [x] to search for an expense using - [x] name - [x] price - [x] tags - [x] date
1.0
Find Expense - To complete Task - [x] to search for an expense using - [x] name - [x] price - [x] tags - [x] date
priority
find expense to complete task to search for an expense using name price tags date
1
234,318
7,719,716,886
IssuesEvent
2018-05-23 20:21:49
aseprite/aseprite
https://api.github.com/repos/aseprite/aseprite
closed
Add clipboard support for Linux
high priority linux
We should be able to copy & paste text and images between Aseprite and other programs like GIMP.
1.0
Add clipboard support for Linux - We should be able to copy & paste text and images between Aseprite and other programs like GIMP.
priority
add clipboard support for linux we should be able to copy paste text and images between aseprite and other programs like gimp
1
280,465
8,682,193,521
IssuesEvent
2018-12-02 05:01:22
MonoOni/binarydist
https://api.github.com/repos/MonoOni/binarydist
opened
Mono profiler is broken
area: JIT os: aix os: i priority: high type: bug
Was low priority until `--trace` started to require it to work. For now, you can revert mono/mono@2b029b3df4a1816880e3f02bacd35cf9a0835e4e, but this won't be sustainable. Need to figure out how to get this working, otherwise debugging will be a LOT harder.
1.0
Mono profiler is broken - Was low priority until `--trace` started to require it to work. For now, you can revert mono/mono@2b029b3df4a1816880e3f02bacd35cf9a0835e4e, but this won't be sustainable. Need to figure out how to get this working, otherwise debugging will be a LOT harder.
priority
mono profiler is broken was low priority until trace started to require it to work for now you can revert mono mono but this won t be sustainable need to figure out how to get this working otherwise debugging will be a lot harder
1
827,441
31,772,912,481
IssuesEvent
2023-09-12 12:53:32
Budibase/budibase
https://api.github.com/repos/Budibase/budibase
closed
[BUDI-7468] Save Row action will attempt to save the image data of the S3 Upload component
bug forms s3 env - production High priority bb-components
**Checklist** * [X] I have searched budibase discussions and github issues to check if my issue already exists **Hosting** * Cloud * Tenant ID: bb.budibase.app **Describe the bug** **Save Row** action will attempt to save the image base64 data contents of the **S3 Upload** component. This can cause a `Request Entity Too Large` if a large amount of image data is unintentionally passed to the `/rows` endpoint. **To Reproduce** Steps to reproduce the behavior: 1. Add a form with a **custom** schema 2. Add an S3 File Upload component 3. Add a button with a **Save Row** action 4. Preview the app and upload an image 5. Look in the network tab and see the raw image data being saved against the `/rows` endpoint **Expected behavior** The S3 Upload component should be ignored from the form schema to prevent it being saved against the `/rows` endpoint and causing unintended errors such as `Request Entity Too Large` **Screenshots** ![Screenshot 2023-09-07 at 21.17.32.png](https://uploads.linear.app/7ecfa297-cc77-4fdf-9117-fa5afdb8f908/9f7590f4-b946-4b9a-acec-bd93c7dd4d89/7660b2b0-5f99-43b9-a201-1f15a0538ed0?signature=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJwYXRoIjoiLzdlY2ZhMjk3LWNjNzctNGZkZi05MTE3LWZhNWFmZGI4ZjkwOC85Zjc1OTBmNC1iOTQ2LTRiOWEtYWNlYy1iZDkzYzdkZDRkODkvNzY2MGIyYjAtNWY5OS00M2I5LWEyMDEtMWYxNWEwNTM4ZWQwIiwiaWF0IjoxNjk0MTE3OTU3LCJleHAiOjE2OTQyMDQzNTd9.DXdMuQ9X54FrcnwSvmGcoihCisOKnDdtugVg7ic4fdc) ![Screenshot 2023-09-07 at 21.18.04.png](https://uploads.linear.app/7ecfa297-cc77-4fdf-9117-fa5afdb8f908/1be98085-3c14-4c3d-89cd-1bc27e388960/0d28228f-dc58-492d-b8c1-a0044084aa83?signature=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJwYXRoIjoiLzdlY2ZhMjk3LWNjNzctNGZkZi05MTE3LWZhNWFmZGI4ZjkwOC8xYmU5ODA4NS0zYzE0LTRjM2QtODljZC0xYmMyN2UzODg5NjAvMGQyODIyOGYtZGM1OC00OTJkLWI4YzEtYTAwNDQwODRhYTgzIiwiaWF0IjoxNjk0MTE3OTU3LCJleHAiOjE2OTQyMDQzNTd9.UXbkcmyvqgwu3K73iFTew9wpx5WXmTN10Z7QJkz2X5A) ![Screenshot 2023-09-07 at 21.19.02.png](https://uploads.linear.app/7ecfa297-cc77-4fdf-9117-fa5afdb8f908/4d4ccf92-4765-45a7-8ca5-3532bd7af97b/53cd0e33-978e-443a-805f-9f277782584f?signature=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJwYXRoIjoiLzdlY2ZhMjk3LWNjNzctNGZkZi05MTE3LWZhNWFmZGI4ZjkwOC80ZDRjY2Y5Mi00NzY1LTQ1YTctOGNhNS0zNTMyYmQ3YWY5N2IvNTNjZDBlMzMtOTc4ZS00NDNhLTgwNWYtOWYyNzc3ODI1ODRmIiwiaWF0IjoxNjk0MTE3OTU3LCJleHAiOjE2OTQyMDQzNTd9.H9tPgmM7f6Wt9TiUCjJ2R5OA636BEE15e52RM0wC6qU) **Additional context** Only seems to be an issue for images file types. Other object types seem to only provide a small amount of metadata json. <sub>From [SyncLinear.com](https://synclinear.com) | [BUDI-7468](https://linear.app/budibase/issue/BUDI-7468/save-row-action-will-attempt-to-save-the-image-data-of-the-s3-upload)</sub>
1.0
[BUDI-7468] Save Row action will attempt to save the image data of the S3 Upload component - **Checklist** * [X] I have searched budibase discussions and github issues to check if my issue already exists **Hosting** * Cloud * Tenant ID: bb.budibase.app **Describe the bug** **Save Row** action will attempt to save the image base64 data contents of the **S3 Upload** component. This can cause a `Request Entity Too Large` if a large amount of image data is unintentionally passed to the `/rows` endpoint. **To Reproduce** Steps to reproduce the behavior: 1. Add a form with a **custom** schema 2. Add an S3 File Upload component 3. Add a button with a **Save Row** action 4. Preview the app and upload an image 5. Look in the network tab and see the raw image data being saved against the `/rows` endpoint **Expected behavior** The S3 Upload component should be ignored from the form schema to prevent it being saved against the `/rows` endpoint and causing unintended errors such as `Request Entity Too Large` **Screenshots** ![Screenshot 2023-09-07 at 21.17.32.png](https://uploads.linear.app/7ecfa297-cc77-4fdf-9117-fa5afdb8f908/9f7590f4-b946-4b9a-acec-bd93c7dd4d89/7660b2b0-5f99-43b9-a201-1f15a0538ed0?signature=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJwYXRoIjoiLzdlY2ZhMjk3LWNjNzctNGZkZi05MTE3LWZhNWFmZGI4ZjkwOC85Zjc1OTBmNC1iOTQ2LTRiOWEtYWNlYy1iZDkzYzdkZDRkODkvNzY2MGIyYjAtNWY5OS00M2I5LWEyMDEtMWYxNWEwNTM4ZWQwIiwiaWF0IjoxNjk0MTE3OTU3LCJleHAiOjE2OTQyMDQzNTd9.DXdMuQ9X54FrcnwSvmGcoihCisOKnDdtugVg7ic4fdc) ![Screenshot 2023-09-07 at 21.18.04.png](https://uploads.linear.app/7ecfa297-cc77-4fdf-9117-fa5afdb8f908/1be98085-3c14-4c3d-89cd-1bc27e388960/0d28228f-dc58-492d-b8c1-a0044084aa83?signature=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJwYXRoIjoiLzdlY2ZhMjk3LWNjNzctNGZkZi05MTE3LWZhNWFmZGI4ZjkwOC8xYmU5ODA4NS0zYzE0LTRjM2QtODljZC0xYmMyN2UzODg5NjAvMGQyODIyOGYtZGM1OC00OTJkLWI4YzEtYTAwNDQwODRhYTgzIiwiaWF0IjoxNjk0MTE3OTU3LCJleHAiOjE2OTQyMDQzNTd9.UXbkcmyvqgwu3K73iFTew9wpx5WXmTN10Z7QJkz2X5A) ![Screenshot 2023-09-07 at 21.19.02.png](https://uploads.linear.app/7ecfa297-cc77-4fdf-9117-fa5afdb8f908/4d4ccf92-4765-45a7-8ca5-3532bd7af97b/53cd0e33-978e-443a-805f-9f277782584f?signature=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJwYXRoIjoiLzdlY2ZhMjk3LWNjNzctNGZkZi05MTE3LWZhNWFmZGI4ZjkwOC80ZDRjY2Y5Mi00NzY1LTQ1YTctOGNhNS0zNTMyYmQ3YWY5N2IvNTNjZDBlMzMtOTc4ZS00NDNhLTgwNWYtOWYyNzc3ODI1ODRmIiwiaWF0IjoxNjk0MTE3OTU3LCJleHAiOjE2OTQyMDQzNTd9.H9tPgmM7f6Wt9TiUCjJ2R5OA636BEE15e52RM0wC6qU) **Additional context** Only seems to be an issue for images file types. Other object types seem to only provide a small amount of metadata json. <sub>From [SyncLinear.com](https://synclinear.com) | [BUDI-7468](https://linear.app/budibase/issue/BUDI-7468/save-row-action-will-attempt-to-save-the-image-data-of-the-s3-upload)</sub>
priority
save row action will attempt to save the image data of the upload component checklist i have searched budibase discussions and github issues to check if my issue already exists hosting cloud tenant id bb budibase app describe the bug save row action will attempt to save the image data contents of the upload component this can cause a request entity too large if a large amount of image data is unintentionally passed to the rows endpoint to reproduce steps to reproduce the behavior add a form with a custom schema add an file upload component add a button with a save row action preview the app and upload an image look in the network tab and see the raw image data being saved against the rows endpoint expected behavior the upload component should be ignored from the form schema to prevent it being saved against the rows endpoint and causing unintended errors such as request entity too large screenshots additional context only seems to be an issue for images file types other object types seem to only provide a small amount of metadata json from
1
669,219
22,616,722,224
IssuesEvent
2022-06-29 23:17:11
paleobot/pbot-dev
https://api.github.com/repos/paleobot/pbot-dev
opened
Administrators Group
enhancement high priority
We need there to be an Administrators Group that allows system administrators to see private data. Among other things, this allow PBot ExCom to see what data have been in the private sphere for a long time and needs to go public.
1.0
Administrators Group - We need there to be an Administrators Group that allows system administrators to see private data. Among other things, this allow PBot ExCom to see what data have been in the private sphere for a long time and needs to go public.
priority
administrators group we need there to be an administrators group that allows system administrators to see private data among other things this allow pbot excom to see what data have been in the private sphere for a long time and needs to go public
1
96,514
3,969,098,400
IssuesEvent
2016-05-03 22:02:24
KrzysztofSendor/python-chess
https://api.github.com/repos/KrzysztofSendor/python-chess
closed
Coordinates translation
difficulty: easy priority: high
Create function in board module for translating notation (eg. a2, c6) to Board array coordinates (eg. Board[0][6], Board[2][2]). Mind that array is counted from upper left with whites in bottom and in standard notation board is counted from bottom left with whites in bottom. Practically the whole code is already written in movement handling function. Just copy it. **Should return a list (an array) of integers.** `input = "a2" output = [0,6]` Create second function that does the same thing but the other way around. **Should return a list from a string given.** **All further created functions should use it.**
1.0
Coordinates translation - Create function in board module for translating notation (eg. a2, c6) to Board array coordinates (eg. Board[0][6], Board[2][2]). Mind that array is counted from upper left with whites in bottom and in standard notation board is counted from bottom left with whites in bottom. Practically the whole code is already written in movement handling function. Just copy it. **Should return a list (an array) of integers.** `input = "a2" output = [0,6]` Create second function that does the same thing but the other way around. **Should return a list from a string given.** **All further created functions should use it.**
priority
coordinates translation create function in board module for translating notation eg to board array coordinates eg board board mind that array is counted from upper left with whites in bottom and in standard notation board is counted from bottom left with whites in bottom practically the whole code is already written in movement handling function just copy it should return a list an array of integers input output create second function that does the same thing but the other way around should return a list from a string given all further created functions should use it
1
764,299
26,793,891,756
IssuesEvent
2023-02-01 10:27:02
wso2/carbon-analytics
https://api.github.com/repos/wso2/carbon-analytics
closed
[Business Rules] Error shown while deploying the business rule
Type/Bug Severity/Major Priority/High Component/BusinessRules
**Description:** When creating a new business rule from a template, and press save & deploy an error 'Failed to deploy' is shown in the UI and dashboard terminal while deploying the app. `Failed to deploy business rule sweetProd org.wso2.carbon.business.rules.core.exceptions.SiddhiAppsApiHelperException: Unexpected error occurred during saving the siddhi app '@App:name('sweetprod_0')` But the business rule get successfully deployed. ![image](https://user-images.githubusercontent.com/20179540/43074597-7df781f4-8e9b-11e8-89b4-d6d02130a67d.png) **Suggested Labels:** Bug, Business Rules **Suggested Assignees:** @minudika **Affected Product Version:** 4.2.0 **OS, DB, other environment details and versions:** Windows10
1.0
[Business Rules] Error shown while deploying the business rule - **Description:** When creating a new business rule from a template, and press save & deploy an error 'Failed to deploy' is shown in the UI and dashboard terminal while deploying the app. `Failed to deploy business rule sweetProd org.wso2.carbon.business.rules.core.exceptions.SiddhiAppsApiHelperException: Unexpected error occurred during saving the siddhi app '@App:name('sweetprod_0')` But the business rule get successfully deployed. ![image](https://user-images.githubusercontent.com/20179540/43074597-7df781f4-8e9b-11e8-89b4-d6d02130a67d.png) **Suggested Labels:** Bug, Business Rules **Suggested Assignees:** @minudika **Affected Product Version:** 4.2.0 **OS, DB, other environment details and versions:** Windows10
priority
error shown while deploying the business rule description when creating a new business rule from a template and press save deploy an error failed to deploy is shown in the ui and dashboard terminal while deploying the app failed to deploy business rule sweetprod org carbon business rules core exceptions siddhiappsapihelperexception unexpected error occurred during saving the siddhi app app name sweetprod but the business rule get successfully deployed suggested labels bug business rules suggested assignees minudika affected product version os db other environment details and versions
1
183,212
6,678,391,666
IssuesEvent
2017-10-05 14:06:36
CCAFS/MARLO
https://api.github.com/repos/CCAFS/MARLO
closed
Harvesting metadata from CIMMYT repositories
Priority - High Type -Task
Branch: dev-rest-metadata - [ ] Create connection class to CIMMYT - [ ] Map XML data to metadata Model - [ ] Add CIMMYT as example in deliverables section
1.0
Harvesting metadata from CIMMYT repositories - Branch: dev-rest-metadata - [ ] Create connection class to CIMMYT - [ ] Map XML data to metadata Model - [ ] Add CIMMYT as example in deliverables section
priority
harvesting metadata from cimmyt repositories branch dev rest metadata create connection class to cimmyt map xml data to metadata model add cimmyt as example in deliverables section
1
382,154
11,301,565,434
IssuesEvent
2020-01-17 15:50:35
mono/monodevelop
https://api.github.com/repos/mono/monodevelop
closed
[Diagnostics] Implement RuleSet file support
Area: C# Binding high-priority not-regression vs-sync
VS normally looks for ruleset files included via MSBuild: https://github.com/dotnet/roslyn/blob/0c62bae6aade582227232183540fd891254a76cb/src/Compilers/Core/Portable/CodeAnalysis.csproj#L14 Support for that is implemented below for VS: http://source.roslyn.io/#Microsoft.VisualStudio.LanguageServices/Implementation/ProjectSystem/RuleSets/VisualStudioRuleSetManager.RuleSetFile.cs,106 > VS bug #577079
1.0
[Diagnostics] Implement RuleSet file support - VS normally looks for ruleset files included via MSBuild: https://github.com/dotnet/roslyn/blob/0c62bae6aade582227232183540fd891254a76cb/src/Compilers/Core/Portable/CodeAnalysis.csproj#L14 Support for that is implemented below for VS: http://source.roslyn.io/#Microsoft.VisualStudio.LanguageServices/Implementation/ProjectSystem/RuleSets/VisualStudioRuleSetManager.RuleSetFile.cs,106 > VS bug #577079
priority
implement ruleset file support vs normally looks for ruleset files included via msbuild support for that is implemented below for vs vs bug
1
211,729
7,204,040,136
IssuesEvent
2018-02-06 11:13:15
xmichelo/Beeftext
https://api.github.com/repos/xmichelo/Beeftext
closed
Error message on first launch.
bug high priority
Due to recent changes, the application now display an error message when it is first launched on PC because there is no combo list file.
1.0
Error message on first launch. - Due to recent changes, the application now display an error message when it is first launched on PC because there is no combo list file.
priority
error message on first launch due to recent changes the application now display an error message when it is first launched on pc because there is no combo list file
1
773,751
27,169,299,611
IssuesEvent
2023-02-17 17:52:45
minio/minio
https://api.github.com/repos/minio/minio
closed
Buckets info and cluster summary not changes
community priority: high
## NOTE Buckets info and cluster summary not changes. Objects count and used space are the same even after few hours of big uploads. I now migrating date from one cluster to another. Before update to the current version information about object count and used space was changed every 3-5 minutes but now it shows same values for the all time after update. This values not changes in web console and also in metrics which sent to prometheus. It's interesting that only total used space metric which sending to prometheus are changes - objects count and buckets used space are the same. For now prometheus shows me 6.4 TB used and web console only 4.4. You can see it at screenshots attached. ## Expected Behavior Information about used space and objects count should be changed according real data. ## Current Behavior For in it shows values which was before update few days ago. ## Regression It works fine before update in version RELEASE.2023-01-12T02-06-16Z (commit-id=7bc95c47a322971aff7d4d4c270dcf28a933e84b) ## Your Environment * Version used (`minio --version`): RELEASE.2023-02-10T18-48-39Z (commit-id=d0f4cc89a5d0ca45e7ba00fbb7733c2fead3f44e) * Server setup and configuration: 4 nodes with 4 disks on each * Operating System and version (`uname -a`): Linux minio-cold-1 5.15.0-60-generic #66-Ubuntu SMP Fri Jan 20 14:29:49 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
1.0
Buckets info and cluster summary not changes - ## NOTE Buckets info and cluster summary not changes. Objects count and used space are the same even after few hours of big uploads. I now migrating date from one cluster to another. Before update to the current version information about object count and used space was changed every 3-5 minutes but now it shows same values for the all time after update. This values not changes in web console and also in metrics which sent to prometheus. It's interesting that only total used space metric which sending to prometheus are changes - objects count and buckets used space are the same. For now prometheus shows me 6.4 TB used and web console only 4.4. You can see it at screenshots attached. ## Expected Behavior Information about used space and objects count should be changed according real data. ## Current Behavior For in it shows values which was before update few days ago. ## Regression It works fine before update in version RELEASE.2023-01-12T02-06-16Z (commit-id=7bc95c47a322971aff7d4d4c270dcf28a933e84b) ## Your Environment * Version used (`minio --version`): RELEASE.2023-02-10T18-48-39Z (commit-id=d0f4cc89a5d0ca45e7ba00fbb7733c2fead3f44e) * Server setup and configuration: 4 nodes with 4 disks on each * Operating System and version (`uname -a`): Linux minio-cold-1 5.15.0-60-generic #66-Ubuntu SMP Fri Jan 20 14:29:49 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
priority
buckets info and cluster summary not changes note buckets info and cluster summary not changes objects count and used space are the same even after few hours of big uploads i now migrating date from one cluster to another before update to the current version information about object count and used space was changed every minutes but now it shows same values for the all time after update this values not changes in web console and also in metrics which sent to prometheus it s interesting that only total used space metric which sending to prometheus are changes objects count and buckets used space are the same for now prometheus shows me tb used and web console only you can see it at screenshots attached expected behavior information about used space and objects count should be changed according real data current behavior for in it shows values which was before update few days ago regression it works fine before update in version release commit id your environment version used minio version release commit id server setup and configuration nodes with disks on each operating system and version uname a linux minio cold generic ubuntu smp fri jan utc gnu linux
1
695,565
23,864,170,046
IssuesEvent
2022-09-07 09:35:14
factly/dega
https://api.github.com/repos/factly/dega
closed
Updating kavach to the latest changes
priority:high studio server
This change includes - - [x] moving spaces to kavach. - [x] moving policies to kavach.
1.0
Updating kavach to the latest changes - This change includes - - [x] moving spaces to kavach. - [x] moving policies to kavach.
priority
updating kavach to the latest changes this change includes moving spaces to kavach moving policies to kavach
1
554,244
16,415,358,180
IssuesEvent
2021-05-19 05:43:46
django-cms/django-cms
https://api.github.com/repos/django-cms/django-cms
closed
Slug uniqueness not checked when moving page
needs expert opinion needs patch priority: high
## Description When page is moved, its slug uniqueness is not checked. Thus two pages with the same slug can be present on the same level. ## Steps to reproduce 1. Create a page at the root level and give it a slug "test". 2. Publish the page 3. Using the Pages manager, copy the page and paste it as a child of the test page 4. Move the page back to the root - we now have two pages with the same slug (and both can be published) ## Expected behaviour Slug should be checked and the page move should either not be allowed or slug should be automatically modified. ## Actual behaviour Two pages with the same slug are allowed on the same level. Both can be published at the same time. The preview link in the Pages admin only points to the original (not to the copy). ## Additional information Situation has been reproduced on the Divio platform using a clean project - https://cms-bug-demo-stage.us.aldryn.io/en/admin/cms/page/?language=en
1.0
Slug uniqueness not checked when moving page - ## Description When page is moved, its slug uniqueness is not checked. Thus two pages with the same slug can be present on the same level. ## Steps to reproduce 1. Create a page at the root level and give it a slug "test". 2. Publish the page 3. Using the Pages manager, copy the page and paste it as a child of the test page 4. Move the page back to the root - we now have two pages with the same slug (and both can be published) ## Expected behaviour Slug should be checked and the page move should either not be allowed or slug should be automatically modified. ## Actual behaviour Two pages with the same slug are allowed on the same level. Both can be published at the same time. The preview link in the Pages admin only points to the original (not to the copy). ## Additional information Situation has been reproduced on the Divio platform using a clean project - https://cms-bug-demo-stage.us.aldryn.io/en/admin/cms/page/?language=en
priority
slug uniqueness not checked when moving page description when page is moved its slug uniqueness is not checked thus two pages with the same slug can be present on the same level steps to reproduce create a page at the root level and give it a slug test publish the page using the pages manager copy the page and paste it as a child of the test page move the page back to the root we now have two pages with the same slug and both can be published expected behaviour slug should be checked and the page move should either not be allowed or slug should be automatically modified actual behaviour two pages with the same slug are allowed on the same level both can be published at the same time the preview link in the pages admin only points to the original not to the copy additional information situation has been reproduced on the divio platform using a clean project
1
197,539
6,960,714,765
IssuesEvent
2017-12-08 05:33:32
OpusCapita/react-crudeditor
https://api.github.com/repos/OpusCapita/react-crudeditor
closed
Allow instance to have fields not mentioned in model.fields
requested priority: high type:enhancement
Server-side may send an instance with fields, some of which are not described in model.fields.
1.0
Allow instance to have fields not mentioned in model.fields - Server-side may send an instance with fields, some of which are not described in model.fields.
priority
allow instance to have fields not mentioned in model fields server side may send an instance with fields some of which are not described in model fields
1
687,961
23,543,718,151
IssuesEvent
2022-08-20 20:05:05
AkjoStudios/AkjoNav
https://api.github.com/repos/AkjoStudios/AkjoNav
closed
[Feature Enhancement] Add common class for main tab panels
priority: high status: planned type: enhancement Official Issue
### In which module is the feature you want to enhance? AkjoNav-DesktopMapper ### Feature Main tab panels ### Description Move all common features of the main tab panels into a separate class. ### Additional Context No additional context/screenshots.
1.0
[Feature Enhancement] Add common class for main tab panels - ### In which module is the feature you want to enhance? AkjoNav-DesktopMapper ### Feature Main tab panels ### Description Move all common features of the main tab panels into a separate class. ### Additional Context No additional context/screenshots.
priority
add common class for main tab panels in which module is the feature you want to enhance akjonav desktopmapper feature main tab panels description move all common features of the main tab panels into a separate class additional context no additional context screenshots
1
593,123
17,947,811,571
IssuesEvent
2021-09-12 06:02:14
kiudee/chess-tuning-tools
https://api.github.com/repos/kiudee/chess-tuning-tools
closed
Divide by zero error when using input warping
bug Priority: High
* Chess Tuning Tools version: **0.6.0b2** * Python version: **3.8** * Operating System: vast.ai instance using the **nvidia/cuda:10.1-cudnn7-devel-ubuntu18.04** image. ### Description I'm trying a tune of three parameters but turning on `warp_input` throws a divide by zero error every iteration and a weird warping for one of the parameters along with weird confidence bound for all parameters. (Everything looks normal without warping.) **Error:** ``` /root/miniconda/envs/tuning/lib/python3.8/site-packages/sklearn/gaussian_process/kernels.py:255: RuntimeWarning: divide by zero encountered in log return np.log(np.hstack(theta)) ``` **Without warping:** ![20200909-182503-280](https://user-images.githubusercontent.com/10537957/92638661-3bc9ee00-f2a0-11ea-9736-1a4341520d02.png) **With warping:** ![20200909-182930-280](https://user-images.githubusercontent.com/10537957/92639126-ff4ac200-f2a0-11ea-9d35-24e4fd74c06b.png) **Warp functions:** ``` Input warping was applied using the following parameters for the beta distributions: {'ScaleQ': (0.191, 0.849), 'Cpuct': (0.471, 1.083), 'FpuValue': (0.635, 1.421)} ``` ![image](https://user-images.githubusercontent.com/10537957/92641035-c8c27680-f2a3-11ea-8f87-89c22c7803a4.png) This warping for `ScaleQ` doesn't make sense but I suspect that it doesn't know what it's doing because the effect of this parameter is still not yet determined very well. The effect isn't nearly as strong as the other parameters and there is an expected interaction among all of them. I'll keep running non-warped `ts` and `vr` sampling for a while and try again once it starts to understand better. Files to reproduce: [DataAndConfig.zip](https://github.com/kiudee/chess-tuning-tools/files/5197274/DataAndConfig.zip) ### What I Did Changing `"warp_inputs": false` to `"warp_inputs": true` in the `config.json` results in the runtime warning: ``` 2020-09-09 18:34:09,833 INFO Got score: 5.6995265465242655 +- 1.9955240492535908 2020-09-09 18:34:09,834 INFO Updating model /root/miniconda/envs/tuning/lib/python3.8/site-packages/sklearn/gaussian_process/kernels.py:255: RuntimeWarning: divide by zero encountered in log return np.log(np.hstack(theta)) 2020-09-09 18:34:28,397 INFO GP sampling finished (18.563089s) 2020-09-09 18:34:28,398 DEBUG GP kernel: 1.54**2 * Matern(length_scale=[0.522, 0.362, 0.276], nu=2.5) + WhiteKernel(noise_level=3.88e-06) 2020-09-09 18:34:28,399 DEBUG Input warping was applied using the following parameters for the beta distributions: {'ScaleQ': (0.191, 0.849), 'Cpuct': (0.471, 1.083), 'FpuValue': (0.635, 1.421)} 2020-09-09 18:34:28,408 INFO Starting iteration 281 2020-09-09 18:34:41,482 INFO Current optimum: {'ScaleQ': 0.3991959134189197, 'Cpuct': 4.700855456794591, 'FpuValue': 1.4722686789734403} 2020-09-09 18:34:41,482 INFO Estimated value: -0.5733 +- 0.3404 2020-09-09 18:34:41,482 INFO 90.0% confidence interval of the value: (-1.1332, -0.0133) 2020-09-09 18:34:41,648 INFO 90.0% confidence intervals of the parameters: Parameter Lower bound Upper bound --------------------------------------------------- ScaleQ 11.758489497804575 11.758489497804575 Cpuct 0.9963961100717935 0.9963961100717935 FpuValue -0.9812151210716575 -0.9812151210716575 ``` --- Related to #92.
1.0
Divide by zero error when using input warping - * Chess Tuning Tools version: **0.6.0b2** * Python version: **3.8** * Operating System: vast.ai instance using the **nvidia/cuda:10.1-cudnn7-devel-ubuntu18.04** image. ### Description I'm trying a tune of three parameters but turning on `warp_input` throws a divide by zero error every iteration and a weird warping for one of the parameters along with weird confidence bound for all parameters. (Everything looks normal without warping.) **Error:** ``` /root/miniconda/envs/tuning/lib/python3.8/site-packages/sklearn/gaussian_process/kernels.py:255: RuntimeWarning: divide by zero encountered in log return np.log(np.hstack(theta)) ``` **Without warping:** ![20200909-182503-280](https://user-images.githubusercontent.com/10537957/92638661-3bc9ee00-f2a0-11ea-9736-1a4341520d02.png) **With warping:** ![20200909-182930-280](https://user-images.githubusercontent.com/10537957/92639126-ff4ac200-f2a0-11ea-9d35-24e4fd74c06b.png) **Warp functions:** ``` Input warping was applied using the following parameters for the beta distributions: {'ScaleQ': (0.191, 0.849), 'Cpuct': (0.471, 1.083), 'FpuValue': (0.635, 1.421)} ``` ![image](https://user-images.githubusercontent.com/10537957/92641035-c8c27680-f2a3-11ea-8f87-89c22c7803a4.png) This warping for `ScaleQ` doesn't make sense but I suspect that it doesn't know what it's doing because the effect of this parameter is still not yet determined very well. The effect isn't nearly as strong as the other parameters and there is an expected interaction among all of them. I'll keep running non-warped `ts` and `vr` sampling for a while and try again once it starts to understand better. Files to reproduce: [DataAndConfig.zip](https://github.com/kiudee/chess-tuning-tools/files/5197274/DataAndConfig.zip) ### What I Did Changing `"warp_inputs": false` to `"warp_inputs": true` in the `config.json` results in the runtime warning: ``` 2020-09-09 18:34:09,833 INFO Got score: 5.6995265465242655 +- 1.9955240492535908 2020-09-09 18:34:09,834 INFO Updating model /root/miniconda/envs/tuning/lib/python3.8/site-packages/sklearn/gaussian_process/kernels.py:255: RuntimeWarning: divide by zero encountered in log return np.log(np.hstack(theta)) 2020-09-09 18:34:28,397 INFO GP sampling finished (18.563089s) 2020-09-09 18:34:28,398 DEBUG GP kernel: 1.54**2 * Matern(length_scale=[0.522, 0.362, 0.276], nu=2.5) + WhiteKernel(noise_level=3.88e-06) 2020-09-09 18:34:28,399 DEBUG Input warping was applied using the following parameters for the beta distributions: {'ScaleQ': (0.191, 0.849), 'Cpuct': (0.471, 1.083), 'FpuValue': (0.635, 1.421)} 2020-09-09 18:34:28,408 INFO Starting iteration 281 2020-09-09 18:34:41,482 INFO Current optimum: {'ScaleQ': 0.3991959134189197, 'Cpuct': 4.700855456794591, 'FpuValue': 1.4722686789734403} 2020-09-09 18:34:41,482 INFO Estimated value: -0.5733 +- 0.3404 2020-09-09 18:34:41,482 INFO 90.0% confidence interval of the value: (-1.1332, -0.0133) 2020-09-09 18:34:41,648 INFO 90.0% confidence intervals of the parameters: Parameter Lower bound Upper bound --------------------------------------------------- ScaleQ 11.758489497804575 11.758489497804575 Cpuct 0.9963961100717935 0.9963961100717935 FpuValue -0.9812151210716575 -0.9812151210716575 ``` --- Related to #92.
priority
divide by zero error when using input warping chess tuning tools version python version operating system vast ai instance using the nvidia cuda devel image description i m trying a tune of three parameters but turning on warp input throws a divide by zero error every iteration and a weird warping for one of the parameters along with weird confidence bound for all parameters everything looks normal without warping error root miniconda envs tuning lib site packages sklearn gaussian process kernels py runtimewarning divide by zero encountered in log return np log np hstack theta without warping with warping warp functions input warping was applied using the following parameters for the beta distributions scaleq cpuct fpuvalue this warping for scaleq doesn t make sense but i suspect that it doesn t know what it s doing because the effect of this parameter is still not yet determined very well the effect isn t nearly as strong as the other parameters and there is an expected interaction among all of them i ll keep running non warped ts and vr sampling for a while and try again once it starts to understand better files to reproduce what i did changing warp inputs false to warp inputs true in the config json results in the runtime warning info got score info updating model root miniconda envs tuning lib site packages sklearn gaussian process kernels py runtimewarning divide by zero encountered in log return np log np hstack theta info gp sampling finished debug gp kernel matern length scale nu whitekernel noise level debug input warping was applied using the following parameters for the beta distributions scaleq cpuct fpuvalue info starting iteration info current optimum scaleq cpuct fpuvalue info estimated value info confidence interval of the value info confidence intervals of the parameters parameter lower bound upper bound scaleq cpuct fpuvalue related to
1
452,538
13,055,436,003
IssuesEvent
2020-07-30 01:40:51
Azure/vscode-kubernetes-tools
https://api.github.com/repos/Azure/vscode-kubernetes-tools
closed
Terminal shell path broken
bug priority-high
Since upgrading to latest version I get a `The terminal shell path ""C:\Program Files (x86)\.vs-kubernetes\tools\kubectl\kubectl".exe" does not exist` The error is only when right clicking and trying to initiate a terminal session... this used to work well before. All other commands work.
1.0
Terminal shell path broken - Since upgrading to latest version I get a `The terminal shell path ""C:\Program Files (x86)\.vs-kubernetes\tools\kubectl\kubectl".exe" does not exist` The error is only when right clicking and trying to initiate a terminal session... this used to work well before. All other commands work.
priority
terminal shell path broken since upgrading to latest version i get a the terminal shell path c program files vs kubernetes tools kubectl kubectl exe does not exist the error is only when right clicking and trying to initiate a terminal session this used to work well before all other commands work
1
481,789
13,891,930,819
IssuesEvent
2020-10-19 11:26:23
CLIxIndia-Dev/clixoer
https://api.github.com/repos/CLIxIndia-Dev/clixoer
closed
The COOL PREVIEW Changes
enhancement frontend highpriority
**The cool preview changes ** This task contains changes suggestion about cool page preview - The arrange the name before the Button in cool page preview - The Affordance icon need to responsive clear in mobile view - change in copy share button ( Twitter )
1.0
The COOL PREVIEW Changes - **The cool preview changes ** This task contains changes suggestion about cool page preview - The arrange the name before the Button in cool page preview - The Affordance icon need to responsive clear in mobile view - change in copy share button ( Twitter )
priority
the cool preview changes the cool preview changes this task contains changes suggestion about cool page preview the arrange the name before the button in cool page preview the affordance icon need to responsive clear in mobile view change in copy share button twitter
1