Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
855
| labels
stringlengths 4
721
| body
stringlengths 1
261k
| index
stringclasses 13
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
240k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
443,322
| 12,792,548,904
|
IssuesEvent
|
2020-07-02 01:40:39
|
lardemua/atom
|
https://api.github.com/repos/lardemua/atom
|
closed
|
Merge general hand eye with calibrate
|
High Priority enhancement help wanted
|
Should do this to advance in the general approach.
Defined with @eupedrosa that the best way is to try to put in calibrate the stuff from hand eye, i.e. to use the calibrate as base.
|
1.0
|
Merge general hand eye with calibrate - Should do this to advance in the general approach.
Defined with @eupedrosa that the best way is to try to put in calibrate the stuff from hand eye, i.e. to use the calibrate as base.
|
priority
|
merge general hand eye with calibrate should do this to advance in the general approach defined with eupedrosa that the best way is to try to put in calibrate the stuff from hand eye i e to use the calibrate as base
| 1
|
141,381
| 5,435,457,609
|
IssuesEvent
|
2017-03-05 17:09:33
|
Templarian/MaterialDesign
|
https://api.github.com/repos/Templarian/MaterialDesign
|
closed
|
Exponent and Root
|
Alias Icon High Priority Icon Request
|
I'm sure this was part of a request at some point but I can't seem to find these icons. An `x²` and `√x` would be really useful. Could also use a decimal point (the dot talked about in #1709 would work), and a degree symbol like `x°`.
|
1.0
|
Exponent and Root - I'm sure this was part of a request at some point but I can't seem to find these icons. An `x²` and `√x` would be really useful. Could also use a decimal point (the dot talked about in #1709 would work), and a degree symbol like `x°`.
|
priority
|
exponent and root i m sure this was part of a request at some point but i can t seem to find these icons an x² and √x would be really useful could also use a decimal point the dot talked about in would work and a degree symbol like x°
| 1
|
131,858
| 5,166,426,990
|
IssuesEvent
|
2017-01-17 16:12:26
|
snaiperskaya96/test-import-repo
|
https://api.github.com/repos/snaiperskaya96/test-import-repo
|
opened
|
Delete old 'auto invoice' module
|
Accepted Enhancement High Priority
|
https://trello.com/c/XWeij4td/487-delete-old-auto-invoice-module
This must be done after all of our Brightpearl accounts are fully upgraded to version 4.90.
|
1.0
|
Delete old 'auto invoice' module - https://trello.com/c/XWeij4td/487-delete-old-auto-invoice-module
This must be done after all of our Brightpearl accounts are fully upgraded to version 4.90.
|
priority
|
delete old auto invoice module this must be done after all of our brightpearl accounts are fully upgraded to version
| 1
|
737,245
| 25,507,848,224
|
IssuesEvent
|
2022-11-28 10:48:37
|
opensquare-network/bounties
|
https://api.github.com/repos/opensquare-network/bounties
|
closed
|
Award extrinsic keep broadcasting
|
bug priority:high
|
1. The award button should be disabled if work not submitted
2. The broadcasting toast keep showing after clicking award button when no work not submitted

|
1.0
|
Award extrinsic keep broadcasting - 1. The award button should be disabled if work not submitted
2. The broadcasting toast keep showing after clicking award button when no work not submitted

|
priority
|
award extrinsic keep broadcasting the award button should be disabled if work not submitted the broadcasting toast keep showing after clicking award button when no work not submitted
| 1
|
32,883
| 2,760,832,254
|
IssuesEvent
|
2015-04-28 14:24:24
|
DLR-SC/tixi
|
https://api.github.com/repos/DLR-SC/tixi
|
closed
|
Function for specifying output stream of error and warning messages
|
auto-migrated Milestone-Release2.1.2 Priority-High Type-Enhancement
|
```
Currently, all error and warnings are printed to stdout. This is sometimes
impractical. E.g. we would like to get TiXI Message inside TiGLViewer.
Therefore we have to pipe the messages into the Logging framework of TiGL.
Adding a new logging framework to Tixi is overkill, but we could provide a
function that allows setting a different output stream.
The function could be
tixiSetStandardOut(FILE * outstream)
```
Original issue reported on code.google.com by `martinsi...@gmail.com` on 14 Nov 2013 at 8:40
|
1.0
|
Function for specifying output stream of error and warning messages - ```
Currently, all error and warnings are printed to stdout. This is sometimes
impractical. E.g. we would like to get TiXI Message inside TiGLViewer.
Therefore we have to pipe the messages into the Logging framework of TiGL.
Adding a new logging framework to Tixi is overkill, but we could provide a
function that allows setting a different output stream.
The function could be
tixiSetStandardOut(FILE * outstream)
```
Original issue reported on code.google.com by `martinsi...@gmail.com` on 14 Nov 2013 at 8:40
|
priority
|
function for specifying output stream of error and warning messages currently all error and warnings are printed to stdout this is sometimes impractical e g we would like to get tixi message inside tiglviewer therefore we have to pipe the messages into the logging framework of tigl adding a new logging framework to tixi is overkill but we could provide a function that allows setting a different output stream the function could be tixisetstandardout file outstream original issue reported on code google com by martinsi gmail com on nov at
| 1
|
780,204
| 27,384,570,300
|
IssuesEvent
|
2023-02-28 12:26:12
|
Azure/mec-app-solution-accelerator
|
https://api.github.com/repos/Azure/mec-app-solution-accelerator
|
closed
|
[doc] Create an .MD explaining how to provision a new camera
|
issue P1 (High priority)
|
Add that as an .MD with a link from README.md explaining how to provision a new camera.
|
1.0
|
[doc] Create an .MD explaining how to provision a new camera - Add that as an .MD with a link from README.md explaining how to provision a new camera.
|
priority
|
create an md explaining how to provision a new camera add that as an md with a link from readme md explaining how to provision a new camera
| 1
|
576,783
| 17,094,640,473
|
IssuesEvent
|
2021-07-08 23:14:36
|
CHOMPStation2/CHOMPStation2
|
https://api.github.com/repos/CHOMPStation2/CHOMPStation2
|
closed
|
Mapping support needed for the change to conveyor belts
|
High Priority Map Edit
|
#### Brief description of the issue
Due to https://github.com/CHOMPStation2/CHOMPStation2/pull/2441, conveyor belts set to diagonal directions are pointing in the wrong directions, commonly problematic at the mining base that relies on it to process materials.
#### What you expected to happen
For conveyor belts to work before the PR merge.
#### What actually happened
Conveyor belts are pointing in the wrong direction if they were set to diagonals.
#### Steps to reproduce
- Step 1 - Go to Mining on Sif
- Step 2 - Check conveyor belts
- Step 3 - Laugh/Cry
#### Code Revision
Server revision: B:-Using TGS- D:-Using TGS-
Commit: 7438d0f486dcc60bafc39dd0e8266db5b55194fd
TGS version: 4.11.1
DMAPI version: 5.3.0
#### Anything else you may wish to add:
- This generally affects all conveyor belts mapped in, just blocks mining from their job.
|
1.0
|
Mapping support needed for the change to conveyor belts - #### Brief description of the issue
Due to https://github.com/CHOMPStation2/CHOMPStation2/pull/2441, conveyor belts set to diagonal directions are pointing in the wrong directions, commonly problematic at the mining base that relies on it to process materials.
#### What you expected to happen
For conveyor belts to work before the PR merge.
#### What actually happened
Conveyor belts are pointing in the wrong direction if they were set to diagonals.
#### Steps to reproduce
- Step 1 - Go to Mining on Sif
- Step 2 - Check conveyor belts
- Step 3 - Laugh/Cry
#### Code Revision
Server revision: B:-Using TGS- D:-Using TGS-
Commit: 7438d0f486dcc60bafc39dd0e8266db5b55194fd
TGS version: 4.11.1
DMAPI version: 5.3.0
#### Anything else you may wish to add:
- This generally affects all conveyor belts mapped in, just blocks mining from their job.
|
priority
|
mapping support needed for the change to conveyor belts brief description of the issue due to conveyor belts set to diagonal directions are pointing in the wrong directions commonly problematic at the mining base that relies on it to process materials what you expected to happen for conveyor belts to work before the pr merge what actually happened conveyor belts are pointing in the wrong direction if they were set to diagonals steps to reproduce step go to mining on sif step check conveyor belts step laugh cry code revision server revision b using tgs d using tgs commit tgs version dmapi version anything else you may wish to add this generally affects all conveyor belts mapped in just blocks mining from their job
| 1
|
614,495
| 19,184,245,408
|
IssuesEvent
|
2021-12-04 23:26:27
|
aaronparker/evergreen
|
https://api.github.com/repos/aaronparker/evergreen
|
closed
|
[Bug]: Microsoft.NET Download URL's changed?
|
bug priority:high
|
### What happened?
Microsoft.NET 'LTS' Channel downloads appear to have moved from https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/....
'Current' Channel still seems to be valid for downloads.
If I go to the Microsoft .NET Download site - the new url appears to be:
https://dotnet.microsoft.com/download/dotnet/thank-you/runtime-desktop-6.0.0-windows-x64-installer
### Version
2111.448
### What PowerShell edition/s are you running Evergreen on?
Windows PowerShell
### Which operating system/s are you running Evergreen on?
Windows Server 2016+
### Have you reviewed the documentation?
- [X] Troubleshooting at: https://stealthpuppy.com/evergreen/troubleshoot/
- [X] Known issues at: https://stealthpuppy.com/evergreen/issues/
### Verbose output
```shell
Get-EvergreenApp -Name Microsoft.NET -Verbose
VERBOSE: Get-EvergreenApp: Function exists: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1.
VERBOSE: Get-EvergreenApp: Dot sourcing: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1.
VERBOSE: Get-FunctionResource: read application resource strings from [C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Man
ifests\Microsoft.NET.json]
VERBOSE: Get-EvergreenApp: Calling: Get-Microsoft.NET.
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UserAgent: Mozilla/5.0 (Windows NT; Windows NT 10.0; en-US) AppleWebKit
/534.6 (KHTML, like Gecko) Chrome/7.0.500.0 Safari/534.6].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Method: Default].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [ErrorAction: Continue].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UseBasicParsing: True].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Uri: https://dotnetcli.blob.core.windows.net/dotnet/Runtime/Current/lat
est.version].
VERBOSE: GET https://dotnetcli.blob.core.windows.net/dotnet/Runtime/Current/latest.version with 0-byte payload
VERBOSE: received 6-byte response of content type text/plain
VERBOSE: Invoke-WebRequestWrapper: Response: [200].
VERBOSE: Invoke-WebRequestWrapper: Content type: [text/plain].
VERBOSE: Invoke-WebRequestWrapper: Returning content of length: [6].
VERBOSE: Get-Microsoft.NET: found version: 5.0.12.
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UserAgent: Mozilla/5.0 (Windows NT; Windows NT 10.0; en-US) AppleWebKit
/534.6 (KHTML, like Gecko) Chrome/7.0.500.0 Safari/534.6].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Method: Default].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [ErrorAction: Continue].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UseBasicParsing: True].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Uri: https://dotnetcli.blob.core.windows.net/dotnet/Runtime/LTS/latest.
version].
VERBOSE: GET https://dotnetcli.blob.core.windows.net/dotnet/Runtime/LTS/latest.version with 0-byte payload
VERBOSE: received 5-byte response of content type text/plain
VERBOSE: Invoke-WebRequestWrapper: Response: [200].
VERBOSE: Invoke-WebRequestWrapper: Content type: [text/plain].
VERBOSE: Invoke-WebRequestWrapper: Returning content of length: [5].
VERBOSE: Get-Microsoft.NET: found version: 6.0.0.
VERBOSE: Get-EvergreenApp: Output result from: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1.
Version Architecture Channel URI
------- ------------ ------- ---
6.0.0 x64 LTS https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/windowsdesktop-runtime-6.0.0-win-x64.exe
6.0.0 x86 LTS https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/windowsdesktop-runtime-6.0.0-win-x86.exe
5.0.12 x64 Current https://dotnetcli.blob.core.windows.net/dotnet/WindowsDesktop/5.0.12/windowsdesktop-runtime-5.0.12-win-...
5.0.12 x86 Current https://dotnetcli.blob.core.windows.net/dotnet/WindowsDesktop/5.0.12/windowsdesktop-runtime-5.0.12-win-...
```
|
1.0
|
[Bug]: Microsoft.NET Download URL's changed? - ### What happened?
Microsoft.NET 'LTS' Channel downloads appear to have moved from https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/....
'Current' Channel still seems to be valid for downloads.
If I go to the Microsoft .NET Download site - the new url appears to be:
https://dotnet.microsoft.com/download/dotnet/thank-you/runtime-desktop-6.0.0-windows-x64-installer
### Version
2111.448
### What PowerShell edition/s are you running Evergreen on?
Windows PowerShell
### Which operating system/s are you running Evergreen on?
Windows Server 2016+
### Have you reviewed the documentation?
- [X] Troubleshooting at: https://stealthpuppy.com/evergreen/troubleshoot/
- [X] Known issues at: https://stealthpuppy.com/evergreen/issues/
### Verbose output
```shell
Get-EvergreenApp -Name Microsoft.NET -Verbose
VERBOSE: Get-EvergreenApp: Function exists: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1.
VERBOSE: Get-EvergreenApp: Dot sourcing: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1.
VERBOSE: Get-FunctionResource: read application resource strings from [C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Man
ifests\Microsoft.NET.json]
VERBOSE: Get-EvergreenApp: Calling: Get-Microsoft.NET.
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UserAgent: Mozilla/5.0 (Windows NT; Windows NT 10.0; en-US) AppleWebKit
/534.6 (KHTML, like Gecko) Chrome/7.0.500.0 Safari/534.6].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Method: Default].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [ErrorAction: Continue].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UseBasicParsing: True].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Uri: https://dotnetcli.blob.core.windows.net/dotnet/Runtime/Current/lat
est.version].
VERBOSE: GET https://dotnetcli.blob.core.windows.net/dotnet/Runtime/Current/latest.version with 0-byte payload
VERBOSE: received 6-byte response of content type text/plain
VERBOSE: Invoke-WebRequestWrapper: Response: [200].
VERBOSE: Invoke-WebRequestWrapper: Content type: [text/plain].
VERBOSE: Invoke-WebRequestWrapper: Returning content of length: [6].
VERBOSE: Get-Microsoft.NET: found version: 5.0.12.
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UserAgent: Mozilla/5.0 (Windows NT; Windows NT 10.0; en-US) AppleWebKit
/534.6 (KHTML, like Gecko) Chrome/7.0.500.0 Safari/534.6].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Method: Default].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [ErrorAction: Continue].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [UseBasicParsing: True].
VERBOSE: Invoke-WebRequestWrapper: Invoke-WebRequest parameter: [Uri: https://dotnetcli.blob.core.windows.net/dotnet/Runtime/LTS/latest.
version].
VERBOSE: GET https://dotnetcli.blob.core.windows.net/dotnet/Runtime/LTS/latest.version with 0-byte payload
VERBOSE: received 5-byte response of content type text/plain
VERBOSE: Invoke-WebRequestWrapper: Response: [200].
VERBOSE: Invoke-WebRequestWrapper: Content type: [text/plain].
VERBOSE: Invoke-WebRequestWrapper: Returning content of length: [5].
VERBOSE: Get-Microsoft.NET: found version: 6.0.0.
VERBOSE: Get-EvergreenApp: Output result from: C:\Program Files\WindowsPowerShell\Modules\Evergreen\2111.488\Apps\Get-Microsoft.NET.ps1.
Version Architecture Channel URI
------- ------------ ------- ---
6.0.0 x64 LTS https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/windowsdesktop-runtime-6.0.0-win-x64.exe
6.0.0 x86 LTS https://dotnetcli.azureedge.net/dotnet/Runtime/6.0.0/windowsdesktop-runtime-6.0.0-win-x86.exe
5.0.12 x64 Current https://dotnetcli.blob.core.windows.net/dotnet/WindowsDesktop/5.0.12/windowsdesktop-runtime-5.0.12-win-...
5.0.12 x86 Current https://dotnetcli.blob.core.windows.net/dotnet/WindowsDesktop/5.0.12/windowsdesktop-runtime-5.0.12-win-...
```
|
priority
|
microsoft net download url s changed what happened microsoft net lts channel downloads appear to have moved from current channel still seems to be valid for downloads if i go to the microsoft net download site the new url appears to be version what powershell edition s are you running evergreen on windows powershell which operating system s are you running evergreen on windows server have you reviewed the documentation troubleshooting at known issues at verbose output shell get evergreenapp name microsoft net verbose verbose get evergreenapp function exists c program files windowspowershell modules evergreen apps get microsoft net verbose get evergreenapp dot sourcing c program files windowspowershell modules evergreen apps get microsoft net verbose get functionresource read application resource strings from c program files windowspowershell modules evergreen man ifests microsoft net json verbose get evergreenapp calling get microsoft net verbose invoke webrequestwrapper invoke webrequest parameter useragent mozilla windows nt windows nt en us applewebkit khtml like gecko chrome safari verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter uri est version verbose get with byte payload verbose received byte response of content type text plain verbose invoke webrequestwrapper response verbose invoke webrequestwrapper content type verbose invoke webrequestwrapper returning content of length verbose get microsoft net found version verbose invoke webrequestwrapper invoke webrequest parameter useragent mozilla windows nt windows nt en us applewebkit khtml like gecko chrome safari verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter verbose invoke webrequestwrapper invoke webrequest parameter uri version verbose get with byte payload verbose received byte response of content type text plain verbose invoke webrequestwrapper response verbose invoke webrequestwrapper content type verbose invoke webrequestwrapper returning content of length verbose get microsoft net found version verbose get evergreenapp output result from c program files windowspowershell modules evergreen apps get microsoft net version architecture channel uri lts lts current current
| 1
|
282,606
| 8,708,460,740
|
IssuesEvent
|
2018-12-06 10:58:09
|
pablotabares/decide
|
https://api.github.com/repos/pablotabares/decide
|
closed
|
Add poll creation functionality
|
bot enhancement priority: high
|
Telegram bot must provide commands to create a poll, including questions and answers.
|
1.0
|
Add poll creation functionality - Telegram bot must provide commands to create a poll, including questions and answers.
|
priority
|
add poll creation functionality telegram bot must provide commands to create a poll including questions and answers
| 1
|
635,471
| 20,403,353,866
|
IssuesEvent
|
2022-02-23 00:26:01
|
CoEDL/nyingarn-workspace
|
https://api.github.com/repos/CoEDL/nyingarn-workspace
|
closed
|
Users need to be able to download specific files and delete specific files
|
enhancement priority-high
|
Implement ability to see resource files, download them (e.g. digivol csv) and delete specific files.
|
1.0
|
Users need to be able to download specific files and delete specific files - Implement ability to see resource files, download them (e.g. digivol csv) and delete specific files.
|
priority
|
users need to be able to download specific files and delete specific files implement ability to see resource files download them e g digivol csv and delete specific files
| 1
|
373,325
| 11,042,216,416
|
IssuesEvent
|
2019-12-09 08:40:39
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Issue in inferRecordFieldType method
|
Area/Language Component/Compiler Points/1 Priority/High Type/Improvement
|
**Description:**
In the Types class, we can infer the type of a record field using `inferRecordFieldType` method. `anydata` is not considered in this method, `any` will be returned in such cases.
|
1.0
|
Issue in inferRecordFieldType method - **Description:**
In the Types class, we can infer the type of a record field using `inferRecordFieldType` method. `anydata` is not considered in this method, `any` will be returned in such cases.
|
priority
|
issue in inferrecordfieldtype method description in the types class we can infer the type of a record field using inferrecordfieldtype method anydata is not considered in this method any will be returned in such cases
| 1
|
441,258
| 12,710,053,444
|
IssuesEvent
|
2020-06-23 13:20:29
|
RonAsis/Wsep202
|
https://api.github.com/repos/RonAsis/Wsep202
|
opened
|
bug- not responsive enough - can't appoint all users as owner
|
High priority bug
|
and doesn't tell why- there is no error message
|
1.0
|
bug- not responsive enough - can't appoint all users as owner - and doesn't tell why- there is no error message
|
priority
|
bug not responsive enough can t appoint all users as owner and doesn t tell why there is no error message
| 1
|
243,593
| 7,859,496,551
|
IssuesEvent
|
2018-06-21 16:46:50
|
minio/minio-go
|
https://api.github.com/repos/minio/minio-go
|
closed
|
Error while running Azure tests on Mint
|
priority: high
|
When I was running Azure gateway tests on Mint, I got the following error:
```
{
"args": {
"bucketName": "minio-go-test-mj6c6n45dw0bpa4e",
"objectName": "test-object",
"opts": "",
"size": -1
},
"duration": 289,
"function": "PutObject(bucketName, objectName, reader, size, opts)",
"message": "Expected content-language 'en-US' doesn't match with StatObject return value",
"name": "minio-go: testPutObjectWithContentLanguage",
"status": "FAIL"
}
```
|
1.0
|
Error while running Azure tests on Mint - When I was running Azure gateway tests on Mint, I got the following error:
```
{
"args": {
"bucketName": "minio-go-test-mj6c6n45dw0bpa4e",
"objectName": "test-object",
"opts": "",
"size": -1
},
"duration": 289,
"function": "PutObject(bucketName, objectName, reader, size, opts)",
"message": "Expected content-language 'en-US' doesn't match with StatObject return value",
"name": "minio-go: testPutObjectWithContentLanguage",
"status": "FAIL"
}
```
|
priority
|
error while running azure tests on mint when i was running azure gateway tests on mint i got the following error args bucketname minio go test objectname test object opts size duration function putobject bucketname objectname reader size opts message expected content language en us doesn t match with statobject return value name minio go testputobjectwithcontentlanguage status fail
| 1
|
797,763
| 28,154,687,278
|
IssuesEvent
|
2023-04-03 06:18:02
|
AY2223S2-CS2113-T15-4/tp
|
https://api.github.com/repos/AY2223S2-CS2113-T15-4/tp
|
closed
|
[PE-D][Tester A] There is some logging that is disrupting the user experience when using commands
|
type.Bug priority.High severity.High
|

<!--session: 1680252479707-92304da9-1923-4959-b973-1d1a8a61eabd-->
<!--Version: Web v3.4.7-->
-------------
Labels: `type.FunctionalityBug` `severity.Medium`
original: SSzeWen/ped#1
|
1.0
|
[PE-D][Tester A] There is some logging that is disrupting the user experience when using commands - 
<!--session: 1680252479707-92304da9-1923-4959-b973-1d1a8a61eabd-->
<!--Version: Web v3.4.7-->
-------------
Labels: `type.FunctionalityBug` `severity.Medium`
original: SSzeWen/ped#1
|
priority
|
there is some logging that is disrupting the user experience when using commands labels type functionalitybug severity medium original sszewen ped
| 1
|
581,808
| 17,332,278,413
|
IssuesEvent
|
2021-07-28 05:12:54
|
sacloud/terraform-provider-sakuracloud
|
https://api.github.com/repos/sacloud/terraform-provider-sakuracloud
|
closed
|
外部ツール経由でプラン変更された場合のID変更の追跡
|
area/resources priority/high v2
|
### 概要
さくらのクラウドではプラン変更時にリソースのIDが変更される。
対象リソース:
- サーバ
- ルータ
- ELB
Terraformからプラン変更を行なった場合、IDが変更されることへの対応が実装されているが、
外部ツール(AutoScalerなど)からプラン変更が行われた場合にリソースが追跡できなくなる。
これは、各リソースのReadでIDを元にさくらのクラウドAPIを用いてリソースの情報を参照しているため。
(IDで検索し404が返ってきた場合はリソースが削除されたとみなす)
追跡できなくなってもimportすることで復帰させることが出来るが運用的に煩雑。
このため、何らかのルールにしたがって変更前のIDをメタデータとしてリソースに保持しておき、IDでの検索が404になった場合にはメタデータを利用して検索するようにフォールバックする。
### 実装案
`@previous-id=123456789012`のようなタグで変更前のIDを表す。
Terraform側はまずIDでの検索を試し、404になった場合は同一ゾーンのリソースから`@previous-id=<現在保持しているID>`というタグを条件に検索する。
API呼び出し例: `GET /server?{"Filter":{"Tags.Name":"@previous-id=123456789012"}}`
|
1.0
|
外部ツール経由でプラン変更された場合のID変更の追跡 - ### 概要
さくらのクラウドではプラン変更時にリソースのIDが変更される。
対象リソース:
- サーバ
- ルータ
- ELB
Terraformからプラン変更を行なった場合、IDが変更されることへの対応が実装されているが、
外部ツール(AutoScalerなど)からプラン変更が行われた場合にリソースが追跡できなくなる。
これは、各リソースのReadでIDを元にさくらのクラウドAPIを用いてリソースの情報を参照しているため。
(IDで検索し404が返ってきた場合はリソースが削除されたとみなす)
追跡できなくなってもimportすることで復帰させることが出来るが運用的に煩雑。
このため、何らかのルールにしたがって変更前のIDをメタデータとしてリソースに保持しておき、IDでの検索が404になった場合にはメタデータを利用して検索するようにフォールバックする。
### 実装案
`@previous-id=123456789012`のようなタグで変更前のIDを表す。
Terraform側はまずIDでの検索を試し、404になった場合は同一ゾーンのリソースから`@previous-id=<現在保持しているID>`というタグを条件に検索する。
API呼び出し例: `GET /server?{"Filter":{"Tags.Name":"@previous-id=123456789012"}}`
|
priority
|
外部ツール経由でプラン変更された場合のid変更の追跡 概要 さくらのクラウドではプラン変更時にリソースのidが変更される。 対象リソース サーバ ルータ elb terraformからプラン変更を行なった場合、idが変更されることへの対応が実装されているが、 外部ツール autoscalerなど からプラン変更が行われた場合にリソースが追跡できなくなる。 これは、各リソースのreadでidを元にさくらのクラウドapiを用いてリソースの情報を参照しているため。 追跡できなくなってもimportすることで復帰させることが出来るが運用的に煩雑。 このため、何らかのルールにしたがって変更前のidをメタデータとしてリソースに保持しておき、 。 実装案 previous id のようなタグで変更前のidを表す。 terraform側はまずidでの検索を試し、 previous id というタグを条件に検索する。 api呼び出し例 get server filter tags name previous id
| 1
|
448,881
| 12,959,526,980
|
IssuesEvent
|
2020-07-20 13:08:14
|
wso2/micro-integrator
|
https://api.github.com/repos/wso2/micro-integrator
|
closed
|
Inconsistency in printing logs while using the payloadFactory mediator
|
Priority/High Severity/Major
|
**Description:**
1. Please find the following proxy service.
```
<?xml version="1.0" encoding="UTF-8"?>
<proxy xmlns="http://ws.apache.org/ns/synapse" name="test" startOnLoad="true" statistics="disable" trace="disable" transports="http,https">
<target>
<inSequence>
<payloadFactory media-type="xml">
<format>
<values>Test123</values>
</format>
<args/>
</payloadFactory>
<log level="full"/>
<log>
<property expression="//*[local-name()='values']" name="objects4"/>
</log>
<respond/>
</inSequence>
</target>
<description/>
</proxy>
```
2. Deploy it inside wso2mi-1.1.0 and invoke it.
Expected result
` <values>Test123</values>`
Actual result
`Test123`
**Suggested Labels:**
wso2mi-1.1.0, payloadFactory
**Affected Product Version:**
wso2mi-1.1.0
**OS, DB, other environment details and versions:**
Linnux
**Related Issues:**
https://github.com/wso2/product-ei/issues/2092
|
1.0
|
Inconsistency in printing logs while using the payloadFactory mediator - **Description:**
1. Please find the following proxy service.
```
<?xml version="1.0" encoding="UTF-8"?>
<proxy xmlns="http://ws.apache.org/ns/synapse" name="test" startOnLoad="true" statistics="disable" trace="disable" transports="http,https">
<target>
<inSequence>
<payloadFactory media-type="xml">
<format>
<values>Test123</values>
</format>
<args/>
</payloadFactory>
<log level="full"/>
<log>
<property expression="//*[local-name()='values']" name="objects4"/>
</log>
<respond/>
</inSequence>
</target>
<description/>
</proxy>
```
2. Deploy it inside wso2mi-1.1.0 and invoke it.
Expected result
` <values>Test123</values>`
Actual result
`Test123`
**Suggested Labels:**
wso2mi-1.1.0, payloadFactory
**Affected Product Version:**
wso2mi-1.1.0
**OS, DB, other environment details and versions:**
Linnux
**Related Issues:**
https://github.com/wso2/product-ei/issues/2092
|
priority
|
inconsistency in printing logs while using the payloadfactory mediator description please find the following proxy service deploy it inside and invoke it expected result actual result suggested labels payloadfactory affected product version os db other environment details and versions linnux related issues
| 1
|
185,292
| 6,720,769,436
|
IssuesEvent
|
2017-10-16 09:07:03
|
kedgeproject/kedge
|
https://api.github.com/repos/kedgeproject/kedge
|
closed
|
replicas set to 0 if not specified
|
kind/bug kind/task priority/high
|
If replicas is not set in Kedge file, the generated output shouldn't have it either.
Now it defaults to 0, which leads to a confusing situation where your containers are not started when DC id deployed.
```
name: foo
controller: deploymentconfig
containers:
- image: quay.io/tomkral/sleeper
```
```
▶ ./kedge generate -f test.yaml
---
apiVersion: v1
kind: DeploymentConfig
metadata:
creationTimestamp: null
name: foo
spec:
replicas: 0
strategy:
resources: {}
template:
metadata:
creationTimestamp: null
spec:
containers:
- image: quay.io/tomkral/sleeper
name: foo
resources: {}
test: false
triggers: null
status:
availableReplicas: 0
latestVersion: 0
observedGeneration: 0
replicas: 0
unavailableReplicas: 0
updatedReplicas: 0
```
replicas shouldn't be set if it's not set in Kedge file, OpenShift default will be used on the cluster side.
|
1.0
|
replicas set to 0 if not specified - If replicas is not set in Kedge file, the generated output shouldn't have it either.
Now it defaults to 0, which leads to a confusing situation where your containers are not started when DC id deployed.
```
name: foo
controller: deploymentconfig
containers:
- image: quay.io/tomkral/sleeper
```
```
▶ ./kedge generate -f test.yaml
---
apiVersion: v1
kind: DeploymentConfig
metadata:
creationTimestamp: null
name: foo
spec:
replicas: 0
strategy:
resources: {}
template:
metadata:
creationTimestamp: null
spec:
containers:
- image: quay.io/tomkral/sleeper
name: foo
resources: {}
test: false
triggers: null
status:
availableReplicas: 0
latestVersion: 0
observedGeneration: 0
replicas: 0
unavailableReplicas: 0
updatedReplicas: 0
```
replicas shouldn't be set if it's not set in Kedge file, OpenShift default will be used on the cluster side.
|
priority
|
replicas set to if not specified if replicas is not set in kedge file the generated output shouldn t have it either now it defaults to which leads to a confusing situation where your containers are not started when dc id deployed name foo controller deploymentconfig containers image quay io tomkral sleeper ▶ kedge generate f test yaml apiversion kind deploymentconfig metadata creationtimestamp null name foo spec replicas strategy resources template metadata creationtimestamp null spec containers image quay io tomkral sleeper name foo resources test false triggers null status availablereplicas latestversion observedgeneration replicas unavailablereplicas updatedreplicas replicas shouldn t be set if it s not set in kedge file openshift default will be used on the cluster side
| 1
|
517,364
| 15,007,623,422
|
IssuesEvent
|
2021-01-31 05:52:59
|
Left-on-Read/app
|
https://api.github.com/repos/Left-on-Read/app
|
closed
|
Implement one filter
|
high priority (p1) setup
|
Implement filtering - we can start with just a single filter, such as by date.
~This should be redux-driven.~
|
1.0
|
Implement one filter - Implement filtering - we can start with just a single filter, such as by date.
~This should be redux-driven.~
|
priority
|
implement one filter implement filtering we can start with just a single filter such as by date this should be redux driven
| 1
|
473,405
| 13,641,942,594
|
IssuesEvent
|
2020-09-25 14:50:47
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
CPU memory leak when using torch.no_grad()
|
high priority module: autograd topic: memory usage triage review triaged
|
## 🐛 Bug
If use `torch.no_grad()` block, the cpu memory will continually increase untill OOM kill happens.
But once remove the `no_grad`, everything would be all right.
I tried del loss or put the validation step into a function, but the memory leak still happens.
Is my code wrong? or a BUG?
## To Reproduce
Steps to reproduce the behavior:
1. Validation step using `torch.no_grad()`.
2. Using `CrossEntropyLoss()` as criterion.
3. CPU RAM continually increasing occurs a OOM kill by system(Ubuntu 18.04).
PyTorch 1.6
```
for epoch in range(10):
net.train()
# Good training.
for data in trainloader:
inputs, labels = data['images'], data['masks']
for idx in range(0, len(inputs), 7):
optimizer.zero_grad()
outputs = net(inputs[idx:idx + 7])
loss = criterion(outputs, labels[idx:idx + 7])
loss.backward()
optimizer.step()
# Bad validation.
net.eval()
test_loss = 0.0
test_times = 0
for data in testloader:
# !!!!!!!!👇
with torch.no_grad():
inputs, labels = data['images'], data['masks']
for idx in range(0, len(inputs), 7):
# or put no_grad here, leaking still happens.
outputs = net(inputs[idx:idx + 7])
loss = criterion(outputs, labels[idx:idx + 7])
test_loss += loss.item()
test_times += 1
test_loss /= test_times
```
## Expected behavior
Normally valid without increasing CPU RAM.
## Environment
- PyTorch Version (e.g., 1.0): 1.6
- OS (e.g., Linux): Ubuntu 18.04
- How you installed PyTorch (`conda`, `pip`, source): conda
- Build command you used (if compiling from source): NaN
- Python version: 3.8.5
- CUDA/cuDNN version: 10.2
- GPU models and configuration: Tesla V100 16G
- Any other relevant information: with system RAM 128G
## Additional context
Thanks for all your excellent work!
cc @ezyang @gchanan @zou3519 @albanD @gqchen @pearu @nikitaved
|
1.0
|
CPU memory leak when using torch.no_grad() - ## 🐛 Bug
If use `torch.no_grad()` block, the cpu memory will continually increase untill OOM kill happens.
But once remove the `no_grad`, everything would be all right.
I tried del loss or put the validation step into a function, but the memory leak still happens.
Is my code wrong? or a BUG?
## To Reproduce
Steps to reproduce the behavior:
1. Validation step using `torch.no_grad()`.
2. Using `CrossEntropyLoss()` as criterion.
3. CPU RAM continually increasing occurs a OOM kill by system(Ubuntu 18.04).
PyTorch 1.6
```
for epoch in range(10):
net.train()
# Good training.
for data in trainloader:
inputs, labels = data['images'], data['masks']
for idx in range(0, len(inputs), 7):
optimizer.zero_grad()
outputs = net(inputs[idx:idx + 7])
loss = criterion(outputs, labels[idx:idx + 7])
loss.backward()
optimizer.step()
# Bad validation.
net.eval()
test_loss = 0.0
test_times = 0
for data in testloader:
# !!!!!!!!👇
with torch.no_grad():
inputs, labels = data['images'], data['masks']
for idx in range(0, len(inputs), 7):
# or put no_grad here, leaking still happens.
outputs = net(inputs[idx:idx + 7])
loss = criterion(outputs, labels[idx:idx + 7])
test_loss += loss.item()
test_times += 1
test_loss /= test_times
```
## Expected behavior
Normally valid without increasing CPU RAM.
## Environment
- PyTorch Version (e.g., 1.0): 1.6
- OS (e.g., Linux): Ubuntu 18.04
- How you installed PyTorch (`conda`, `pip`, source): conda
- Build command you used (if compiling from source): NaN
- Python version: 3.8.5
- CUDA/cuDNN version: 10.2
- GPU models and configuration: Tesla V100 16G
- Any other relevant information: with system RAM 128G
## Additional context
Thanks for all your excellent work!
cc @ezyang @gchanan @zou3519 @albanD @gqchen @pearu @nikitaved
|
priority
|
cpu memory leak when using torch no grad 🐛 bug if use torch no grad block the cpu memory will continually increase untill oom kill happens but once remove the no grad everything would be all right i tried del loss or put the validation step into a function but the memory leak still happens is my code wrong or a bug to reproduce steps to reproduce the behavior validation step using torch no grad using crossentropyloss as criterion cpu ram continually increasing occurs a oom kill by system ubuntu pytorch for epoch in range net train good training for data in trainloader inputs labels data data for idx in range len inputs optimizer zero grad outputs net inputs loss criterion outputs labels loss backward optimizer step bad validation net eval test loss test times for data in testloader 👇 with torch no grad inputs labels data data for idx in range len inputs or put no grad here leaking still happens outputs net inputs loss criterion outputs labels test loss loss item test times test loss test times expected behavior normally valid without increasing cpu ram environment pytorch version e g os e g linux ubuntu how you installed pytorch conda pip source conda build command you used if compiling from source nan python version cuda cudnn version gpu models and configuration tesla any other relevant information with system ram additional context thanks for all your excellent work cc ezyang gchanan alband gqchen pearu nikitaved
| 1
|
617,097
| 19,342,641,181
|
IssuesEvent
|
2021-12-15 07:18:14
|
ballerina-platform/ballerina-dev-website
|
https://api.github.com/repos/ballerina-platform/ballerina-dev-website
|
closed
|
Add Content on How to Write a Connector in Bio
|
Priority/Highest Area/Docs Type/Task Points/1
|
**Description:**
Need to update [1] according to the latest Swan Lake changes and add the content to Bio.
[1] https://medium.com/ballerina-techblog/how-to-write-a-client-endpoint-in-ballerina-3c24c185ffaf
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, Browser, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
|
1.0
|
Add Content on How to Write a Connector in Bio - **Description:**
Need to update [1] according to the latest Swan Lake changes and add the content to Bio.
[1] https://medium.com/ballerina-techblog/how-to-write-a-client-endpoint-in-ballerina-3c24c185ffaf
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, Browser, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
|
priority
|
add content on how to write a connector in bio description need to update according to the latest swan lake changes and add the content to bio suggested labels suggested assignees affected product version os browser other environment details and versions steps to reproduce related issues
| 1
|
711,306
| 24,457,641,549
|
IssuesEvent
|
2022-10-07 08:17:32
|
AY2223S1-CS2113-F11-4/tp
|
https://api.github.com/repos/AY2223S1-CS2113-F11-4/tp
|
closed
|
Adding New Prescription Record: add
|
type.Story priority.High
|
As a doctor / user, I want to be able to add a a new prescription for a patient so that I can retrieve it for their future visits and records
<img width="499" alt="image" src="https://user-images.githubusercontent.com/31297758/193443112-74662d47-2998-405d-9315-1da60dec36d5.png">
|
1.0
|
Adding New Prescription Record: add - As a doctor / user, I want to be able to add a a new prescription for a patient so that I can retrieve it for their future visits and records
<img width="499" alt="image" src="https://user-images.githubusercontent.com/31297758/193443112-74662d47-2998-405d-9315-1da60dec36d5.png">
|
priority
|
adding new prescription record add as a doctor user i want to be able to add a a new prescription for a patient so that i can retrieve it for their future visits and records img width alt image src
| 1
|
312,047
| 9,542,320,880
|
IssuesEvent
|
2019-05-01 03:08:45
|
openmsupply/mobile
|
https://api.github.com/repos/openmsupply/mobile
|
opened
|
Customer requisition finalisation crash
|
Bug Effort small Ivory Coast (phase 1) Priority: High
|
Build Number: 2.3.0-rc0 dev+apk
Description:

Reproducible: yes
Reproduction Steps:
1. receive a customer requisition
2. enter a value
3. finalise it
4. some loading spinner and a RSoD (in dev, crash in apk)
Comments: There were items in my customer requisition that my store didn't actually have visible in their store.
|
1.0
|
Customer requisition finalisation crash - Build Number: 2.3.0-rc0 dev+apk
Description:

Reproducible: yes
Reproduction Steps:
1. receive a customer requisition
2. enter a value
3. finalise it
4. some loading spinner and a RSoD (in dev, crash in apk)
Comments: There were items in my customer requisition that my store didn't actually have visible in their store.
|
priority
|
customer requisition finalisation crash build number dev apk description reproducible yes reproduction steps receive a customer requisition enter a value finalise it some loading spinner and a rsod in dev crash in apk comments there were items in my customer requisition that my store didn t actually have visible in their store
| 1
|
460,117
| 13,205,116,210
|
IssuesEvent
|
2020-08-14 17:15:56
|
googleinterns/bazel-rules-fuzzing
|
https://api.github.com/repos/googleinterns/bazel-rules-fuzzing
|
closed
|
Enable regression support in the launcher script
|
high priority
|
## Expected Behavior
Launcher needs to support running regression test without continuous fuzzing test.
To achieve this, a string_flag `engine` will be added to decide the launcher's behavior. After the launcher is modified, a new rule `regression_launcher` will be needed to start the launcher in the regression mode.
## Actual Behavior
Only continuous fuzzing test mode is supported
|
1.0
|
Enable regression support in the launcher script - ## Expected Behavior
Launcher needs to support running regression test without continuous fuzzing test.
To achieve this, a string_flag `engine` will be added to decide the launcher's behavior. After the launcher is modified, a new rule `regression_launcher` will be needed to start the launcher in the regression mode.
## Actual Behavior
Only continuous fuzzing test mode is supported
|
priority
|
enable regression support in the launcher script expected behavior launcher needs to support running regression test without continuous fuzzing test to achieve this a string flag engine will be added to decide the launcher s behavior after the launcher is modified a new rule regression launcher will be needed to start the launcher in the regression mode actual behavior only continuous fuzzing test mode is supported
| 1
|
36,066
| 2,795,249,985
|
IssuesEvent
|
2015-05-11 20:59:18
|
Arabidopsis-Information-Portal/adama
|
https://api.github.com/repos/Arabidopsis-Information-Portal/adama
|
closed
|
Restrict names of namespaces to current allowed characters
|
bug high priority
|
In Adama 0.3 the namespace together with the service name correspond to a container image. The only characters allowed for a container image are [a-z0-9_.-]. This is an implementation detail and it **should not** leak to the user (see #45 and #3). While we fix those issues, Adama should refuse to create a namespace with forbidden characters.
Right now, Adama accepts namespaces with forbidden characters, and it refuses service names with those characters. However, even if the service name is compliant, being under a non-compliant namespace produces an invalid image name.
|
1.0
|
Restrict names of namespaces to current allowed characters - In Adama 0.3 the namespace together with the service name correspond to a container image. The only characters allowed for a container image are [a-z0-9_.-]. This is an implementation detail and it **should not** leak to the user (see #45 and #3). While we fix those issues, Adama should refuse to create a namespace with forbidden characters.
Right now, Adama accepts namespaces with forbidden characters, and it refuses service names with those characters. However, even if the service name is compliant, being under a non-compliant namespace produces an invalid image name.
|
priority
|
restrict names of namespaces to current allowed characters in adama the namespace together with the service name correspond to a container image the only characters allowed for a container image are this is an implementation detail and it should not leak to the user see and while we fix those issues adama should refuse to create a namespace with forbidden characters right now adama accepts namespaces with forbidden characters and it refuses service names with those characters however even if the service name is compliant being under a non compliant namespace produces an invalid image name
| 1
|
744,886
| 25,959,460,957
|
IssuesEvent
|
2022-12-18 17:54:16
|
zigtools/zls
|
https://api.github.com/repos/zigtools/zls
|
closed
|
Integer overflow when parsing incomplete function
|
bug priority:high fuzzing result
|
### Zig Version
0.11.0-dev.782+0b4461d97
### Zig Language Server Version
3526f5fb84b89b6327fc21c6b836b842bf4db90d
### Steps to Reproduce
Open the following file:
```zig
AtEnd() void {
_ = @import("std");
const expect =
try expect(result[1]);
var buf = [_:0]u8{ 'a', 'b'}, 42.0f};
};
});
if (builtin.zig_backend == .stage2_wasm) return error.SkipZigTest; // TODO
try expect(@as(u32, 0xFE00BE00), structPtr.*);
}
pub fn set_checksum_size = 1000;
v.func_field(0) == 2);
try expect(ctz(@as(u121, 3)),
const tmp = ref.*;
\\#define foo "a string";
.is_var_args);
try testOneCtzVector(u128, 64, @splat(4, @as(u24, 0x6a2c48)));
const X = struct {
\\const std = @import("std");
pub fn main() void {
var bad: f128 = 0.0_;
_ = @Type(@typeInfo(Foo).Union.tag_type == f32);
try expect(length == std.mem.eql(u8, messages here is a VarDecl in scope
}
\\0: first arg
\\ return error.SkipZigTest; // TODO
comptime try expect(G3.Fn.return_trace;
return Foo {.x = 13};
}
var some_struct_param_type() void {
_ = v;
};
try testTruncWithVectors() !void {
var target: [*c]u8 = @ptrCast([*]u8, &l.array);
try expect(a == -1);
try S.entry(true, false };
_ = c.printf("0.000000000FFFFFFFF000000000000p+3
.a = .{
S.declaration of label 'blk'
// :8:12: error: union declared here
// :24:20: error: loop in named 'c' in enum 'tmp.Letter'
// :1:7: error: C import block
// :5:10: error: parameters cannot cast into pointer" {
const X = struct field" {
var a: c_int = @ptrCast(*u32, &bytes);
try expect(result == 1) catch unreachable;
}
// error
// backend=stage2
// target=native
//
// :1:8: error: this is a longer message, "integer overflow = @addWithOverflow(u64, a, b, &res));
\\}
try expectEqual = std.testing.expectEqualStrings = std.fs;
pub const B = enum(u8) {
const x: f32 = 0.0;
var default value stored to trigger the bug
//a second vector initialization syntax
// :2:1: error: expectedBitSize = 80;
comptime {
return error union by return" {
const S = extern struct {
const std = @import("behavior = .Pipe;
\\}
});
const i = g + h; // 100
return 1234;
var array = array[i];
fn foo(a: bool, b: bool = true,
}
}
// error
// backend=stage2
// target=native
//
// :2:5: note: control flow inside runtime isNan(nan_times_zero));
.build_modes = true,
};
_ = c;
b1_6: u1,
resume frame;
try expectError(error.FailedToCreateEntry;
defer stdout.print("All your code here, or load and store" {
if (comptime {
const msg = @ptrToInt("Hello, World!\n")),
);
}
fn b() void {
if (true) {} else |err| {
.{ .name = "unsigned int choose[1][1] == 'a');
}
test "double nested unpacked",
\\source.zig:11:5: [address] in foo (test)
}
};
var array1: [4]u8 = "aoeu";
try std.testing.expect;
const expr" {
if (builtin.zig_backend == .stage2_x86_64) return false;
comptime try expect(12 == default initialization syntax
// :2:12: error: expected { 0, 0 }, @as(u64, 0x3ff000000000000000000000000000000000000000000000FFFFFFFFFFFFFFFFFFFFFFFFFF);
try expect(comptime { _ = Foo; }
// error
// backend=stage2
// target=native
//
// :4:30: error: expected type 'anyframe->i32);
try consume_tuple(t2 ++ .{0}, 2);
}
};
.step = b.step("test", write_src = b.addObjectFromWriteFile(src_basename).?);
\\4: "
.fnPtr = bar,
\\pub export fn entry() void {
_ = ignore;
\\int foo();
}
const A = struct {
\\ mov $14, %rdx
fn foo() void {
try expect(mem.eql(u8, message, "integer = u32;
\\pub export fn entry3() callconv(.Async) void {
var array = [_]u8{ .y = 2 };
const zig_args.append(list);
\\ if (builtin.zig_backend == .stage2_sparc64) return x;
_ = stack_trace: ?*std.builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO
var b: i8 = -18;
try foo();
try std.testing.ex
```
### Expected Behavior
It doesn't integer underflow.
### Actual Behavior
```log
thread 16460 panic: integer overflow
C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:2192:61: 0x7ff6a4baee34 in fullCall (zls.exe.obj)
const maybe_async_token = tree.firstToken(info.fn_expr) - 1;
^
C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:1883:24: 0x7ff6a4b3f2af in callOne (zls.exe.obj)
.fn_expr = data.lhs,
^
C:\Programming\Zig\buzz\repos\zls\src\ast.zig:1086:24: 0x7ff6a4a7810a in callFull (zls.exe.obj)
=> tree.callOne(buf, node),
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2964:38: 0x7ff6a49d7864 in makeScopeInternal (zls.exe.obj)
const call = ast.callFull(tree, node_idx, &buf).?;
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3025:64: 0x7ff6a49d8166 in makeScopeInternal (zls.exe.obj)
try makeScopeInternal(allocator, context, field.ast.type_expr);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2560:30: 0x7ff6a4a76a93 in makeInnerScope (zls.exe.obj)
try makeScopeInternal(allocator, context, decl);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2631:31: 0x7ff6a49d4df6 in makeScopeInternal (zls.exe.obj)
try makeInnerScope(allocator, context, node_idx);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2496:33: 0x7ff6a49d41cd in makeDocumentScope (zls.exe.obj)
.enums = &document_scope.enum_completions,
^
C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:612:65: 0x7ff6a4a6493f in createDocument (zls.exe.obj)
var document_scope = try analysis.makeDocumentScope(self.allocator, tree);
^
C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:158:39: 0x7ff6a49d3268 in openDocument (zls.exe.obj)
handle.* = try self.createDocument(duped_uri, duped_text, true);
^
C:\Programming\Zig\buzz\repos\zls\src\Server.zig:1867:111: 0x7ff6a49d2e2f in openDocumentHandler__anon_12308 (zls.exe.obj)
const handle = try server.document_store.openDocument(req.params.textDocument.uri, req.params.textDocument.text);
^
C:\Programming\Zig\buzz\repos\zls\src\Server.zig:2964:35: 0x7ff6a4a36475 in processJsonRpc__anon_10459 (zls.exe.obj)
method_info[2](server, writer, id, request_obj) catch |err| {
^
C:\Programming\Zig\buzz\repos\zls\src\main.zig:51:34: 0x7ff6a4a3d754 in loop (zls.exe.obj)
try server.processJsonRpc(writer, buffer);
^
C:\Programming\Zig\buzz\repos\zls\src\main.zig:281:13: 0x7ff6a4a3dbd2 in main (zls.exe.obj)
try loop(&server);
^
C:\Programming\Zig\zig-from-the-website\lib\std\start.zig:385:41: 0x7ff6a4a3e077 in WinStartup (zls.exe.obj)
std.debug.maybeEnableSegfaultHandler();
^
???:?:?: 0x7ffab1d9559f in ??? (???)
???:?:?: 0x7ffab2c0485a in ??? (???)
```
(By the way this *might be* a Zig issue - shouldn't the parser always error instead of underflowing, or is this just super cursed?)
|
1.0
|
Integer overflow when parsing incomplete function - ### Zig Version
0.11.0-dev.782+0b4461d97
### Zig Language Server Version
3526f5fb84b89b6327fc21c6b836b842bf4db90d
### Steps to Reproduce
Open the following file:
```zig
AtEnd() void {
_ = @import("std");
const expect =
try expect(result[1]);
var buf = [_:0]u8{ 'a', 'b'}, 42.0f};
};
});
if (builtin.zig_backend == .stage2_wasm) return error.SkipZigTest; // TODO
try expect(@as(u32, 0xFE00BE00), structPtr.*);
}
pub fn set_checksum_size = 1000;
v.func_field(0) == 2);
try expect(ctz(@as(u121, 3)),
const tmp = ref.*;
\\#define foo "a string";
.is_var_args);
try testOneCtzVector(u128, 64, @splat(4, @as(u24, 0x6a2c48)));
const X = struct {
\\const std = @import("std");
pub fn main() void {
var bad: f128 = 0.0_;
_ = @Type(@typeInfo(Foo).Union.tag_type == f32);
try expect(length == std.mem.eql(u8, messages here is a VarDecl in scope
}
\\0: first arg
\\ return error.SkipZigTest; // TODO
comptime try expect(G3.Fn.return_trace;
return Foo {.x = 13};
}
var some_struct_param_type() void {
_ = v;
};
try testTruncWithVectors() !void {
var target: [*c]u8 = @ptrCast([*]u8, &l.array);
try expect(a == -1);
try S.entry(true, false };
_ = c.printf("0.000000000FFFFFFFF000000000000p+3
.a = .{
S.declaration of label 'blk'
// :8:12: error: union declared here
// :24:20: error: loop in named 'c' in enum 'tmp.Letter'
// :1:7: error: C import block
// :5:10: error: parameters cannot cast into pointer" {
const X = struct field" {
var a: c_int = @ptrCast(*u32, &bytes);
try expect(result == 1) catch unreachable;
}
// error
// backend=stage2
// target=native
//
// :1:8: error: this is a longer message, "integer overflow = @addWithOverflow(u64, a, b, &res));
\\}
try expectEqual = std.testing.expectEqualStrings = std.fs;
pub const B = enum(u8) {
const x: f32 = 0.0;
var default value stored to trigger the bug
//a second vector initialization syntax
// :2:1: error: expectedBitSize = 80;
comptime {
return error union by return" {
const S = extern struct {
const std = @import("behavior = .Pipe;
\\}
});
const i = g + h; // 100
return 1234;
var array = array[i];
fn foo(a: bool, b: bool = true,
}
}
// error
// backend=stage2
// target=native
//
// :2:5: note: control flow inside runtime isNan(nan_times_zero));
.build_modes = true,
};
_ = c;
b1_6: u1,
resume frame;
try expectError(error.FailedToCreateEntry;
defer stdout.print("All your code here, or load and store" {
if (comptime {
const msg = @ptrToInt("Hello, World!\n")),
);
}
fn b() void {
if (true) {} else |err| {
.{ .name = "unsigned int choose[1][1] == 'a');
}
test "double nested unpacked",
\\source.zig:11:5: [address] in foo (test)
}
};
var array1: [4]u8 = "aoeu";
try std.testing.expect;
const expr" {
if (builtin.zig_backend == .stage2_x86_64) return false;
comptime try expect(12 == default initialization syntax
// :2:12: error: expected { 0, 0 }, @as(u64, 0x3ff000000000000000000000000000000000000000000000FFFFFFFFFFFFFFFFFFFFFFFFFF);
try expect(comptime { _ = Foo; }
// error
// backend=stage2
// target=native
//
// :4:30: error: expected type 'anyframe->i32);
try consume_tuple(t2 ++ .{0}, 2);
}
};
.step = b.step("test", write_src = b.addObjectFromWriteFile(src_basename).?);
\\4: "
.fnPtr = bar,
\\pub export fn entry() void {
_ = ignore;
\\int foo();
}
const A = struct {
\\ mov $14, %rdx
fn foo() void {
try expect(mem.eql(u8, message, "integer = u32;
\\pub export fn entry3() callconv(.Async) void {
var array = [_]u8{ .y = 2 };
const zig_args.append(list);
\\ if (builtin.zig_backend == .stage2_sparc64) return x;
_ = stack_trace: ?*std.builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO
var b: i8 = -18;
try foo();
try std.testing.ex
```
### Expected Behavior
It doesn't integer underflow.
### Actual Behavior
```log
thread 16460 panic: integer overflow
C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:2192:61: 0x7ff6a4baee34 in fullCall (zls.exe.obj)
const maybe_async_token = tree.firstToken(info.fn_expr) - 1;
^
C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:1883:24: 0x7ff6a4b3f2af in callOne (zls.exe.obj)
.fn_expr = data.lhs,
^
C:\Programming\Zig\buzz\repos\zls\src\ast.zig:1086:24: 0x7ff6a4a7810a in callFull (zls.exe.obj)
=> tree.callOne(buf, node),
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2964:38: 0x7ff6a49d7864 in makeScopeInternal (zls.exe.obj)
const call = ast.callFull(tree, node_idx, &buf).?;
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3025:64: 0x7ff6a49d8166 in makeScopeInternal (zls.exe.obj)
try makeScopeInternal(allocator, context, field.ast.type_expr);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2560:30: 0x7ff6a4a76a93 in makeInnerScope (zls.exe.obj)
try makeScopeInternal(allocator, context, decl);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2631:31: 0x7ff6a49d4df6 in makeScopeInternal (zls.exe.obj)
try makeInnerScope(allocator, context, node_idx);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2496:33: 0x7ff6a49d41cd in makeDocumentScope (zls.exe.obj)
.enums = &document_scope.enum_completions,
^
C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:612:65: 0x7ff6a4a6493f in createDocument (zls.exe.obj)
var document_scope = try analysis.makeDocumentScope(self.allocator, tree);
^
C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:158:39: 0x7ff6a49d3268 in openDocument (zls.exe.obj)
handle.* = try self.createDocument(duped_uri, duped_text, true);
^
C:\Programming\Zig\buzz\repos\zls\src\Server.zig:1867:111: 0x7ff6a49d2e2f in openDocumentHandler__anon_12308 (zls.exe.obj)
const handle = try server.document_store.openDocument(req.params.textDocument.uri, req.params.textDocument.text);
^
C:\Programming\Zig\buzz\repos\zls\src\Server.zig:2964:35: 0x7ff6a4a36475 in processJsonRpc__anon_10459 (zls.exe.obj)
method_info[2](server, writer, id, request_obj) catch |err| {
^
C:\Programming\Zig\buzz\repos\zls\src\main.zig:51:34: 0x7ff6a4a3d754 in loop (zls.exe.obj)
try server.processJsonRpc(writer, buffer);
^
C:\Programming\Zig\buzz\repos\zls\src\main.zig:281:13: 0x7ff6a4a3dbd2 in main (zls.exe.obj)
try loop(&server);
^
C:\Programming\Zig\zig-from-the-website\lib\std\start.zig:385:41: 0x7ff6a4a3e077 in WinStartup (zls.exe.obj)
std.debug.maybeEnableSegfaultHandler();
^
???:?:?: 0x7ffab1d9559f in ??? (???)
???:?:?: 0x7ffab2c0485a in ??? (???)
```
(By the way this *might be* a Zig issue - shouldn't the parser always error instead of underflowing, or is this just super cursed?)
|
priority
|
integer overflow when parsing incomplete function zig version dev zig language server version steps to reproduce open the following file zig atend void import std const expect try expect result var buf a b if builtin zig backend wasm return error skipzigtest todo try expect as structptr pub fn set checksum size v func field try expect ctz as const tmp ref define foo a string is var args try testonectzvector splat as const x struct const std import std pub fn main void var bad type typeinfo foo union tag type try expect length std mem eql messages here is a vardecl in scope first arg return error skipzigtest todo comptime try expect fn return trace return foo x var some struct param type void v try testtruncwithvectors void var target ptrcast l array try expect a try s entry true false c printf a s declaration of label blk error union declared here error loop in named c in enum tmp letter error c import block error parameters cannot cast into pointer const x struct field var a c int ptrcast bytes try expect result catch unreachable error backend target native error this is a longer message integer overflow addwithoverflow a b res try expectequal std testing expectequalstrings std fs pub const b enum const x var default value stored to trigger the bug a second vector initialization syntax error expectedbitsize comptime return error union by return const s extern struct const std import behavior pipe const i g h return var array array fn foo a bool b bool true error backend target native note control flow inside runtime isnan nan times zero build modes true c resume frame try expecterror error failedtocreateentry defer stdout print all your code here or load and store if comptime const msg ptrtoint hello world n fn b void if true else err name unsigned int choose a test double nested unpacked source zig in foo test var aoeu try std testing expect const expr if builtin zig backend return false comptime try expect default initialization syntax error expected as try expect comptime foo error backend target native error expected type anyframe try consume tuple step b step test write src b addobjectfromwritefile src basename fnptr bar pub export fn entry void ignore int foo const a struct mov rdx fn foo void try expect mem eql message integer pub export fn callconv async void var array y const zig args append list if builtin zig backend return x stack trace std builtin zig backend return error skipzigtest todo var b try foo try std testing ex expected behavior it doesn t integer underflow actual behavior log thread panic integer overflow c programming zig zig from the website lib std zig ast zig in fullcall zls exe obj const maybe async token tree firsttoken info fn expr c programming zig zig from the website lib std zig ast zig in callone zls exe obj fn expr data lhs c programming zig buzz repos zls src ast zig in callfull zls exe obj tree callone buf node c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj const call ast callfull tree node idx buf c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj try makescopeinternal allocator context field ast type expr c programming zig buzz repos zls src analysis zig in makeinnerscope zls exe obj try makescopeinternal allocator context decl c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj try makeinnerscope allocator context node idx c programming zig buzz repos zls src analysis zig in makedocumentscope zls exe obj enums document scope enum completions c programming zig buzz repos zls src documentstore zig in createdocument zls exe obj var document scope try analysis makedocumentscope self allocator tree c programming zig buzz repos zls src documentstore zig in opendocument zls exe obj handle try self createdocument duped uri duped text true c programming zig buzz repos zls src server zig in opendocumenthandler anon zls exe obj const handle try server document store opendocument req params textdocument uri req params textdocument text c programming zig buzz repos zls src server zig in processjsonrpc anon zls exe obj method info server writer id request obj catch err c programming zig buzz repos zls src main zig in loop zls exe obj try server processjsonrpc writer buffer c programming zig buzz repos zls src main zig in main zls exe obj try loop server c programming zig zig from the website lib std start zig in winstartup zls exe obj std debug maybeenablesegfaulthandler in in by the way this might be a zig issue shouldn t the parser always error instead of underflowing or is this just super cursed
| 1
|
823,311
| 30,989,685,286
|
IssuesEvent
|
2023-08-09 02:47:47
|
Karooobar/Voyager
|
https://api.github.com/repos/Karooobar/Voyager
|
closed
|
Need to pass storeId, owner ID, from the user who is logged in
|
High Priority
|
Right now in helpers like ItemHelper, CategoryHelper, and storeHelper, we are hardcoding the store id that is 200.
We need to modify this so that we use the store id of the user that is logged in.
|
1.0
|
Need to pass storeId, owner ID, from the user who is logged in - Right now in helpers like ItemHelper, CategoryHelper, and storeHelper, we are hardcoding the store id that is 200.
We need to modify this so that we use the store id of the user that is logged in.
|
priority
|
need to pass storeid owner id from the user who is logged in right now in helpers like itemhelper categoryhelper and storehelper we are hardcoding the store id that is we need to modify this so that we use the store id of the user that is logged in
| 1
|
629,928
| 20,071,470,861
|
IssuesEvent
|
2022-02-04 07:33:07
|
debops/debops
|
https://api.github.com/repos/debops/debops
|
closed
|
bootstrap-sssd and bootstrap-ldap don't seem to be idempotent
|
bug priority: high tag: LDAP
|
Both ``nslcd`` and ``sssd`` contain configuration file generation tasks which include directives like:
```
- name: Generate nslcd configuration
...
when: nslcd__ldap_base_dn|d()
```
and:
```
- name: Generate sssd configuration
...
when: sssd__ldap_base_dn|d()
```
The ``nslcd__ldap_base_dn`` variables and ``sssd__ldap_base_dn`` variables are both defined as:
`'{{ ansible_local.ldap.base_dn|d([]) }}'`
In the ``bootstrap-*`` case, ``ansible_local.ldap.base_dn`` is initially undefined, and later set by the ``debops.ldap`` role.
However, the ``sssd__ldap_base_dn`` and ``nslcd__ldap_base_dn`` variables aren't recalculated, so the configuration files ``/etc/sssd/sssd.conf`` and ``/etc/nslcd.conf`` files aren't generated (in the latter case, a default file is generated by the DPKG package, so the fact the file is present is misleading).
I tried copying the trick from ``debops.ldap``:
```
- name: Take note of the current LDAP configuration
set_fact:
# Track the changes in the configuration state
# between role executions in the same play.
ldap__fact_configured: '{{ ldap__configured }}'
# Re-instantiate dependent variables to evaluate variables that use them.
# Without this, dependent variables may contain outdated configuration.
ldap__fact_dependent_tasks: '{{ ldap__dependent_tasks }}'
tags: [ 'role::ldap:tasks', 'skip::ldap:tasks' ]
```
To update ``sssd__ldap_base_dn`` and ``nslcd__ldap_base_dn``, but that doesn't seem to actually recalculate them. A simple workaround is to run the ``bootstrap-*`` scripts twice for a host...where the second invocation will show:
```
PLAY RECAP ***********************************************************************************
example : ok=200 changed=2 unreachable=0 failed=0 skipped=138 rescued=0 ignored=0
```
(Note the ``changed`` value)
Would be happy to do a PR, but I'm kind of stumped here (tried on ``2.10.7+merged+base+2.10.8+dfsg-1`` from Debian unstable).
|
1.0
|
bootstrap-sssd and bootstrap-ldap don't seem to be idempotent - Both ``nslcd`` and ``sssd`` contain configuration file generation tasks which include directives like:
```
- name: Generate nslcd configuration
...
when: nslcd__ldap_base_dn|d()
```
and:
```
- name: Generate sssd configuration
...
when: sssd__ldap_base_dn|d()
```
The ``nslcd__ldap_base_dn`` variables and ``sssd__ldap_base_dn`` variables are both defined as:
`'{{ ansible_local.ldap.base_dn|d([]) }}'`
In the ``bootstrap-*`` case, ``ansible_local.ldap.base_dn`` is initially undefined, and later set by the ``debops.ldap`` role.
However, the ``sssd__ldap_base_dn`` and ``nslcd__ldap_base_dn`` variables aren't recalculated, so the configuration files ``/etc/sssd/sssd.conf`` and ``/etc/nslcd.conf`` files aren't generated (in the latter case, a default file is generated by the DPKG package, so the fact the file is present is misleading).
I tried copying the trick from ``debops.ldap``:
```
- name: Take note of the current LDAP configuration
set_fact:
# Track the changes in the configuration state
# between role executions in the same play.
ldap__fact_configured: '{{ ldap__configured }}'
# Re-instantiate dependent variables to evaluate variables that use them.
# Without this, dependent variables may contain outdated configuration.
ldap__fact_dependent_tasks: '{{ ldap__dependent_tasks }}'
tags: [ 'role::ldap:tasks', 'skip::ldap:tasks' ]
```
To update ``sssd__ldap_base_dn`` and ``nslcd__ldap_base_dn``, but that doesn't seem to actually recalculate them. A simple workaround is to run the ``bootstrap-*`` scripts twice for a host...where the second invocation will show:
```
PLAY RECAP ***********************************************************************************
example : ok=200 changed=2 unreachable=0 failed=0 skipped=138 rescued=0 ignored=0
```
(Note the ``changed`` value)
Would be happy to do a PR, but I'm kind of stumped here (tried on ``2.10.7+merged+base+2.10.8+dfsg-1`` from Debian unstable).
|
priority
|
bootstrap sssd and bootstrap ldap don t seem to be idempotent both nslcd and sssd contain configuration file generation tasks which include directives like name generate nslcd configuration when nslcd ldap base dn d and name generate sssd configuration when sssd ldap base dn d the nslcd ldap base dn variables and sssd ldap base dn variables are both defined as ansible local ldap base dn d in the bootstrap case ansible local ldap base dn is initially undefined and later set by the debops ldap role however the sssd ldap base dn and nslcd ldap base dn variables aren t recalculated so the configuration files etc sssd sssd conf and etc nslcd conf files aren t generated in the latter case a default file is generated by the dpkg package so the fact the file is present is misleading i tried copying the trick from debops ldap name take note of the current ldap configuration set fact track the changes in the configuration state between role executions in the same play ldap fact configured ldap configured re instantiate dependent variables to evaluate variables that use them without this dependent variables may contain outdated configuration ldap fact dependent tasks ldap dependent tasks tags to update sssd ldap base dn and nslcd ldap base dn but that doesn t seem to actually recalculate them a simple workaround is to run the bootstrap scripts twice for a host where the second invocation will show play recap example ok changed unreachable failed skipped rescued ignored note the changed value would be happy to do a pr but i m kind of stumped here tried on merged base dfsg from debian unstable
| 1
|
343,656
| 10,334,532,770
|
IssuesEvent
|
2019-09-03 08:34:34
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
opened
|
[master-preview] Broken Treasuary with new Government system
|
High Priority
|
1. You can place Treasury and change Tax without rights.
2. Elected Title with access to Treasury doesn't give you access to Treasury

3. It's in another table, but It's directly connected to Treasury: The description in the elected title that you can set the tax. But I think it’s need to describe not only the tax but also Treasury itself.

4. Same trouble at 8.3.0: You can set tax more then 100%. Need to limit it.

|
1.0
|
[master-preview] Broken Treasuary with new Government system - 1. You can place Treasury and change Tax without rights.
2. Elected Title with access to Treasury doesn't give you access to Treasury

3. It's in another table, but It's directly connected to Treasury: The description in the elected title that you can set the tax. But I think it’s need to describe not only the tax but also Treasury itself.

4. Same trouble at 8.3.0: You can set tax more then 100%. Need to limit it.

|
priority
|
broken treasuary with new government system you can place treasury and change tax without rights elected title with access to treasury doesn t give you access to treasury it s in another table but it s directly connected to treasury the description in the elected title that you can set the tax but i think it’s need to describe not only the tax but also treasury itself same trouble at you can set tax more then need to limit it
| 1
|
471,168
| 13,561,468,572
|
IssuesEvent
|
2020-09-18 04:38:40
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
IDP, Add authentication configuration fail
|
Affected/5.11.0-m31 Component/Identity Apps Priority/High Severity/Critical bug console dev-exp
|
**Describe the issue:**
Following check box at IDP , authentication configuration wizard check box are not clickable
1. passive STS authentication configuration wizard check boxes
2. SAML authentication configuration wizard check boxes
console error:
```
vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 Uncaught TypeError: Cannot read property 'toString' of undefined
at Object.onChange (main.babc4f7b.chunk.js?ef057d352ba2713628dc:1)
at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2
at t.handleChange (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2
at Object.s (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at d (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2
```
**How to reproduce:**
1. login to console
2. Navigate to the developer page and identity provider
3. Click to add and and new IDP
4. Click to edit IDP and select Authentication tab
5. Select passive STS the Authenticator and click next
6. click check boxes of these ,
-Passive STS user ID location
-Enable SAML assertion signature validation
- Enable SAML assertion audience validation
**Expected behavior:**
Check box not select as expected
**Environment information**
- Product Version: [e.g., IS 5.11.0-m31]
|
1.0
|
IDP, Add authentication configuration fail - **Describe the issue:**
Following check box at IDP , authentication configuration wizard check box are not clickable
1. passive STS authentication configuration wizard check boxes
2. SAML authentication configuration wizard check boxes
console error:
```
vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2 Uncaught TypeError: Cannot read property 'toString' of undefined
at Object.onChange (main.babc4f7b.chunk.js?ef057d352ba2713628dc:1)
at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at e.exports (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2
at t.handleChange (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2
at Object.s (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at d (vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2)
at vendors~main.22da41b3.chunk.js?ef057d352ba2713628dc:2
```
**How to reproduce:**
1. login to console
2. Navigate to the developer page and identity provider
3. Click to add and and new IDP
4. Click to edit IDP and select Authentication tab
5. Select passive STS the Authenticator and click next
6. click check boxes of these ,
-Passive STS user ID location
-Enable SAML assertion signature validation
- Enable SAML assertion audience validation
**Expected behavior:**
Check box not select as expected
**Environment information**
- Product Version: [e.g., IS 5.11.0-m31]
|
priority
|
idp add authentication configuration fail describe the issue following check box at idp authentication configuration wizard check box are not clickable passive sts authentication configuration wizard check boxes saml authentication configuration wizard check boxes console error vendors main chunk js uncaught typeerror cannot read property tostring of undefined at object onchange main chunk js at e exports vendors main chunk js at e exports vendors main chunk js at e exports vendors main chunk js at vendors main chunk js at t handlechange vendors main chunk js at vendors main chunk js at object s vendors main chunk js at d vendors main chunk js at vendors main chunk js how to reproduce login to console navigate to the developer page and identity provider click to add and and new idp click to edit idp and select authentication tab select passive sts the authenticator and click next click check boxes of these passive sts user id location enable saml assertion signature validation enable saml assertion audience validation expected behavior check box not select as expected environment information product version
| 1
|
405,339
| 11,871,802,750
|
IssuesEvent
|
2020-03-26 14:54:58
|
kenodressel/quarantine-hero
|
https://api.github.com/repos/kenodressel/quarantine-hero
|
closed
|
Reactivate Google GeoLocation API
|
enhancement inprogress priority-high
|
- [ ] Migrate Database (plz --> geolocations)
- [ ] Verify that the website still works when we flip the feature flag
- [ ] Verify that the firebase functions still works when we flip the feature flag
- [ ] make the repo backwards compatible so that we can always go back to plz with out migration
The backwards compatibility is enforce by only allowing users to input a plz in the "askForHelp" form and the "offer-help" form.
|
1.0
|
Reactivate Google GeoLocation API - - [ ] Migrate Database (plz --> geolocations)
- [ ] Verify that the website still works when we flip the feature flag
- [ ] Verify that the firebase functions still works when we flip the feature flag
- [ ] make the repo backwards compatible so that we can always go back to plz with out migration
The backwards compatibility is enforce by only allowing users to input a plz in the "askForHelp" form and the "offer-help" form.
|
priority
|
reactivate google geolocation api migrate database plz geolocations verify that the website still works when we flip the feature flag verify that the firebase functions still works when we flip the feature flag make the repo backwards compatible so that we can always go back to plz with out migration the backwards compatibility is enforce by only allowing users to input a plz in the askforhelp form and the offer help form
| 1
|
576,032
| 17,069,359,417
|
IssuesEvent
|
2021-07-07 11:23:28
|
ParamagicDev/mrujs
|
https://api.github.com/repos/ParamagicDev/mrujs
|
closed
|
Feature: Turbo integration
|
Priority: High Priority: Normal Project Size: Medium Type: Feature
|
#### Describe Problem
currently Turbo + Mrujs have quite a bit of intersection. I'm not even sure if this is a good idea tbh. This is more so a "future possibility"
there will be a lot of trial and error to make this happen. So this issue will serve as a future reference of possible changes that need to be made.
#### Suggest Solution
first change: modify querySelectors.
https://github.com/rails/rails/commit/8de181dac638457564677e02c71a9d04fd3e7fde
|
2.0
|
Feature: Turbo integration - #### Describe Problem
currently Turbo + Mrujs have quite a bit of intersection. I'm not even sure if this is a good idea tbh. This is more so a "future possibility"
there will be a lot of trial and error to make this happen. So this issue will serve as a future reference of possible changes that need to be made.
#### Suggest Solution
first change: modify querySelectors.
https://github.com/rails/rails/commit/8de181dac638457564677e02c71a9d04fd3e7fde
|
priority
|
feature turbo integration describe problem currently turbo mrujs have quite a bit of intersection i m not even sure if this is a good idea tbh this is more so a future possibility there will be a lot of trial and error to make this happen so this issue will serve as a future reference of possible changes that need to be made suggest solution first change modify queryselectors
| 1
|
594,469
| 18,046,341,463
|
IssuesEvent
|
2021-09-19 00:32:13
|
practice-enterprise/api
|
https://api.github.com/repos/practice-enterprise/api
|
opened
|
All requests to DB needs to lowered
|
Priority: High Type: Enhancement
|
- [ ] Adjust polling
- [ ] Lower requests to DB drastically
- [ ] Announcements, user calls
- [ ] ...
|
1.0
|
All requests to DB needs to lowered - - [ ] Adjust polling
- [ ] Lower requests to DB drastically
- [ ] Announcements, user calls
- [ ] ...
|
priority
|
all requests to db needs to lowered adjust polling lower requests to db drastically announcements user calls
| 1
|
343,070
| 10,325,021,752
|
IssuesEvent
|
2019-09-01 14:09:33
|
Eve-ning/vsrg
|
https://api.github.com/repos/Eve-ning/vsrg
|
closed
|
Use directories instead of filters
|
aesthetics high-priority
|
Getting messy with game specific implementations, it's either I file them properly or use another repo
|
1.0
|
Use directories instead of filters - Getting messy with game specific implementations, it's either I file them properly or use another repo
|
priority
|
use directories instead of filters getting messy with game specific implementations it s either i file them properly or use another repo
| 1
|
480,030
| 13,822,378,502
|
IssuesEvent
|
2020-10-13 04:56:03
|
Kedyn/fusliez-notes
|
https://api.github.com/repos/Kedyn/fusliez-notes
|
closed
|
Tree shake the font awesome library
|
Priority: High Status: Accepted Type: Maintenance
|
@fortawesome/free-solid-svg-icons/index.es.js is a very big file. Tree shake it down to the handful of icons we use in the project.
|
1.0
|
Tree shake the font awesome library - @fortawesome/free-solid-svg-icons/index.es.js is a very big file. Tree shake it down to the handful of icons we use in the project.
|
priority
|
tree shake the font awesome library fortawesome free solid svg icons index es js is a very big file tree shake it down to the handful of icons we use in the project
| 1
|
44,786
| 2,914,446,389
|
IssuesEvent
|
2015-06-23 05:12:38
|
CameronProbert/VUW-SWEN303-Assignment4
|
https://api.github.com/repos/CameronProbert/VUW-SWEN303-Assignment4
|
closed
|
Add stats for courts page
|
TODO High Priority
|
Add stats as text for overall/per season and with a team selected and without
|
1.0
|
Add stats for courts page - Add stats as text for overall/per season and with a team selected and without
|
priority
|
add stats for courts page add stats as text for overall per season and with a team selected and without
| 1
|
428,102
| 12,403,100,943
|
IssuesEvent
|
2020-05-21 13:19:47
|
MoonchildProductions/UXP
|
https://api.github.com/repos/MoonchildProductions/UXP
|
closed
|
Fieldset disabled no longer works
|
Assigned C: DOM C: Layout High Priority Regression
|
Regression 28.8.4 -> 28.9.0, still broken in 28.9.3
Potentially a regression from the html5 parser work done? Marking as parser component for now.
http://jsbin.com/bibiqi/1/edit?html,css,output -- test input fields should all have a red background in the disabled fieldset.
|
1.0
|
Fieldset disabled no longer works - Regression 28.8.4 -> 28.9.0, still broken in 28.9.3
Potentially a regression from the html5 parser work done? Marking as parser component for now.
http://jsbin.com/bibiqi/1/edit?html,css,output -- test input fields should all have a red background in the disabled fieldset.
|
priority
|
fieldset disabled no longer works regression still broken in potentially a regression from the parser work done marking as parser component for now test input fields should all have a red background in the disabled fieldset
| 1
|
388,565
| 11,489,266,531
|
IssuesEvent
|
2020-02-11 15:14:55
|
level73/membernet
|
https://api.github.com/repos/level73/membernet
|
closed
|
MANAGE USERS tab not appearing when hovering over
|
Platform: Membernet Priority: High Status: Wontfix Type: Bug
|
Frequently, I am unable to access the manage users tab, it does not load when hovering.
|
1.0
|
MANAGE USERS tab not appearing when hovering over - Frequently, I am unable to access the manage users tab, it does not load when hovering.
|
priority
|
manage users tab not appearing when hovering over frequently i am unable to access the manage users tab it does not load when hovering
| 1
|
110,576
| 4,434,766,224
|
IssuesEvent
|
2016-08-18 04:59:29
|
OpenSRP/opensrp-client
|
https://api.github.com/repos/OpenSRP/opensrp-client
|
opened
|
Migrate Bidan app drishti codebase to OpenSRP codebase
|
High Priority INDONESIA
|
Currently Bidan app using old dristhi codebase.
|
1.0
|
Migrate Bidan app drishti codebase to OpenSRP codebase - Currently Bidan app using old dristhi codebase.
|
priority
|
migrate bidan app drishti codebase to opensrp codebase currently bidan app using old dristhi codebase
| 1
|
238,141
| 7,775,278,206
|
IssuesEvent
|
2018-06-05 01:50:52
|
ESGF/esgf-compute-wps
|
https://api.github.com/repos/ESGF/esgf-compute-wps
|
opened
|
Expose regrid and domain as a per operation setting in workflow
|
feature priority/high
|
**Is your feature request related to a problem? Please describe.**
You cannot set the regrid or domain for individual operations in a workflow.
**Describe the solution you'd like**
When you configure an operation you should be able to set the domain or regrid.
**Describe alternatives you've considered**
n/a
**Additional context**
n/a
|
1.0
|
Expose regrid and domain as a per operation setting in workflow - **Is your feature request related to a problem? Please describe.**
You cannot set the regrid or domain for individual operations in a workflow.
**Describe the solution you'd like**
When you configure an operation you should be able to set the domain or regrid.
**Describe alternatives you've considered**
n/a
**Additional context**
n/a
|
priority
|
expose regrid and domain as a per operation setting in workflow is your feature request related to a problem please describe you cannot set the regrid or domain for individual operations in a workflow describe the solution you d like when you configure an operation you should be able to set the domain or regrid describe alternatives you ve considered n a additional context n a
| 1
|
336,209
| 10,173,198,972
|
IssuesEvent
|
2019-08-08 12:36:04
|
ecwinc/pact-server
|
https://api.github.com/repos/ecwinc/pact-server
|
closed
|
Easy way to upload data
|
high priority
|
Idea: type in some admin password into browser page and get redirected to "upload" page where one can upload geoJSON files, sending their content to the database. Will be necessary once we have all the new data from Jae (uploading the raw JSON from a script will take a long time)
|
1.0
|
Easy way to upload data - Idea: type in some admin password into browser page and get redirected to "upload" page where one can upload geoJSON files, sending their content to the database. Will be necessary once we have all the new data from Jae (uploading the raw JSON from a script will take a long time)
|
priority
|
easy way to upload data idea type in some admin password into browser page and get redirected to upload page where one can upload geojson files sending their content to the database will be necessary once we have all the new data from jae uploading the raw json from a script will take a long time
| 1
|
175,592
| 6,552,460,010
|
IssuesEvent
|
2017-09-05 18:26:38
|
vtyulb/BSA-Analytics
|
https://api.github.com/repos/vtyulb/BSA-Analytics
|
closed
|
не работает precise search
|
High priority
|
После обновления программы precise search выдает ошибку на любых файлах как на сервере, так и на локальном компьютере
|
1.0
|
не работает precise search - После обновления программы precise search выдает ошибку на любых файлах как на сервере, так и на локальном компьютере
|
priority
|
не работает precise search после обновления программы precise search выдает ошибку на любых файлах как на сервере так и на локальном компьютере
| 1
|
281,839
| 8,700,292,993
|
IssuesEvent
|
2018-12-05 08:15:16
|
AICrowd/ai-crowd-3
|
https://api.github.com/repos/AICrowd/ai-crowd-3
|
closed
|
500 after middle button clicking on a video in leaderboard
|
June 2018 bug high priority
|
_From @AdamStelmaszczyk on July 06, 2017 13:35_
After middle button clicking (which opens a link in a new tab) on any video image on [leaderboard page](https://www.crowdai.org/challenges/nips-2017-learning-to-run/leaderboards), 500 "We're sorry, but something went wrong" page always comes up, [example](https://www.crowdai.org/challenges/nips-2017-learning-to-run/leaderboards/video_modal?submission_id=493).
Left button clicking works (because it only opens the modal window, no URL change).
_Copied from original issue: crowdAI/crowdai#243_
|
1.0
|
500 after middle button clicking on a video in leaderboard - _From @AdamStelmaszczyk on July 06, 2017 13:35_
After middle button clicking (which opens a link in a new tab) on any video image on [leaderboard page](https://www.crowdai.org/challenges/nips-2017-learning-to-run/leaderboards), 500 "We're sorry, but something went wrong" page always comes up, [example](https://www.crowdai.org/challenges/nips-2017-learning-to-run/leaderboards/video_modal?submission_id=493).
Left button clicking works (because it only opens the modal window, no URL change).
_Copied from original issue: crowdAI/crowdai#243_
|
priority
|
after middle button clicking on a video in leaderboard from adamstelmaszczyk on july after middle button clicking which opens a link in a new tab on any video image on we re sorry but something went wrong page always comes up left button clicking works because it only opens the modal window no url change copied from original issue crowdai crowdai
| 1
|
745,456
| 25,985,017,621
|
IssuesEvent
|
2022-12-19 22:52:09
|
gmlc-dispatches/dispatches
|
https://api.github.com/repos/gmlc-dispatches/dispatches
|
closed
|
Test failures in RE DoubleLoopOptimization.ipynb: ValueError: Error evaluating Param value (UnitOnT0State[309_WIND_1])
|
Priority:High
|
This occurred with #164, possibly as a consequence of updating to IDAES 2.0.0b2:
```
ValueError: Error evaluating Param value (UnitOnT0State[309_WIND_1]):
The Param value is currently set to an invalid value. This is
typically from a scalar Param or mutable Indexed Param without
an initial or default value.
```
See https://github.com/gmlc-dispatches/dispatches/actions/runs/3678468960/jobs/6221778238#step:5:385 for more details.
|
1.0
|
Test failures in RE DoubleLoopOptimization.ipynb: ValueError: Error evaluating Param value (UnitOnT0State[309_WIND_1]) - This occurred with #164, possibly as a consequence of updating to IDAES 2.0.0b2:
```
ValueError: Error evaluating Param value (UnitOnT0State[309_WIND_1]):
The Param value is currently set to an invalid value. This is
typically from a scalar Param or mutable Indexed Param without
an initial or default value.
```
See https://github.com/gmlc-dispatches/dispatches/actions/runs/3678468960/jobs/6221778238#step:5:385 for more details.
|
priority
|
test failures in re doubleloopoptimization ipynb valueerror error evaluating param value this occurred with possibly as a consequence of updating to idaes valueerror error evaluating param value the param value is currently set to an invalid value this is typically from a scalar param or mutable indexed param without an initial or default value see for more details
| 1
|
701,193
| 24,090,101,591
|
IssuesEvent
|
2022-09-19 14:08:46
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
Modal Component: CSS Blur filter causing performance issues
|
[Type] Bug [Priority] High General Interface [Type] Performance [Feature] Component System
|
### Description
When trying to use any modal with the new blurred design using Gutenberg 14.0.2 and WordPress 6.0.2, there's noticeable lag. This happens in the Site Editor and Post Editor when doing things like: opening pattern explorer, opening template part inserter, and opening preferences. I can't tell if this is related to the new design or something else entirely so let me know if the title needs to be updated :)
### Step-by-step reproduction instructions
1. Open the Site Editor.
2. Open the Inserter, go to the patterns tab, and hit "explore". Notice lag.
3. Close out of that experience.
4. Open preferences from the three dot menu. Notice lag.
5. Close out of that experience.
6. Add a template part block and choose the option to select from a current template.
7. See modal pop up and notice lag.
### Screenshots, screen recording, code snippet
https://user-images.githubusercontent.com/26996883/188512881-4e06f49a-be65-4f3c-b8cf-690c612e670c.mov
### Environment info
- WordPress 6.0.2
- TT2 and Tove (two block themes)
- Gutenberg 14.0.2
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
Yes
|
1.0
|
Modal Component: CSS Blur filter causing performance issues - ### Description
When trying to use any modal with the new blurred design using Gutenberg 14.0.2 and WordPress 6.0.2, there's noticeable lag. This happens in the Site Editor and Post Editor when doing things like: opening pattern explorer, opening template part inserter, and opening preferences. I can't tell if this is related to the new design or something else entirely so let me know if the title needs to be updated :)
### Step-by-step reproduction instructions
1. Open the Site Editor.
2. Open the Inserter, go to the patterns tab, and hit "explore". Notice lag.
3. Close out of that experience.
4. Open preferences from the three dot menu. Notice lag.
5. Close out of that experience.
6. Add a template part block and choose the option to select from a current template.
7. See modal pop up and notice lag.
### Screenshots, screen recording, code snippet
https://user-images.githubusercontent.com/26996883/188512881-4e06f49a-be65-4f3c-b8cf-690c612e670c.mov
### Environment info
- WordPress 6.0.2
- TT2 and Tove (two block themes)
- Gutenberg 14.0.2
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
Yes
|
priority
|
modal component css blur filter causing performance issues description when trying to use any modal with the new blurred design using gutenberg and wordpress there s noticeable lag this happens in the site editor and post editor when doing things like opening pattern explorer opening template part inserter and opening preferences i can t tell if this is related to the new design or something else entirely so let me know if the title needs to be updated step by step reproduction instructions open the site editor open the inserter go to the patterns tab and hit explore notice lag close out of that experience open preferences from the three dot menu notice lag close out of that experience add a template part block and choose the option to select from a current template see modal pop up and notice lag screenshots screen recording code snippet environment info wordpress and tove two block themes gutenberg please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg yes
| 1
|
356,856
| 10,598,338,286
|
IssuesEvent
|
2019-10-10 04:26:26
|
harmony-one/harmony
|
https://api.github.com/repos/harmony-one/harmony
|
closed
|
Joining testnet fails
|
bug high priority
|
**Describe the bug**
Running a node on AWS to join testnet fails
**To Reproduce**
Log on to the AWS Insance and run
```
curl -O https://raw.githubusercontent.com/harmony-one/harmony/master/scripts/node.sh
chmod +x node.sh
sudo ./node.sh
```
**Expected behavior**
Node joins succesfully
**Environment (please complete the following information):**
- AWS EC2 instance
**Additional context**
Below is the error log
```
INFO [04-03|17:20:36.685] [SYNC] Current Block port=9000 ip=34.222.46.182 blockHex=0x811737e250857a0faaacefcd94e066ef01a8adede7873d8cd3a8104edd291a56
ERROR[04-03|17:20:36.690] cannot create group sig mask: key not found port=9000 ip=34.222.46.182
DEBUG[04-03|17:20:36.697] Trie cache stats after commit misses=0 unloads=0
DEBUG[04-03|17:20:36.697] Inserted new block number=1 hash=6b1e9e…3c63d4 uncles=0 txs=3 gas=1454996 elapsed=11.641ms
INFO [04-03|17:20:36.697] Imported new chain segment blocks=1 txs=3 mgas=1.455 elapsed=11.893ms mgasps=122.336 number=1 hash=6b1e9e…3c63d4 age=1d12h51m cache=13.37kB
INFO [04-03|17:20:36.697] [SYNC] new block added to blockchain port=9000 ip=34.222.46.182 blockHeight=1 blockHex=0x6b1e9e06ce095a10acfd44fd51accaa7b1d10489c3e30890f8404a30093c63d4
INFO [04-03|17:20:36.698] [SYNC] Current Block port=9000 ip=34.222.46.182 blockHex=0x6b1e9e06ce095a10acfd44fd51accaa7b1d10489c3e30890f8404a30093c63d4
ERROR[04-03|17:20:36.699] cannot create group sig mask: key not found port=9000 ip=34.222.46.182
ERROR[04-03|17:20:36.700]
########## BAD BLOCK #########
Chain config: {ChainID: 0 Homestead: 0 DAO: <nil> DAOSupport: false EIP150: 0 EIP155: 0 EIP158: 0 Byzantium: 0 Constantinople: 0 ConstantinopleFix: 0 Engine: ethash}
Number: 2
Hash: 0x10e4c3caa8562e8867adab7bb46c48fcc34b069dc94e98de7466aff188d3e4be
&{[] 1 77048 [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0] [] [129 146 152 85 228 159 49 19 1 173 88 245 109 97 14 71 99 244 250 58 146 204 164 231 41 159 233 211 35 78 3 228] [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0] 77048}
Error: invalid merkle root (remote: 237abd2e63ae0cf8a2bc1801d3b183d83d599c8d57da3522a7f45339f21e8d46 local: 25dd7137dc20eb62813dc812fb7a900f4eb6f616c174b184a7a2cb6af53dc930)
##############################
DEBUG[04-03|17:20:36.700] Error adding new block to blockchain port=9000 ip=34.222.46.182 Error="invalid merkle root (remote: 237abd2e63ae0cf8a2bc1801d3b183d83d599c8d57da3522a7f45339f21e8d46 local: 25dd7137dc20eb62813dc812fb7a900f4eb6f616c174b184a7a2cb6af53dc930)"
INFO [04-03|17:20:36.704] [SYNC] clean up inactive peers port=9000 ip=34.222.46.182 activeNumber=7
```
|
1.0
|
Joining testnet fails - **Describe the bug**
Running a node on AWS to join testnet fails
**To Reproduce**
Log on to the AWS Insance and run
```
curl -O https://raw.githubusercontent.com/harmony-one/harmony/master/scripts/node.sh
chmod +x node.sh
sudo ./node.sh
```
**Expected behavior**
Node joins succesfully
**Environment (please complete the following information):**
- AWS EC2 instance
**Additional context**
Below is the error log
```
INFO [04-03|17:20:36.685] [SYNC] Current Block port=9000 ip=34.222.46.182 blockHex=0x811737e250857a0faaacefcd94e066ef01a8adede7873d8cd3a8104edd291a56
ERROR[04-03|17:20:36.690] cannot create group sig mask: key not found port=9000 ip=34.222.46.182
DEBUG[04-03|17:20:36.697] Trie cache stats after commit misses=0 unloads=0
DEBUG[04-03|17:20:36.697] Inserted new block number=1 hash=6b1e9e…3c63d4 uncles=0 txs=3 gas=1454996 elapsed=11.641ms
INFO [04-03|17:20:36.697] Imported new chain segment blocks=1 txs=3 mgas=1.455 elapsed=11.893ms mgasps=122.336 number=1 hash=6b1e9e…3c63d4 age=1d12h51m cache=13.37kB
INFO [04-03|17:20:36.697] [SYNC] new block added to blockchain port=9000 ip=34.222.46.182 blockHeight=1 blockHex=0x6b1e9e06ce095a10acfd44fd51accaa7b1d10489c3e30890f8404a30093c63d4
INFO [04-03|17:20:36.698] [SYNC] Current Block port=9000 ip=34.222.46.182 blockHex=0x6b1e9e06ce095a10acfd44fd51accaa7b1d10489c3e30890f8404a30093c63d4
ERROR[04-03|17:20:36.699] cannot create group sig mask: key not found port=9000 ip=34.222.46.182
ERROR[04-03|17:20:36.700]
########## BAD BLOCK #########
Chain config: {ChainID: 0 Homestead: 0 DAO: <nil> DAOSupport: false EIP150: 0 EIP155: 0 EIP158: 0 Byzantium: 0 Constantinople: 0 ConstantinopleFix: 0 Engine: ethash}
Number: 2
Hash: 0x10e4c3caa8562e8867adab7bb46c48fcc34b069dc94e98de7466aff188d3e4be
&{[] 1 77048 [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0] [] [129 146 152 85 228 159 49 19 1 173 88 245 109 97 14 71 99 244 250 58 146 204 164 231 41 159 233 211 35 78 3 228] [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0] 77048}
Error: invalid merkle root (remote: 237abd2e63ae0cf8a2bc1801d3b183d83d599c8d57da3522a7f45339f21e8d46 local: 25dd7137dc20eb62813dc812fb7a900f4eb6f616c174b184a7a2cb6af53dc930)
##############################
DEBUG[04-03|17:20:36.700] Error adding new block to blockchain port=9000 ip=34.222.46.182 Error="invalid merkle root (remote: 237abd2e63ae0cf8a2bc1801d3b183d83d599c8d57da3522a7f45339f21e8d46 local: 25dd7137dc20eb62813dc812fb7a900f4eb6f616c174b184a7a2cb6af53dc930)"
INFO [04-03|17:20:36.704] [SYNC] clean up inactive peers port=9000 ip=34.222.46.182 activeNumber=7
```
|
priority
|
joining testnet fails describe the bug running a node on aws to join testnet fails to reproduce log on to the aws insance and run curl o chmod x node sh sudo node sh expected behavior node joins succesfully environment please complete the following information aws instance additional context below is the error log info current block port ip blockhex error cannot create group sig mask key not found port ip debug trie cache stats after commit misses unloads debug inserted new block number hash … uncles txs gas elapsed info imported new chain segment blocks txs mgas elapsed mgasps number hash … age cache info new block added to blockchain port ip blockheight blockhex info current block port ip blockhex error cannot create group sig mask key not found port ip error bad block chain config chainid homestead dao daosupport false byzantium constantinople constantinoplefix engine ethash number hash error invalid merkle root remote local debug error adding new block to blockchain port ip error invalid merkle root remote local info clean up inactive peers port ip activenumber
| 1
|
519,040
| 15,039,098,965
|
IssuesEvent
|
2021-02-02 18:13:46
|
qutebrowser/qutebrowser
|
https://api.github.com/repos/qutebrowser/qutebrowser
|
closed
|
content.cookies.store=false stopped working with 2.0
|
priority: 0 - high status: needs triage
|
**Version info**:
qutebrowser v2.0.1
Git commit:
Backend: QtWebEngine (Chromium 83.0.4103.122)
Qt: 5.15.2
**Does the bug happen if you start with `--temp-basedir`?**:
Needs a session.
**Description**
Cookies are stored with `content.cookies.store` set to `false`.
**How to reproduce**
Set `content.cookies.store` to false, login some, e.g. GitHub. Restart. Still logged in.
|
1.0
|
content.cookies.store=false stopped working with 2.0 - **Version info**:
qutebrowser v2.0.1
Git commit:
Backend: QtWebEngine (Chromium 83.0.4103.122)
Qt: 5.15.2
**Does the bug happen if you start with `--temp-basedir`?**:
Needs a session.
**Description**
Cookies are stored with `content.cookies.store` set to `false`.
**How to reproduce**
Set `content.cookies.store` to false, login some, e.g. GitHub. Restart. Still logged in.
|
priority
|
content cookies store false stopped working with version info qutebrowser git commit backend qtwebengine chromium qt does the bug happen if you start with temp basedir needs a session description cookies are stored with content cookies store set to false how to reproduce set content cookies store to false login some e g github restart still logged in
| 1
|
381,802
| 11,288,245,649
|
IssuesEvent
|
2020-01-16 07:21:26
|
cvfadmin/COVE
|
https://api.github.com/repos/cvfadmin/COVE
|
closed
|
Pagination breaks when user zooms in our out of page.
|
bug high priority
|
cmd +/cmd -
**Proposed Solution: Switch to per page pagination instead of infinite scroll**
|
1.0
|
Pagination breaks when user zooms in our out of page. - cmd +/cmd -
**Proposed Solution: Switch to per page pagination instead of infinite scroll**
|
priority
|
pagination breaks when user zooms in our out of page cmd cmd proposed solution switch to per page pagination instead of infinite scroll
| 1
|
161,153
| 6,110,016,318
|
IssuesEvent
|
2017-06-21 14:15:42
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
closed
|
[studio] Delete of a published page not working
|
bug Priority: Highest!
|
Using Editorial bp:
1. Create a new temp landing page (e.g. Health 2).
2. Publish the new page.
3. Delete the page.
Result: the page is never deleted. Same thing happens if you edit and existing page, publish and then try to delete it.
|
1.0
|
[studio] Delete of a published page not working - Using Editorial bp:
1. Create a new temp landing page (e.g. Health 2).
2. Publish the new page.
3. Delete the page.
Result: the page is never deleted. Same thing happens if you edit and existing page, publish and then try to delete it.
|
priority
|
delete of a published page not working using editorial bp create a new temp landing page e g health publish the new page delete the page result the page is never deleted same thing happens if you edit and existing page publish and then try to delete it
| 1
|
364,656
| 10,771,904,273
|
IssuesEvent
|
2019-11-02 11:14:08
|
AY1920S1-CS2103T-T11-1/main
|
https://api.github.com/repos/AY1920S1-CS2103T-T11-1/main
|
closed
|
Change how Price field is parsed
|
priority.High
|
Make Price field not accept 0
Improve error message for negative price values
Block more than 2d.p. for price inputs
Related to #212 #170
|
1.0
|
Change how Price field is parsed - Make Price field not accept 0
Improve error message for negative price values
Block more than 2d.p. for price inputs
Related to #212 #170
|
priority
|
change how price field is parsed make price field not accept improve error message for negative price values block more than p for price inputs related to
| 1
|
522,141
| 15,158,052,413
|
IssuesEvent
|
2021-02-12 00:13:26
|
NOAA-GSL/MATS
|
https://api.github.com/repos/NOAA-GSL/MATS
|
opened
|
Add plot timeline button/capability to MATS timeseries
|
Priority: High Project: MATS Type: Feature
|
---
Author Name: **molly.b.smith** (@mollybsmith-noaa)
Original Redmine Issue: 56360, https://vlab.ncep.noaa.gov/redmine/issues/56360
Original Date: 2018-10-16
Original Assignee: randy.pierce
---
We would like to add the capability to show what times have data for a particular set of parameters in MATS.
|
1.0
|
Add plot timeline button/capability to MATS timeseries - ---
Author Name: **molly.b.smith** (@mollybsmith-noaa)
Original Redmine Issue: 56360, https://vlab.ncep.noaa.gov/redmine/issues/56360
Original Date: 2018-10-16
Original Assignee: randy.pierce
---
We would like to add the capability to show what times have data for a particular set of parameters in MATS.
|
priority
|
add plot timeline button capability to mats timeseries author name molly b smith mollybsmith noaa original redmine issue original date original assignee randy pierce we would like to add the capability to show what times have data for a particular set of parameters in mats
| 1
|
627,920
| 19,956,823,230
|
IssuesEvent
|
2022-01-28 00:54:24
|
BlockySurvival/issue-tracker
|
https://api.github.com/repos/BlockySurvival/issue-tracker
|
closed
|
players can create new accounts with empty passwords
|
bug high priority security
|
empty passwords are a disaster. the setting is correct, so this is one more thing multicraft is ignoring for unknown reasons.
|
1.0
|
players can create new accounts with empty passwords - empty passwords are a disaster. the setting is correct, so this is one more thing multicraft is ignoring for unknown reasons.
|
priority
|
players can create new accounts with empty passwords empty passwords are a disaster the setting is correct so this is one more thing multicraft is ignoring for unknown reasons
| 1
|
416,717
| 12,150,817,387
|
IssuesEvent
|
2020-04-24 18:42:00
|
woocommerce/woocommerce-android
|
https://api.github.com/repos/woocommerce/woocommerce-android
|
closed
|
4.1-rc-1: Crash in login discovery view
|
Priority: High [Type] Crash
|
Sentry Issue: [WOOCOMMERCE-ANDROID-GY](https://sentry.io/organizations/a8c/issues/1627191543/?referrer=github_integration)
```
ClassCastException: com.google.android.material.appbar.AppBarLayout cannot be cast to androidx.appcompat.widget.Toolbar
at com.woocommerce.android.ui.login.LoginDiscoveryErrorFragment.onViewCreated(LoginDiscoveryErrorFragment.kt:90)
...
(14 additional frame(s) were not displayed)
com.google.android.material.appbar.AppBarLayout cannot be cast to androidx.appcompat.widget.Toolbar
```
|
1.0
|
4.1-rc-1: Crash in login discovery view - Sentry Issue: [WOOCOMMERCE-ANDROID-GY](https://sentry.io/organizations/a8c/issues/1627191543/?referrer=github_integration)
```
ClassCastException: com.google.android.material.appbar.AppBarLayout cannot be cast to androidx.appcompat.widget.Toolbar
at com.woocommerce.android.ui.login.LoginDiscoveryErrorFragment.onViewCreated(LoginDiscoveryErrorFragment.kt:90)
...
(14 additional frame(s) were not displayed)
com.google.android.material.appbar.AppBarLayout cannot be cast to androidx.appcompat.widget.Toolbar
```
|
priority
|
rc crash in login discovery view sentry issue classcastexception com google android material appbar appbarlayout cannot be cast to androidx appcompat widget toolbar at com woocommerce android ui login logindiscoveryerrorfragment onviewcreated logindiscoveryerrorfragment kt additional frame s were not displayed com google android material appbar appbarlayout cannot be cast to androidx appcompat widget toolbar
| 1
|
291,516
| 8,926,230,710
|
IssuesEvent
|
2019-01-22 03:08:34
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
Master: "plan research" tutorial broken
|
High Priority
|
1) checking "show all" not counts, maybe refers to https://github.com/StrangeLoopGames/EcoIssues/issues/10702
2) "mouse over" - no tooltip, no complete :(


|
1.0
|
Master: "plan research" tutorial broken - 1) checking "show all" not counts, maybe refers to https://github.com/StrangeLoopGames/EcoIssues/issues/10702
2) "mouse over" - no tooltip, no complete :(


|
priority
|
master plan research tutorial broken checking show all not counts maybe refers to mouse over no tooltip no complete
| 1
|
145,340
| 5,565,128,486
|
IssuesEvent
|
2017-03-26 11:25:32
|
metasfresh/metasfresh
|
https://api.github.com/repos/metasfresh/metasfresh
|
opened
|
Manufacturing Order UI
|
enhancement high priority
|
### Is this a bug or feature request?
Feature Request
### What is the current behavior?
Currently, there is no Manufacturing Order UI Window in metasfresh.
### What is the expected or desired behavior?
Create a generic window w/ Grid View that lists all nonprocessed Manufacturing Orders.
This Window is the starting point for https://github.com/metasfresh/metasfresh-webui-frontend/issues/580.
|
1.0
|
Manufacturing Order UI - ### Is this a bug or feature request?
Feature Request
### What is the current behavior?
Currently, there is no Manufacturing Order UI Window in metasfresh.
### What is the expected or desired behavior?
Create a generic window w/ Grid View that lists all nonprocessed Manufacturing Orders.
This Window is the starting point for https://github.com/metasfresh/metasfresh-webui-frontend/issues/580.
|
priority
|
manufacturing order ui is this a bug or feature request feature request what is the current behavior currently there is no manufacturing order ui window in metasfresh what is the expected or desired behavior create a generic window w grid view that lists all nonprocessed manufacturing orders this window is the starting point for
| 1
|
258,032
| 8,150,301,950
|
IssuesEvent
|
2018-08-22 12:34:51
|
metasfresh/metasfresh
|
https://api.github.com/repos/metasfresh/metasfresh
|
closed
|
Support Doctype Text templates also on purchase order
|
branch:master priority:high type:enhancement
|
### Is this a bug or feature request?
FR
### What is the current behavior?
metasfresh supports putting header and footer text tempates into sales order and invoice
http://docs.metasfresh.org/webui_collection/EN/Print_text_on_documents-doctype.html
if you try this for purchase order the text template will be reapplied on completion.
case:
1. configure header and footer text for doctype purchase order
1. create a new purchase order with lines
1. OK: text is copied to fields in purchase order
1. edit the text and put something else
1. OK: text is displayed
1. complete the order
1. NOK: text is set back to default
### What is the expected or desired behavior?
we do want to support that also in purchase order
todo:
- change that behaviour
- create an automatic test
|
1.0
|
Support Doctype Text templates also on purchase order - ### Is this a bug or feature request?
FR
### What is the current behavior?
metasfresh supports putting header and footer text tempates into sales order and invoice
http://docs.metasfresh.org/webui_collection/EN/Print_text_on_documents-doctype.html
if you try this for purchase order the text template will be reapplied on completion.
case:
1. configure header and footer text for doctype purchase order
1. create a new purchase order with lines
1. OK: text is copied to fields in purchase order
1. edit the text and put something else
1. OK: text is displayed
1. complete the order
1. NOK: text is set back to default
### What is the expected or desired behavior?
we do want to support that also in purchase order
todo:
- change that behaviour
- create an automatic test
|
priority
|
support doctype text templates also on purchase order is this a bug or feature request fr what is the current behavior metasfresh supports putting header and footer text tempates into sales order and invoice if you try this for purchase order the text template will be reapplied on completion case configure header and footer text for doctype purchase order create a new purchase order with lines ok text is copied to fields in purchase order edit the text and put something else ok text is displayed complete the order nok text is set back to default what is the expected or desired behavior we do want to support that also in purchase order todo change that behaviour create an automatic test
| 1
|
194,985
| 6,901,413,083
|
IssuesEvent
|
2017-11-25 06:55:30
|
metasfresh/metasfresh-webui-frontend
|
https://api.github.com/repos/metasfresh/metasfresh-webui-frontend
|
closed
|
Navigation Menu Favorite stars wrong
|
branch:master priority:high type:bug
|
### Is this a bug or feature request?
Bug
### What is the current behavior?
The Bookmark Favorites in Navigation menu are behaving weirdly.
#### Which are the steps to reproduce?
1. Open the Navigation Menu and hover over an entry. Move to grey star to bookmark it. Star disappears. NOK
2. Move over an entry until you see a dark grey star and move out to the right. A light grey star remains. NOK
3. Open a Window, then alt+1. Next to window name try to bookmark the window there. Does not work. NOK.
### What is the expected or desired behavior?
Please fix.


|
1.0
|
Navigation Menu Favorite stars wrong - ### Is this a bug or feature request?
Bug
### What is the current behavior?
The Bookmark Favorites in Navigation menu are behaving weirdly.
#### Which are the steps to reproduce?
1. Open the Navigation Menu and hover over an entry. Move to grey star to bookmark it. Star disappears. NOK
2. Move over an entry until you see a dark grey star and move out to the right. A light grey star remains. NOK
3. Open a Window, then alt+1. Next to window name try to bookmark the window there. Does not work. NOK.
### What is the expected or desired behavior?
Please fix.


|
priority
|
navigation menu favorite stars wrong is this a bug or feature request bug what is the current behavior the bookmark favorites in navigation menu are behaving weirdly which are the steps to reproduce open the navigation menu and hover over an entry move to grey star to bookmark it star disappears nok move over an entry until you see a dark grey star and move out to the right a light grey star remains nok open a window then alt next to window name try to bookmark the window there does not work nok what is the expected or desired behavior please fix
| 1
|
417,814
| 12,179,634,853
|
IssuesEvent
|
2020-04-28 11:02:06
|
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
|
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
|
closed
|
OG troops on religion change
|
:beetle: bug :beetle: :exclamation: priority high
|
<!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Your mod version is:**
78a3f6f458e30148cfec06e599d14843de3f69aa
**What expansions do you have installed?**
All
**Are you using any submods/mods? If so, which?**
No
**Please explain your issue in as much detail as possible:**
Reported by @KoiKarp
> you should lose armies and generals requested from old gods if you convert, atm you keep them after coverting.
**Steps to reproduce the issue:**
Get army from old gods, convert to other religion.
**Upload an attachment below: .zip of your save, or screenshots:**
|
1.0
|
OG troops on religion change - <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Your mod version is:**
78a3f6f458e30148cfec06e599d14843de3f69aa
**What expansions do you have installed?**
All
**Are you using any submods/mods? If so, which?**
No
**Please explain your issue in as much detail as possible:**
Reported by @KoiKarp
> you should lose armies and generals requested from old gods if you convert, atm you keep them after coverting.
**Steps to reproduce the issue:**
Get army from old gods, convert to other religion.
**Upload an attachment below: .zip of your save, or screenshots:**
|
priority
|
og troops on religion change do not remove pre existing lines your mod version is what expansions do you have installed all are you using any submods mods if so which no please explain your issue in as much detail as possible reported by koikarp you should lose armies and generals requested from old gods if you convert atm you keep them after coverting steps to reproduce the issue get army from old gods convert to other religion upload an attachment below zip of your save or screenshots
| 1
|
154,621
| 5,923,178,999
|
IssuesEvent
|
2017-05-23 07:09:32
|
openxc/openxc-ios-framework
|
https://api.github.com/repos/openxc/openxc-ios-framework
|
closed
|
Command support
|
Priority:High Status:In-progress Type:Enhancement
|
framework currently only supports "device_id" and "version" commands. Need to add the remaining from https://github.com/openxc/openxc-message-format/blob/master/JSON.mkd#commands. Some of these may come along with protobuf support.
|
1.0
|
Command support - framework currently only supports "device_id" and "version" commands. Need to add the remaining from https://github.com/openxc/openxc-message-format/blob/master/JSON.mkd#commands. Some of these may come along with protobuf support.
|
priority
|
command support framework currently only supports device id and version commands need to add the remaining from some of these may come along with protobuf support
| 1
|
504,812
| 14,621,309,770
|
IssuesEvent
|
2020-12-22 21:23:25
|
nfdi4plants/Swate
|
https://api.github.com/repos/nfdi4plants/Swate
|
closed
|
Provide validation information via XML metadata
|
Area: Excel Priority: High Type: Enhancement Type: Interop
|
It is possible to set metadata of cells via office.js. Especially for users that create templates for others to use, information on the proposed cell contents is helpful. This would be set via a JSON string ~~in the XML tag~~ with the key "validationParams" and should have the following fields:
- **importance** - an integer between 0 and 100
- **content type** - one of : ontologyTerm, text, number, int, decimal, url, boolean
- if the content type is ontologyTerm, asdd the isA restriction if set: term[restriction]
- [x] Create a JSON schema for this
- [x] Create a form to generate this for a column
_Edit: Office.js only supports custom tags, not custom attributes on existing tags. Therefore, we need to investigate alternative places to save this information. For now we can add the validation table to the app model and later use that data to generate the file/tags/etc we need._
_Edit2(by @Freymaurer): As of the latest decision all swate validation data will now be saved in the excel custom xml, as desribed [here](https://github.com/nfdi4plants/Swate/issues/45#issuecomment-733794910)._
- [x] Add validation data to the app model.
- [x] Swate can write validation data into excel custom xml.
|
1.0
|
Provide validation information via XML metadata - It is possible to set metadata of cells via office.js. Especially for users that create templates for others to use, information on the proposed cell contents is helpful. This would be set via a JSON string ~~in the XML tag~~ with the key "validationParams" and should have the following fields:
- **importance** - an integer between 0 and 100
- **content type** - one of : ontologyTerm, text, number, int, decimal, url, boolean
- if the content type is ontologyTerm, asdd the isA restriction if set: term[restriction]
- [x] Create a JSON schema for this
- [x] Create a form to generate this for a column
_Edit: Office.js only supports custom tags, not custom attributes on existing tags. Therefore, we need to investigate alternative places to save this information. For now we can add the validation table to the app model and later use that data to generate the file/tags/etc we need._
_Edit2(by @Freymaurer): As of the latest decision all swate validation data will now be saved in the excel custom xml, as desribed [here](https://github.com/nfdi4plants/Swate/issues/45#issuecomment-733794910)._
- [x] Add validation data to the app model.
- [x] Swate can write validation data into excel custom xml.
|
priority
|
provide validation information via xml metadata it is possible to set metadata of cells via office js especially for users that create templates for others to use information on the proposed cell contents is helpful this would be set via a json string in the xml tag with the key validationparams and should have the following fields importance an integer between and content type one of ontologyterm text number int decimal url boolean if the content type is ontologyterm asdd the isa restriction if set term create a json schema for this create a form to generate this for a column edit office js only supports custom tags not custom attributes on existing tags therefore we need to investigate alternative places to save this information for now we can add the validation table to the app model and later use that data to generate the file tags etc we need by freymaurer as of the latest decision all swate validation data will now be saved in the excel custom xml as desribed add validation data to the app model swate can write validation data into excel custom xml
| 1
|
439,950
| 12,690,975,678
|
IssuesEvent
|
2020-06-21 14:50:10
|
sodafoundation/SIM
|
https://api.github.com/repos/sodafoundation/SIM
|
closed
|
Create Virtual environment for deploying Dolphin
|
Feature High Priority
|
<!-- This form is for bug reports and feature requests! -->
**Is this a BUG REPORT or FEATURE REQUEST?**:
/ feature
**What happened**:
Need to run dolphin in virtual environment as to to keep dependencies required by project by creating isolated python virtual environments.
**What you expected to happen**:
Dolphin should work independtly even if there are other pyhton projects which uses ddifferent modules.
**How to reproduce it (as minimally and precisely as possible)**:
NA
**Anything else we need to know?**:
NA
**Environment**:
- Dolphin(release/branch) version:
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
|
1.0
|
Create Virtual environment for deploying Dolphin - <!-- This form is for bug reports and feature requests! -->
**Is this a BUG REPORT or FEATURE REQUEST?**:
/ feature
**What happened**:
Need to run dolphin in virtual environment as to to keep dependencies required by project by creating isolated python virtual environments.
**What you expected to happen**:
Dolphin should work independtly even if there are other pyhton projects which uses ddifferent modules.
**How to reproduce it (as minimally and precisely as possible)**:
NA
**Anything else we need to know?**:
NA
**Environment**:
- Dolphin(release/branch) version:
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
|
priority
|
create virtual environment for deploying dolphin is this a bug report or feature request feature what happened need to run dolphin in virtual environment as to to keep dependencies required by project by creating isolated python virtual environments what you expected to happen dolphin should work independtly even if there are other pyhton projects which uses ddifferent modules how to reproduce it as minimally and precisely as possible na anything else we need to know na environment dolphin release branch version os e g from etc os release kernel e g uname a install tools others
| 1
|
551,234
| 16,165,221,570
|
IssuesEvent
|
2021-05-01 10:51:06
|
sopra-fs21-group-01/server
|
https://api.github.com/repos/sopra-fs21-group-01/server
|
closed
|
Implement Exceptions
|
high priority
|
## acceptance criteria:
-no default exceptions occurs during login/registration
## estimated time: 1h
|
1.0
|
Implement Exceptions - ## acceptance criteria:
-no default exceptions occurs during login/registration
## estimated time: 1h
|
priority
|
implement exceptions acceptance criteria no default exceptions occurs during login registration estimated time
| 1
|
148,448
| 5,682,304,805
|
IssuesEvent
|
2017-04-13 09:20:14
|
openxpki/openxpki
|
https://api.github.com/repos/openxpki/openxpki
|
closed
|
Add useful alias name to chain certificates in Java KeyStore
|
feature request high priority
|
At least some versions of keytool add the ISSUER name as alias to a chain certificate which leads to confusion. Proposed solution is to add the subject (CN or full DN) to all certificates in the store.
|
1.0
|
Add useful alias name to chain certificates in Java KeyStore - At least some versions of keytool add the ISSUER name as alias to a chain certificate which leads to confusion. Proposed solution is to add the subject (CN or full DN) to all certificates in the store.
|
priority
|
add useful alias name to chain certificates in java keystore at least some versions of keytool add the issuer name as alias to a chain certificate which leads to confusion proposed solution is to add the subject cn or full dn to all certificates in the store
| 1
|
633,644
| 20,261,129,441
|
IssuesEvent
|
2022-02-15 07:33:46
|
rtCamp/rtMedia
|
https://api.github.com/repos/rtCamp/rtMedia
|
closed
|
Blank media page is visible in Twenty Twenty-Two theme
|
bug Priority/High
|
```
Environment :
Tested on pre-release Branch
WordPress Version: 5.9
BuddyPress Version: 10.0.0
rtMedia Version: 4.6.9
theme - Twenty Twenty-Two
```
The blank media page is displayed in the new theme. But on others theme page is rendered properly.
Note: no console error and warning are displayed. Issue observed on both legacy and Nouveau template
https://user-images.githubusercontent.com/44057535/151926551-1beac4ed-61f5-4adc-9054-960985334c92.mp4
|
1.0
|
Blank media page is visible in Twenty Twenty-Two theme - ```
Environment :
Tested on pre-release Branch
WordPress Version: 5.9
BuddyPress Version: 10.0.0
rtMedia Version: 4.6.9
theme - Twenty Twenty-Two
```
The blank media page is displayed in the new theme. But on others theme page is rendered properly.
Note: no console error and warning are displayed. Issue observed on both legacy and Nouveau template
https://user-images.githubusercontent.com/44057535/151926551-1beac4ed-61f5-4adc-9054-960985334c92.mp4
|
priority
|
blank media page is visible in twenty twenty two theme environment tested on pre release branch wordpress version buddypress version rtmedia version theme twenty twenty two the blank media page is displayed in the new theme but on others theme page is rendered properly note no console error and warning are displayed issue observed on both legacy and nouveau template
| 1
|
719,686
| 24,767,036,250
|
IssuesEvent
|
2022-10-22 17:05:48
|
ufosc/Club_Website_2
|
https://api.github.com/repos/ufosc/Club_Website_2
|
closed
|
Set up EJS linting
|
good first issue help wanted high priority difficulty: easy
|
**Todo:** Set up an EJS linter. This is an easy way to automatically style all our html/EJS code and provide more meaningful error messages for EJS (the ejs module is designed to be minimal so they don't serve detailed errors without a linter).
Here is the module: https://www.npmjs.com/package/ejs-lint
You'll then need to add the linting and testing commands to `package.json`:
```json
"scripts": {
"start": "nodemon app.js",
"test": "standard",
"lint": "standard && YOUR_TEST_CMD_HERE" ,
"fix": "standard --fix && YOUR_AUTO_LINT_CMD_HERE"
},
```
|
1.0
|
Set up EJS linting - **Todo:** Set up an EJS linter. This is an easy way to automatically style all our html/EJS code and provide more meaningful error messages for EJS (the ejs module is designed to be minimal so they don't serve detailed errors without a linter).
Here is the module: https://www.npmjs.com/package/ejs-lint
You'll then need to add the linting and testing commands to `package.json`:
```json
"scripts": {
"start": "nodemon app.js",
"test": "standard",
"lint": "standard && YOUR_TEST_CMD_HERE" ,
"fix": "standard --fix && YOUR_AUTO_LINT_CMD_HERE"
},
```
|
priority
|
set up ejs linting todo set up an ejs linter this is an easy way to automatically style all our html ejs code and provide more meaningful error messages for ejs the ejs module is designed to be minimal so they don t serve detailed errors without a linter here is the module you ll then need to add the linting and testing commands to package json json scripts start nodemon app js test standard lint standard your test cmd here fix standard fix your auto lint cmd here
| 1
|
564,275
| 16,722,372,859
|
IssuesEvent
|
2021-06-10 08:52:50
|
GluuFederation/community-edition-setup
|
https://api.github.com/repos/GluuFederation/community-edition-setup
|
closed
|
Import o_metric.ldif after o=backend is created
|
High Priority bug exclude from release notes
|
The template specify ou=statistic,o=metric but there’s no parent o=metric entry there.
The import sequence is what matters here
**o_metric.ldif** (contains o=metric) -> **base.ldif**
Reported by @iromli
|
1.0
|
Import o_metric.ldif after o=backend is created - The template specify ou=statistic,o=metric but there’s no parent o=metric entry there.
The import sequence is what matters here
**o_metric.ldif** (contains o=metric) -> **base.ldif**
Reported by @iromli
|
priority
|
import o metric ldif after o backend is created the template specify ou statistic o metric but there’s no parent o metric entry there the import sequence is what matters here o metric ldif contains o metric base ldif reported by iromli
| 1
|
83,359
| 3,634,000,962
|
IssuesEvent
|
2016-02-11 16:29:23
|
ngageoint/hootenanny
|
https://api.github.com/repos/ngageoint/hootenanny
|
closed
|
TDS translations incorrectly assign the "divider" tag for divided roads.
|
Category: Core Priority: High Status: In Progress Type: Bug
|
Hoot uses the "divider=yes" tag to decide if a way needs to be split into two separate ways prior to conflation.
The SEP attribute in MGCP and TDS is used to indicate that a road is divided. With MGCP, a single piece of geometry can be tagged with SEP = True to indicate that it is a divided road.
In TDS, the extraction spec states that
'''
The opposing travelled ways of a divided highway shall be extracted as separate Road features. Horizontal separation is what classifies a Road as a divided highway. There is no minimum separation criterion for a divided highway. Each of the Road delineations shall be attributed as 'Divided' is 'True' (SEP = 1001) [i.e., "Divided"]. If a divided highway also shows a vertical barrier, each of the road delineations shall be attributed as 'Median Present' is 'True' (MES = 1001) [i.e., "With Median"].
'''
For TDS, the SEP attribute should not be translated to the "divider" tag as this will cause Hoot to split an already separated road.
|
1.0
|
TDS translations incorrectly assign the "divider" tag for divided roads. - Hoot uses the "divider=yes" tag to decide if a way needs to be split into two separate ways prior to conflation.
The SEP attribute in MGCP and TDS is used to indicate that a road is divided. With MGCP, a single piece of geometry can be tagged with SEP = True to indicate that it is a divided road.
In TDS, the extraction spec states that
'''
The opposing travelled ways of a divided highway shall be extracted as separate Road features. Horizontal separation is what classifies a Road as a divided highway. There is no minimum separation criterion for a divided highway. Each of the Road delineations shall be attributed as 'Divided' is 'True' (SEP = 1001) [i.e., "Divided"]. If a divided highway also shows a vertical barrier, each of the road delineations shall be attributed as 'Median Present' is 'True' (MES = 1001) [i.e., "With Median"].
'''
For TDS, the SEP attribute should not be translated to the "divider" tag as this will cause Hoot to split an already separated road.
|
priority
|
tds translations incorrectly assign the divider tag for divided roads hoot uses the divider yes tag to decide if a way needs to be split into two separate ways prior to conflation the sep attribute in mgcp and tds is used to indicate that a road is divided with mgcp a single piece of geometry can be tagged with sep true to indicate that it is a divided road in tds the extraction spec states that the opposing travelled ways of a divided highway shall be extracted as separate road features horizontal separation is what classifies a road as a divided highway there is no minimum separation criterion for a divided highway each of the road delineations shall be attributed as divided is true sep if a divided highway also shows a vertical barrier each of the road delineations shall be attributed as median present is true mes for tds the sep attribute should not be translated to the divider tag as this will cause hoot to split an already separated road
| 1
|
195,242
| 6,905,757,759
|
IssuesEvent
|
2017-11-27 08:44:01
|
sevin7676/SurvivalArenaTracking
|
https://api.github.com/repos/sevin7676/SurvivalArenaTracking
|
opened
|
Tasks are often far too difficult to complete
|
Priority#1: High Status#3: In Progress Type#3: Bug
|
Tasks are often far too difficult to complete or require an unreasonable amount of time for the reward provided. (reported by many league 5 players)
Reported 10/16/2017
Dev response: (multiple responses - said they are looking into it)
|
1.0
|
Tasks are often far too difficult to complete - Tasks are often far too difficult to complete or require an unreasonable amount of time for the reward provided. (reported by many league 5 players)
Reported 10/16/2017
Dev response: (multiple responses - said they are looking into it)
|
priority
|
tasks are often far too difficult to complete tasks are often far too difficult to complete or require an unreasonable amount of time for the reward provided reported by many league players reported dev response multiple responses said they are looking into it
| 1
|
268,766
| 8,411,452,218
|
IssuesEvent
|
2018-10-12 13:56:40
|
CS2113-AY1819S1-W12-2/main
|
https://api.github.com/repos/CS2113-AY1819S1-W12-2/main
|
opened
|
As a user I can download the notes of a specific module in one command
|
priority.high type.enhancement
|
This is to avoid downloading unnecessary notes.
|
1.0
|
As a user I can download the notes of a specific module in one command - This is to avoid downloading unnecessary notes.
|
priority
|
as a user i can download the notes of a specific module in one command this is to avoid downloading unnecessary notes
| 1
|
315,696
| 9,631,087,979
|
IssuesEvent
|
2019-05-15 13:32:48
|
code4romania/monitorizare-vot-android
|
https://api.github.com/repos/code4romania/monitorizare-vot-android
|
closed
|
Update questions data model - add form id
|
android bug enhancement good first issue help wanted high priority
|
Questions should store form id & question id.
Swagger for API can be found [here](https://mv-mobile-prod.azurewebsites.net/swagger/ui/index.html).
Best practices
- codebase language should be English
- we :two_hearts: unit tests
|
1.0
|
Update questions data model - add form id - Questions should store form id & question id.
Swagger for API can be found [here](https://mv-mobile-prod.azurewebsites.net/swagger/ui/index.html).
Best practices
- codebase language should be English
- we :two_hearts: unit tests
|
priority
|
update questions data model add form id questions should store form id question id swagger for api can be found best practices codebase language should be english we two hearts unit tests
| 1
|
719,564
| 24,764,109,189
|
IssuesEvent
|
2022-10-22 09:27:20
|
devvsakib/hacktoberfest-react-project
|
https://api.github.com/repos/devvsakib/hacktoberfest-react-project
|
closed
|
[BRANCH - "convert-react"] DELETE PROJECTS FOLDER
|
enhancement hacktoberfest hacktoberfest-accepted [priority: high]
|
**Navigate to folder**
Check 2nd Branch, name- convert-react
Clone this, navigate to main-website folder,this is our main website we are converting into react website,
## Feature?
No
## Bug?
No
Task:
Delete PROJECTS folder. we will add projects later, though we have "main" branch with "PROJECTS" folder so we don't need to have it in our "convert-react" branch. Please be aware, and make pr to convert-react branch
You can join our discord server - let's build together [Join](https://discord.gg/xwRbDwnbMr)
|
1.0
|
[BRANCH - "convert-react"] DELETE PROJECTS FOLDER - **Navigate to folder**
Check 2nd Branch, name- convert-react
Clone this, navigate to main-website folder,this is our main website we are converting into react website,
## Feature?
No
## Bug?
No
Task:
Delete PROJECTS folder. we will add projects later, though we have "main" branch with "PROJECTS" folder so we don't need to have it in our "convert-react" branch. Please be aware, and make pr to convert-react branch
You can join our discord server - let's build together [Join](https://discord.gg/xwRbDwnbMr)
|
priority
|
delete projects folder navigate to folder check branch name convert react clone this navigate to main website folder this is our main website we are converting into react website feature no bug no task delete projects folder we will add projects later though we have main branch with projects folder so we don t need to have it in our convert react branch please be aware and make pr to convert react branch you can join our discord server let s build together
| 1
|
63,369
| 3,194,713,668
|
IssuesEvent
|
2015-09-30 13:38:04
|
fusioninventory/fusioninventory-for-glpi
|
https://api.github.com/repos/fusioninventory/fusioninventory-for-glpi
|
closed
|
Missing update process to split Tasks with multiple TaskJobs
|
Category: Tasks Component: For junior contributor Component: Found in version Priority: High Status: Closed Tracker: Bug
|
---
Author Name: **Kevin Roy** (@kiniou)
Original Redmine Issue: 2018, http://forge.fusioninventory.org/issues/2018
Original Date: 2013-03-20
Original Assignee: David Durieux
---
Since multijobs task are abandonned, taskjobs should be splitted into one task per job.
|
1.0
|
Missing update process to split Tasks with multiple TaskJobs - ---
Author Name: **Kevin Roy** (@kiniou)
Original Redmine Issue: 2018, http://forge.fusioninventory.org/issues/2018
Original Date: 2013-03-20
Original Assignee: David Durieux
---
Since multijobs task are abandonned, taskjobs should be splitted into one task per job.
|
priority
|
missing update process to split tasks with multiple taskjobs author name kevin roy kiniou original redmine issue original date original assignee david durieux since multijobs task are abandonned taskjobs should be splitted into one task per job
| 1
|
354,725
| 10,571,432,535
|
IssuesEvent
|
2019-10-07 07:06:43
|
AY1920S1-CS2103-T11-2/main
|
https://api.github.com/repos/AY1920S1-CS2103-T11-2/main
|
closed
|
Initial code refactor
|
priority.High status.Ongoing type.Task
|
Refactor code to prepare for MAMS implementation, Rename all relevant classes.
|
1.0
|
Initial code refactor - Refactor code to prepare for MAMS implementation, Rename all relevant classes.
|
priority
|
initial code refactor refactor code to prepare for mams implementation rename all relevant classes
| 1
|
116,482
| 4,702,775,375
|
IssuesEvent
|
2016-10-13 04:17:45
|
CS2103AUG2016-T14-C4/main
|
https://api.github.com/repos/CS2103AUG2016-T14-C4/main
|
opened
|
As a user, I can type commands without a certain order in arguments
|
priority.high type.task
|
so that I only need to remember the flags for each command and not the order
|
1.0
|
As a user, I can type commands without a certain order in arguments - so that I only need to remember the flags for each command and not the order
|
priority
|
as a user i can type commands without a certain order in arguments so that i only need to remember the flags for each command and not the order
| 1
|
506,566
| 14,667,971,182
|
IssuesEvent
|
2020-12-29 20:00:23
|
mike-eason/solaris
|
https://api.github.com/repos/mike-eason/solaris
|
closed
|
As Solaris I would like to upgrade the server infrastructure
|
enhancement estimate: 13 priority: high status: done
|
- Upgrade Ubuntu
- Upgrade MongoDB
Might as well spin up a new server afresh.
This may resolve the issue with the messages not being marked as read.
|
1.0
|
As Solaris I would like to upgrade the server infrastructure - - Upgrade Ubuntu
- Upgrade MongoDB
Might as well spin up a new server afresh.
This may resolve the issue with the messages not being marked as read.
|
priority
|
as solaris i would like to upgrade the server infrastructure upgrade ubuntu upgrade mongodb might as well spin up a new server afresh this may resolve the issue with the messages not being marked as read
| 1
|
115,286
| 4,662,351,494
|
IssuesEvent
|
2016-10-05 03:14:00
|
communityshare/communityshare
|
https://api.github.com/repos/communityshare/communityshare
|
closed
|
Alert text shown briefly on page load
|
[Effort] Good first issues [Priority] High [Skills] Angular [Type] Bug [Type] UX
|
When loading the app home page, there is an alert that briefly shows before the rest of the HTML is rendered. It's empty and looks like a bug when viewing the page. This alert should not appear.
Likely resolved inside of `static/index.html`

This may have been introduced in in [this commit](https://github.com/communityshare/communityshare/commit/d085fb091cdfeaeffd682c73906f9bc9f8e4a68e).
|
1.0
|
Alert text shown briefly on page load - When loading the app home page, there is an alert that briefly shows before the rest of the HTML is rendered. It's empty and looks like a bug when viewing the page. This alert should not appear.
Likely resolved inside of `static/index.html`

This may have been introduced in in [this commit](https://github.com/communityshare/communityshare/commit/d085fb091cdfeaeffd682c73906f9bc9f8e4a68e).
|
priority
|
alert text shown briefly on page load when loading the app home page there is an alert that briefly shows before the rest of the html is rendered it s empty and looks like a bug when viewing the page this alert should not appear likely resolved inside of static index html this may have been introduced in in
| 1
|
98,845
| 4,031,973,513
|
IssuesEvent
|
2016-05-18 18:59:16
|
neviko/Diff-sign-project
|
https://api.github.com/repos/neviko/Diff-sign-project
|
closed
|
style and design of the learning mode
|
4 - Done point: 5 priority: high
|
design an construct the style of all categories of learning mode
learning house
learning playground
learning general
learning letters
|
1.0
|
style and design of the learning mode - design an construct the style of all categories of learning mode
learning house
learning playground
learning general
learning letters
|
priority
|
style and design of the learning mode design an construct the style of all categories of learning mode learning house learning playground learning general learning letters
| 1
|
514,477
| 14,939,910,888
|
IssuesEvent
|
2021-01-25 17:33:08
|
mlee97/SOEN-390-Team5
|
https://api.github.com/repos/mlee97/SOEN-390-Team5
|
opened
|
[USER STORY]: As an ERP user, I want to login, so that I can access the appropriate interface.
|
priority: high risk: high story point: 5 user story
|
# State the Related Minimal Requirement:
The system should allow only registered users to access the information, possible with different accesses for different roles.
|
1.0
|
[USER STORY]: As an ERP user, I want to login, so that I can access the appropriate interface. - # State the Related Minimal Requirement:
The system should allow only registered users to access the information, possible with different accesses for different roles.
|
priority
|
as an erp user i want to login so that i can access the appropriate interface state the related minimal requirement the system should allow only registered users to access the information possible with different accesses for different roles
| 1
|
312,012
| 9,541,888,911
|
IssuesEvent
|
2019-05-01 00:21:11
|
certbot/certbot
|
https://api.github.com/repos/certbot/certbot
|
closed
|
Fedora apachectl 2.4.39 no longer supports passing arguments to httpd
|
area: apache current sprint priority: high
|
## My operating system is (include version):
Fedora 29
## I installed Certbot with (certbot-auto, OS package manager, pip, etc):
OS package manager
Additionally: httpd-2.4.39
## I ran this command and it produced this output:
certbot renew --apache
apachectl: The "-v" option is not supported.
## Certbot's behavior differed from what I expected because:
certbot gets invalid output from apachectl
## Here is a Certbot log showing the issue (if available):
/usr/sbin/apachectl doesn't support -v, nor -t -D DUMP_RUN_CFG, DUMP_INCLUDES, DUMP_MODULES
Hacking apachectl (or a different one in the path) to specifically allow these options makes it work.
Running httpd -v or -t -D DUMP_RUN_CFG does work, apachectl doesn't.
## Here is the relevant nginx server block or Apache virtualhost for the domain I am configuring:
not relevant, change to /usr/sbin/apachectl, probably in 2.4.39. httpd (and config) is fine, apachectl is not.
not sure if I should file a bug here, or with Fedora
Issue #5439 seems very similar, but appears to be for restart of apache only. This issue occurs before the certificate -- certbot is still verifying if apache is installed.
|
1.0
|
Fedora apachectl 2.4.39 no longer supports passing arguments to httpd - ## My operating system is (include version):
Fedora 29
## I installed Certbot with (certbot-auto, OS package manager, pip, etc):
OS package manager
Additionally: httpd-2.4.39
## I ran this command and it produced this output:
certbot renew --apache
apachectl: The "-v" option is not supported.
## Certbot's behavior differed from what I expected because:
certbot gets invalid output from apachectl
## Here is a Certbot log showing the issue (if available):
/usr/sbin/apachectl doesn't support -v, nor -t -D DUMP_RUN_CFG, DUMP_INCLUDES, DUMP_MODULES
Hacking apachectl (or a different one in the path) to specifically allow these options makes it work.
Running httpd -v or -t -D DUMP_RUN_CFG does work, apachectl doesn't.
## Here is the relevant nginx server block or Apache virtualhost for the domain I am configuring:
not relevant, change to /usr/sbin/apachectl, probably in 2.4.39. httpd (and config) is fine, apachectl is not.
not sure if I should file a bug here, or with Fedora
Issue #5439 seems very similar, but appears to be for restart of apache only. This issue occurs before the certificate -- certbot is still verifying if apache is installed.
|
priority
|
fedora apachectl no longer supports passing arguments to httpd my operating system is include version fedora i installed certbot with certbot auto os package manager pip etc os package manager additionally httpd i ran this command and it produced this output certbot renew apache apachectl the v option is not supported certbot s behavior differed from what i expected because certbot gets invalid output from apachectl here is a certbot log showing the issue if available usr sbin apachectl doesn t support v nor t d dump run cfg dump includes dump modules hacking apachectl or a different one in the path to specifically allow these options makes it work running httpd v or t d dump run cfg does work apachectl doesn t here is the relevant nginx server block or apache virtualhost for the domain i am configuring not relevant change to usr sbin apachectl probably in httpd and config is fine apachectl is not not sure if i should file a bug here or with fedora issue seems very similar but appears to be for restart of apache only this issue occurs before the certificate certbot is still verifying if apache is installed
| 1
|
561,813
| 16,624,605,863
|
IssuesEvent
|
2021-06-03 07:59:46
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
[LS] No documentation provided on hover for Record Fields
|
Area/Hover Priority/High SwanLakeDump Team/LanguageServer Type/Bug
|
**Description:**
$subject
**Steps to reproduce:**
```ballerina
public type NodeCredential record {|
# IP Address
string ip;
int port;
string username;
constants:NodeType nodetype;
# Optfield
string optField ?;
|};
```
|
1.0
|
[LS] No documentation provided on hover for Record Fields - **Description:**
$subject
**Steps to reproduce:**
```ballerina
public type NodeCredential record {|
# IP Address
string ip;
int port;
string username;
constants:NodeType nodetype;
# Optfield
string optField ?;
|};
```
|
priority
|
no documentation provided on hover for record fields description subject steps to reproduce ballerina public type nodecredential record ip address string ip int port string username constants nodetype nodetype optfield string optfield
| 1
|
76,754
| 3,491,912,585
|
IssuesEvent
|
2016-01-04 17:48:18
|
ceylon/ceylon
|
https://api.github.com/repos/ceylon/ceylon
|
closed
|
Broken if-then-else expression in JavaScript compilation
|
bug c-compiler-js high priority
|
The following code, which perfectly works when compiling to JVM breaks in the JavaScript version of the program:
```
"Run the module `ifelse`."
shared void run() {
Integer x = 1;
Integer y = 2;
String text = (if (x > 0) then x.string + "a " else "") + y.string + "b";
assert (text == "1a 2b");
}
```
When extracting the expression `(if (x > 0) then x.string + "a " else "")` to a new local variable, everything is fine in JavaScript, too.
|
1.0
|
Broken if-then-else expression in JavaScript compilation - The following code, which perfectly works when compiling to JVM breaks in the JavaScript version of the program:
```
"Run the module `ifelse`."
shared void run() {
Integer x = 1;
Integer y = 2;
String text = (if (x > 0) then x.string + "a " else "") + y.string + "b";
assert (text == "1a 2b");
}
```
When extracting the expression `(if (x > 0) then x.string + "a " else "")` to a new local variable, everything is fine in JavaScript, too.
|
priority
|
broken if then else expression in javascript compilation the following code which perfectly works when compiling to jvm breaks in the javascript version of the program run the module ifelse shared void run integer x integer y string text if x then x string a else y string b assert text when extracting the expression if x then x string a else to a new local variable everything is fine in javascript too
| 1
|
752,981
| 26,335,998,190
|
IssuesEvent
|
2023-01-10 14:23:33
|
vaticle/typedb
|
https://api.github.com/repos/vaticle/typedb
|
closed
|
Improve error message for MacOS M1 chip users
|
type: bug priority: high
|
## Problem to Solve
MacBook M1 users get the following unhelpful error message on server startup:
```
================================================================
________ __ __ _____ _______ _____ _____
|__ __|\ \ / /| _ \ | _ || _ \ | _ \
| | \ \/ / | | | || | |__|| | | || | | |
| | \ / | |/ / | |___ | | | || |/ /
| | | | | __/ | ___| | | | || _ \
| | | | | | | | __ | | | || | | |
| | | | | | | |_| || |/ / | |/ /
|__| |__| |__| |_______||_____/ |_____/
A STRONGLY-TYPED DATABASE
================================================================
Exception in thread "main" java.lang.UnsatisfiedLinkError: /private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: dlopen(/private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib, 1): no suitable image found. Did find:
/private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: mach-o, but wrong architecture
/private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: mach-o, but wrong architecture
at java.base/jdk.internal.loader.NativeLibraries.load(Native Method)
at java.base/jdk.internal.loader.NativeLibraries$NativeLibraryImpl.open(NativeLibraries.java:384)
at java.base/jdk.internal.loader.NativeLibraries.loadLibrary(NativeLibraries.java:228)
at java.base/jdk.internal.loader.NativeLibraries.loadLibrary(NativeLibraries.java:170)
at java.base/java.lang.ClassLoader.loadLibrary(ClassLoader.java:2389)
at java.base/java.lang.Runtime.load0(Runtime.java:755)
at java.base/java.lang.System.load(System.java:1953)
at org.rocksdb.NativeLibraryLoader.loadLibraryFromJar(NativeLibraryLoader.java:79)
at org.rocksdb.NativeLibraryLoader.loadLibrary(NativeLibraryLoader.java:57)
at org.rocksdb.RocksDB.loadLibrary(RocksDB.java:69)
at org.rocksdb.RocksDB.<clinit>(RocksDB.java:38)
at com.vaticle.typedb.core.rocks.RocksTypeDB.<clinit>(RocksTypeDB.java:47)
at com.vaticle.typedb.core.rocks.RocksFactory.typedb(RocksFactory.java:34)
at com.vaticle.typedb.core.server.TypeDBServer.<init>(TypeDBServer.java:88)
at com.vaticle.typedb.core.server.TypeDBServer.<init>(TypeDBServer.java:72)
at com.vaticle.typedb.core.server.TypeDBServer.runServer(TypeDBServer.java:244)
at com.vaticle.typedb.core.server.TypeDBServer.main(TypeDBServer.java:225)
```
## Proposed Solution
We should detect MacOS M1 architecture running on a non-x64 JDK and give a better error message, something like: "The installed JDK is not compatible with TypeDB, please use a JDK build that targets x64 architecture. See https://forum.vaticle.com/t/running-typedb-on-an-m1-macbook/53 for details"
|
1.0
|
Improve error message for MacOS M1 chip users - ## Problem to Solve
MacBook M1 users get the following unhelpful error message on server startup:
```
================================================================
________ __ __ _____ _______ _____ _____
|__ __|\ \ / /| _ \ | _ || _ \ | _ \
| | \ \/ / | | | || | |__|| | | || | | |
| | \ / | |/ / | |___ | | | || |/ /
| | | | | __/ | ___| | | | || _ \
| | | | | | | | __ | | | || | | |
| | | | | | | |_| || |/ / | |/ /
|__| |__| |__| |_______||_____/ |_____/
A STRONGLY-TYPED DATABASE
================================================================
Exception in thread "main" java.lang.UnsatisfiedLinkError: /private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: dlopen(/private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib, 1): no suitable image found. Did find:
/private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: mach-o, but wrong architecture
/private/var/folders/ry/bnfxqrqd1bd4mxtvgvnsr9j80000gn/T/librocksdbjni18708461303295440.jnilib: mach-o, but wrong architecture
at java.base/jdk.internal.loader.NativeLibraries.load(Native Method)
at java.base/jdk.internal.loader.NativeLibraries$NativeLibraryImpl.open(NativeLibraries.java:384)
at java.base/jdk.internal.loader.NativeLibraries.loadLibrary(NativeLibraries.java:228)
at java.base/jdk.internal.loader.NativeLibraries.loadLibrary(NativeLibraries.java:170)
at java.base/java.lang.ClassLoader.loadLibrary(ClassLoader.java:2389)
at java.base/java.lang.Runtime.load0(Runtime.java:755)
at java.base/java.lang.System.load(System.java:1953)
at org.rocksdb.NativeLibraryLoader.loadLibraryFromJar(NativeLibraryLoader.java:79)
at org.rocksdb.NativeLibraryLoader.loadLibrary(NativeLibraryLoader.java:57)
at org.rocksdb.RocksDB.loadLibrary(RocksDB.java:69)
at org.rocksdb.RocksDB.<clinit>(RocksDB.java:38)
at com.vaticle.typedb.core.rocks.RocksTypeDB.<clinit>(RocksTypeDB.java:47)
at com.vaticle.typedb.core.rocks.RocksFactory.typedb(RocksFactory.java:34)
at com.vaticle.typedb.core.server.TypeDBServer.<init>(TypeDBServer.java:88)
at com.vaticle.typedb.core.server.TypeDBServer.<init>(TypeDBServer.java:72)
at com.vaticle.typedb.core.server.TypeDBServer.runServer(TypeDBServer.java:244)
at com.vaticle.typedb.core.server.TypeDBServer.main(TypeDBServer.java:225)
```
## Proposed Solution
We should detect MacOS M1 architecture running on a non-x64 JDK and give a better error message, something like: "The installed JDK is not compatible with TypeDB, please use a JDK build that targets x64 architecture. See https://forum.vaticle.com/t/running-typedb-on-an-m1-macbook/53 for details"
|
priority
|
improve error message for macos chip users problem to solve macbook users get the following unhelpful error message on server startup a strongly typed database exception in thread main java lang unsatisfiedlinkerror private var folders ry t jnilib dlopen private var folders ry t jnilib no suitable image found did find private var folders ry t jnilib mach o but wrong architecture private var folders ry t jnilib mach o but wrong architecture at java base jdk internal loader nativelibraries load native method at java base jdk internal loader nativelibraries nativelibraryimpl open nativelibraries java at java base jdk internal loader nativelibraries loadlibrary nativelibraries java at java base jdk internal loader nativelibraries loadlibrary nativelibraries java at java base java lang classloader loadlibrary classloader java at java base java lang runtime runtime java at java base java lang system load system java at org rocksdb nativelibraryloader loadlibraryfromjar nativelibraryloader java at org rocksdb nativelibraryloader loadlibrary nativelibraryloader java at org rocksdb rocksdb loadlibrary rocksdb java at org rocksdb rocksdb rocksdb java at com vaticle typedb core rocks rockstypedb rockstypedb java at com vaticle typedb core rocks rocksfactory typedb rocksfactory java at com vaticle typedb core server typedbserver typedbserver java at com vaticle typedb core server typedbserver typedbserver java at com vaticle typedb core server typedbserver runserver typedbserver java at com vaticle typedb core server typedbserver main typedbserver java proposed solution we should detect macos architecture running on a non jdk and give a better error message something like the installed jdk is not compatible with typedb please use a jdk build that targets architecture see for details
| 1
|
62,794
| 3,193,234,367
|
IssuesEvent
|
2015-09-30 03:02:18
|
DarkstarProject/darkstar
|
https://api.github.com/repos/DarkstarProject/darkstar
|
closed
|
Maat's Cap SMN fight
|
High Priority
|
If the pet keeps auto attacking Maat and kills him while the ending CS plays (cant move) it does not count as a win for the Maat's cap quest.
|
1.0
|
Maat's Cap SMN fight - If the pet keeps auto attacking Maat and kills him while the ending CS plays (cant move) it does not count as a win for the Maat's cap quest.
|
priority
|
maat s cap smn fight if the pet keeps auto attacking maat and kills him while the ending cs plays cant move it does not count as a win for the maat s cap quest
| 1
|
305,012
| 9,358,442,940
|
IssuesEvent
|
2019-04-02 02:26:41
|
CS2103-AY1819S2-W17-4/main
|
https://api.github.com/repos/CS2103-AY1819S2-W17-4/main
|
closed
|
Find Expense
|
priority.High status.Ongoing type.Epic type.Story
|
To complete Task
- [x] to search for an expense using
- [x] name
- [x] price
- [x] tags
- [x] date
|
1.0
|
Find Expense - To complete Task
- [x] to search for an expense using
- [x] name
- [x] price
- [x] tags
- [x] date
|
priority
|
find expense to complete task to search for an expense using name price tags date
| 1
|
234,318
| 7,719,716,886
|
IssuesEvent
|
2018-05-23 20:21:49
|
aseprite/aseprite
|
https://api.github.com/repos/aseprite/aseprite
|
closed
|
Add clipboard support for Linux
|
high priority linux
|
We should be able to copy & paste text and images between Aseprite and other programs like GIMP.
|
1.0
|
Add clipboard support for Linux - We should be able to copy & paste text and images between Aseprite and other programs like GIMP.
|
priority
|
add clipboard support for linux we should be able to copy paste text and images between aseprite and other programs like gimp
| 1
|
280,465
| 8,682,193,521
|
IssuesEvent
|
2018-12-02 05:01:22
|
MonoOni/binarydist
|
https://api.github.com/repos/MonoOni/binarydist
|
opened
|
Mono profiler is broken
|
area: JIT os: aix os: i priority: high type: bug
|
Was low priority until `--trace` started to require it to work. For now, you can revert mono/mono@2b029b3df4a1816880e3f02bacd35cf9a0835e4e, but this won't be sustainable. Need to figure out how to get this working, otherwise debugging will be a LOT harder.
|
1.0
|
Mono profiler is broken - Was low priority until `--trace` started to require it to work. For now, you can revert mono/mono@2b029b3df4a1816880e3f02bacd35cf9a0835e4e, but this won't be sustainable. Need to figure out how to get this working, otherwise debugging will be a LOT harder.
|
priority
|
mono profiler is broken was low priority until trace started to require it to work for now you can revert mono mono but this won t be sustainable need to figure out how to get this working otherwise debugging will be a lot harder
| 1
|
827,441
| 31,772,912,481
|
IssuesEvent
|
2023-09-12 12:53:32
|
Budibase/budibase
|
https://api.github.com/repos/Budibase/budibase
|
closed
|
[BUDI-7468] Save Row action will attempt to save the image data of the S3 Upload component
|
bug forms s3 env - production High priority bb-components
|
**Checklist**
* [X] I have searched budibase discussions and github issues to check if my issue already exists
**Hosting**
* Cloud
* Tenant ID: bb.budibase.app
**Describe the bug**
**Save Row** action will attempt to save the image base64 data contents of the **S3 Upload** component.
This can cause a `Request Entity Too Large` if a large amount of image data is unintentionally passed to the `/rows` endpoint.
**To Reproduce**
Steps to reproduce the behavior:
1. Add a form with a **custom** schema
2. Add an S3 File Upload component
3. Add a button with a **Save Row** action
4. Preview the app and upload an image
5. Look in the network tab and see the raw image data being saved against the `/rows` endpoint
**Expected behavior**
The S3 Upload component should be ignored from the form schema to prevent it being saved against the `/rows` endpoint and causing unintended errors such as `Request Entity Too Large`
**Screenshots**



**Additional context**
Only seems to be an issue for images file types. Other object types seem to only provide a small amount of metadata json.
<sub>From [SyncLinear.com](https://synclinear.com) | [BUDI-7468](https://linear.app/budibase/issue/BUDI-7468/save-row-action-will-attempt-to-save-the-image-data-of-the-s3-upload)</sub>
|
1.0
|
[BUDI-7468] Save Row action will attempt to save the image data of the S3 Upload component - **Checklist**
* [X] I have searched budibase discussions and github issues to check if my issue already exists
**Hosting**
* Cloud
* Tenant ID: bb.budibase.app
**Describe the bug**
**Save Row** action will attempt to save the image base64 data contents of the **S3 Upload** component.
This can cause a `Request Entity Too Large` if a large amount of image data is unintentionally passed to the `/rows` endpoint.
**To Reproduce**
Steps to reproduce the behavior:
1. Add a form with a **custom** schema
2. Add an S3 File Upload component
3. Add a button with a **Save Row** action
4. Preview the app and upload an image
5. Look in the network tab and see the raw image data being saved against the `/rows` endpoint
**Expected behavior**
The S3 Upload component should be ignored from the form schema to prevent it being saved against the `/rows` endpoint and causing unintended errors such as `Request Entity Too Large`
**Screenshots**



**Additional context**
Only seems to be an issue for images file types. Other object types seem to only provide a small amount of metadata json.
<sub>From [SyncLinear.com](https://synclinear.com) | [BUDI-7468](https://linear.app/budibase/issue/BUDI-7468/save-row-action-will-attempt-to-save-the-image-data-of-the-s3-upload)</sub>
|
priority
|
save row action will attempt to save the image data of the upload component checklist i have searched budibase discussions and github issues to check if my issue already exists hosting cloud tenant id bb budibase app describe the bug save row action will attempt to save the image data contents of the upload component this can cause a request entity too large if a large amount of image data is unintentionally passed to the rows endpoint to reproduce steps to reproduce the behavior add a form with a custom schema add an file upload component add a button with a save row action preview the app and upload an image look in the network tab and see the raw image data being saved against the rows endpoint expected behavior the upload component should be ignored from the form schema to prevent it being saved against the rows endpoint and causing unintended errors such as request entity too large screenshots additional context only seems to be an issue for images file types other object types seem to only provide a small amount of metadata json from
| 1
|
669,219
| 22,616,722,224
|
IssuesEvent
|
2022-06-29 23:17:11
|
paleobot/pbot-dev
|
https://api.github.com/repos/paleobot/pbot-dev
|
opened
|
Administrators Group
|
enhancement high priority
|
We need there to be an Administrators Group that allows system administrators to see private data.
Among other things, this allow PBot ExCom to see what data have been in the private sphere for a long time and needs to go public.
|
1.0
|
Administrators Group - We need there to be an Administrators Group that allows system administrators to see private data.
Among other things, this allow PBot ExCom to see what data have been in the private sphere for a long time and needs to go public.
|
priority
|
administrators group we need there to be an administrators group that allows system administrators to see private data among other things this allow pbot excom to see what data have been in the private sphere for a long time and needs to go public
| 1
|
96,514
| 3,969,098,400
|
IssuesEvent
|
2016-05-03 22:02:24
|
KrzysztofSendor/python-chess
|
https://api.github.com/repos/KrzysztofSendor/python-chess
|
closed
|
Coordinates translation
|
difficulty: easy priority: high
|
Create function in board module for translating notation (eg. a2, c6) to Board array coordinates (eg. Board[0][6], Board[2][2]). Mind that array is counted from upper left with whites in bottom and in standard notation board is counted from bottom left with whites in bottom.
Practically the whole code is already written in movement handling function. Just copy it.
**Should return a list (an array) of integers.** `input = "a2" output = [0,6]`
Create second function that does the same thing but the other way around.
**Should return a list from a string given.**
**All further created functions should use it.**
|
1.0
|
Coordinates translation - Create function in board module for translating notation (eg. a2, c6) to Board array coordinates (eg. Board[0][6], Board[2][2]). Mind that array is counted from upper left with whites in bottom and in standard notation board is counted from bottom left with whites in bottom.
Practically the whole code is already written in movement handling function. Just copy it.
**Should return a list (an array) of integers.** `input = "a2" output = [0,6]`
Create second function that does the same thing but the other way around.
**Should return a list from a string given.**
**All further created functions should use it.**
|
priority
|
coordinates translation create function in board module for translating notation eg to board array coordinates eg board board mind that array is counted from upper left with whites in bottom and in standard notation board is counted from bottom left with whites in bottom practically the whole code is already written in movement handling function just copy it should return a list an array of integers input output create second function that does the same thing but the other way around should return a list from a string given all further created functions should use it
| 1
|
764,299
| 26,793,891,756
|
IssuesEvent
|
2023-02-01 10:27:02
|
wso2/carbon-analytics
|
https://api.github.com/repos/wso2/carbon-analytics
|
closed
|
[Business Rules] Error shown while deploying the business rule
|
Type/Bug Severity/Major Priority/High Component/BusinessRules
|
**Description:**
When creating a new business rule from a template, and press save & deploy an error 'Failed to deploy' is shown in the UI and dashboard terminal while deploying the app.
`Failed to deploy business rule sweetProd org.wso2.carbon.business.rules.core.exceptions.SiddhiAppsApiHelperException: Unexpected error occurred during saving the siddhi app '@App:name('sweetprod_0')`
But the business rule get successfully deployed.

**Suggested Labels:**
Bug, Business Rules
**Suggested Assignees:**
@minudika
**Affected Product Version:**
4.2.0
**OS, DB, other environment details and versions:**
Windows10
|
1.0
|
[Business Rules] Error shown while deploying the business rule - **Description:**
When creating a new business rule from a template, and press save & deploy an error 'Failed to deploy' is shown in the UI and dashboard terminal while deploying the app.
`Failed to deploy business rule sweetProd org.wso2.carbon.business.rules.core.exceptions.SiddhiAppsApiHelperException: Unexpected error occurred during saving the siddhi app '@App:name('sweetprod_0')`
But the business rule get successfully deployed.

**Suggested Labels:**
Bug, Business Rules
**Suggested Assignees:**
@minudika
**Affected Product Version:**
4.2.0
**OS, DB, other environment details and versions:**
Windows10
|
priority
|
error shown while deploying the business rule description when creating a new business rule from a template and press save deploy an error failed to deploy is shown in the ui and dashboard terminal while deploying the app failed to deploy business rule sweetprod org carbon business rules core exceptions siddhiappsapihelperexception unexpected error occurred during saving the siddhi app app name sweetprod but the business rule get successfully deployed suggested labels bug business rules suggested assignees minudika affected product version os db other environment details and versions
| 1
|
183,212
| 6,678,391,666
|
IssuesEvent
|
2017-10-05 14:06:36
|
CCAFS/MARLO
|
https://api.github.com/repos/CCAFS/MARLO
|
closed
|
Harvesting metadata from CIMMYT repositories
|
Priority - High Type -Task
|
Branch: dev-rest-metadata
- [ ] Create connection class to CIMMYT
- [ ] Map XML data to metadata Model
- [ ] Add CIMMYT as example in deliverables section
|
1.0
|
Harvesting metadata from CIMMYT repositories - Branch: dev-rest-metadata
- [ ] Create connection class to CIMMYT
- [ ] Map XML data to metadata Model
- [ ] Add CIMMYT as example in deliverables section
|
priority
|
harvesting metadata from cimmyt repositories branch dev rest metadata create connection class to cimmyt map xml data to metadata model add cimmyt as example in deliverables section
| 1
|
382,154
| 11,301,565,434
|
IssuesEvent
|
2020-01-17 15:50:35
|
mono/monodevelop
|
https://api.github.com/repos/mono/monodevelop
|
closed
|
[Diagnostics] Implement RuleSet file support
|
Area: C# Binding high-priority not-regression vs-sync
|
VS normally looks for ruleset files included via MSBuild:
https://github.com/dotnet/roslyn/blob/0c62bae6aade582227232183540fd891254a76cb/src/Compilers/Core/Portable/CodeAnalysis.csproj#L14
Support for that is implemented below for VS:
http://source.roslyn.io/#Microsoft.VisualStudio.LanguageServices/Implementation/ProjectSystem/RuleSets/VisualStudioRuleSetManager.RuleSetFile.cs,106
> VS bug #577079
|
1.0
|
[Diagnostics] Implement RuleSet file support - VS normally looks for ruleset files included via MSBuild:
https://github.com/dotnet/roslyn/blob/0c62bae6aade582227232183540fd891254a76cb/src/Compilers/Core/Portable/CodeAnalysis.csproj#L14
Support for that is implemented below for VS:
http://source.roslyn.io/#Microsoft.VisualStudio.LanguageServices/Implementation/ProjectSystem/RuleSets/VisualStudioRuleSetManager.RuleSetFile.cs,106
> VS bug #577079
|
priority
|
implement ruleset file support vs normally looks for ruleset files included via msbuild support for that is implemented below for vs vs bug
| 1
|
211,729
| 7,204,040,136
|
IssuesEvent
|
2018-02-06 11:13:15
|
xmichelo/Beeftext
|
https://api.github.com/repos/xmichelo/Beeftext
|
closed
|
Error message on first launch.
|
bug high priority
|
Due to recent changes, the application now display an error message when it is first launched on PC because there is no combo list file.
|
1.0
|
Error message on first launch. - Due to recent changes, the application now display an error message when it is first launched on PC because there is no combo list file.
|
priority
|
error message on first launch due to recent changes the application now display an error message when it is first launched on pc because there is no combo list file
| 1
|
773,751
| 27,169,299,611
|
IssuesEvent
|
2023-02-17 17:52:45
|
minio/minio
|
https://api.github.com/repos/minio/minio
|
closed
|
Buckets info and cluster summary not changes
|
community priority: high
|
## NOTE
Buckets info and cluster summary not changes. Objects count and used space are the same even after few hours of big uploads. I now migrating date from one cluster to another. Before update to the current version information about object count and used space was changed every 3-5 minutes but now it shows same values for the all time after update. This values not changes in web console and also in metrics which sent to prometheus. It's interesting that only total used space metric which sending to prometheus are changes - objects count and buckets used space are the same. For now prometheus shows me 6.4 TB used and web console only 4.4. You can see it at screenshots attached.
## Expected Behavior
Information about used space and objects count should be changed according real data.
## Current Behavior
For in it shows values which was before update few days ago.
## Regression
It works fine before update in version RELEASE.2023-01-12T02-06-16Z (commit-id=7bc95c47a322971aff7d4d4c270dcf28a933e84b)
## Your Environment
* Version used (`minio --version`): RELEASE.2023-02-10T18-48-39Z (commit-id=d0f4cc89a5d0ca45e7ba00fbb7733c2fead3f44e)
* Server setup and configuration: 4 nodes with 4 disks on each
* Operating System and version (`uname -a`): Linux minio-cold-1 5.15.0-60-generic #66-Ubuntu SMP Fri Jan 20 14:29:49 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
|
1.0
|
Buckets info and cluster summary not changes - ## NOTE
Buckets info and cluster summary not changes. Objects count and used space are the same even after few hours of big uploads. I now migrating date from one cluster to another. Before update to the current version information about object count and used space was changed every 3-5 minutes but now it shows same values for the all time after update. This values not changes in web console and also in metrics which sent to prometheus. It's interesting that only total used space metric which sending to prometheus are changes - objects count and buckets used space are the same. For now prometheus shows me 6.4 TB used and web console only 4.4. You can see it at screenshots attached.
## Expected Behavior
Information about used space and objects count should be changed according real data.
## Current Behavior
For in it shows values which was before update few days ago.
## Regression
It works fine before update in version RELEASE.2023-01-12T02-06-16Z (commit-id=7bc95c47a322971aff7d4d4c270dcf28a933e84b)
## Your Environment
* Version used (`minio --version`): RELEASE.2023-02-10T18-48-39Z (commit-id=d0f4cc89a5d0ca45e7ba00fbb7733c2fead3f44e)
* Server setup and configuration: 4 nodes with 4 disks on each
* Operating System and version (`uname -a`): Linux minio-cold-1 5.15.0-60-generic #66-Ubuntu SMP Fri Jan 20 14:29:49 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
|
priority
|
buckets info and cluster summary not changes note buckets info and cluster summary not changes objects count and used space are the same even after few hours of big uploads i now migrating date from one cluster to another before update to the current version information about object count and used space was changed every minutes but now it shows same values for the all time after update this values not changes in web console and also in metrics which sent to prometheus it s interesting that only total used space metric which sending to prometheus are changes objects count and buckets used space are the same for now prometheus shows me tb used and web console only you can see it at screenshots attached expected behavior information about used space and objects count should be changed according real data current behavior for in it shows values which was before update few days ago regression it works fine before update in version release commit id your environment version used minio version release commit id server setup and configuration nodes with disks on each operating system and version uname a linux minio cold generic ubuntu smp fri jan utc gnu linux
| 1
|
695,565
| 23,864,170,046
|
IssuesEvent
|
2022-09-07 09:35:14
|
factly/dega
|
https://api.github.com/repos/factly/dega
|
closed
|
Updating kavach to the latest changes
|
priority:high studio server
|
This change includes -
- [x] moving spaces to kavach.
- [x] moving policies to kavach.
|
1.0
|
Updating kavach to the latest changes - This change includes -
- [x] moving spaces to kavach.
- [x] moving policies to kavach.
|
priority
|
updating kavach to the latest changes this change includes moving spaces to kavach moving policies to kavach
| 1
|
554,244
| 16,415,358,180
|
IssuesEvent
|
2021-05-19 05:43:46
|
django-cms/django-cms
|
https://api.github.com/repos/django-cms/django-cms
|
closed
|
Slug uniqueness not checked when moving page
|
needs expert opinion needs patch priority: high
|
## Description
When page is moved, its slug uniqueness is not checked. Thus two pages with the same slug can be present on the same level.
## Steps to reproduce
1. Create a page at the root level and give it a slug "test".
2. Publish the page
3. Using the Pages manager, copy the page and paste it as a child of the test page
4. Move the page back to the root - we now have two pages with the same slug (and both can be published)
## Expected behaviour
Slug should be checked and the page move should either not be allowed or slug should be automatically modified.
## Actual behaviour
Two pages with the same slug are allowed on the same level. Both can be published at the same time. The preview link in the Pages admin only points to the original (not to the copy).
## Additional information
Situation has been reproduced on the Divio platform using a clean project - https://cms-bug-demo-stage.us.aldryn.io/en/admin/cms/page/?language=en
|
1.0
|
Slug uniqueness not checked when moving page - ## Description
When page is moved, its slug uniqueness is not checked. Thus two pages with the same slug can be present on the same level.
## Steps to reproduce
1. Create a page at the root level and give it a slug "test".
2. Publish the page
3. Using the Pages manager, copy the page and paste it as a child of the test page
4. Move the page back to the root - we now have two pages with the same slug (and both can be published)
## Expected behaviour
Slug should be checked and the page move should either not be allowed or slug should be automatically modified.
## Actual behaviour
Two pages with the same slug are allowed on the same level. Both can be published at the same time. The preview link in the Pages admin only points to the original (not to the copy).
## Additional information
Situation has been reproduced on the Divio platform using a clean project - https://cms-bug-demo-stage.us.aldryn.io/en/admin/cms/page/?language=en
|
priority
|
slug uniqueness not checked when moving page description when page is moved its slug uniqueness is not checked thus two pages with the same slug can be present on the same level steps to reproduce create a page at the root level and give it a slug test publish the page using the pages manager copy the page and paste it as a child of the test page move the page back to the root we now have two pages with the same slug and both can be published expected behaviour slug should be checked and the page move should either not be allowed or slug should be automatically modified actual behaviour two pages with the same slug are allowed on the same level both can be published at the same time the preview link in the pages admin only points to the original not to the copy additional information situation has been reproduced on the divio platform using a clean project
| 1
|
197,539
| 6,960,714,765
|
IssuesEvent
|
2017-12-08 05:33:32
|
OpusCapita/react-crudeditor
|
https://api.github.com/repos/OpusCapita/react-crudeditor
|
closed
|
Allow instance to have fields not mentioned in model.fields
|
requested priority: high type:enhancement
|
Server-side may send an instance with fields, some of which are not described in model.fields.
|
1.0
|
Allow instance to have fields not mentioned in model.fields - Server-side may send an instance with fields, some of which are not described in model.fields.
|
priority
|
allow instance to have fields not mentioned in model fields server side may send an instance with fields some of which are not described in model fields
| 1
|
687,961
| 23,543,718,151
|
IssuesEvent
|
2022-08-20 20:05:05
|
AkjoStudios/AkjoNav
|
https://api.github.com/repos/AkjoStudios/AkjoNav
|
closed
|
[Feature Enhancement] Add common class for main tab panels
|
priority: high status: planned type: enhancement Official Issue
|
### In which module is the feature you want to enhance?
AkjoNav-DesktopMapper
### Feature
Main tab panels
### Description
Move all common features of the main tab panels into a separate class.
### Additional Context
No additional context/screenshots.
|
1.0
|
[Feature Enhancement] Add common class for main tab panels - ### In which module is the feature you want to enhance?
AkjoNav-DesktopMapper
### Feature
Main tab panels
### Description
Move all common features of the main tab panels into a separate class.
### Additional Context
No additional context/screenshots.
|
priority
|
add common class for main tab panels in which module is the feature you want to enhance akjonav desktopmapper feature main tab panels description move all common features of the main tab panels into a separate class additional context no additional context screenshots
| 1
|
593,123
| 17,947,811,571
|
IssuesEvent
|
2021-09-12 06:02:14
|
kiudee/chess-tuning-tools
|
https://api.github.com/repos/kiudee/chess-tuning-tools
|
closed
|
Divide by zero error when using input warping
|
bug Priority: High
|
* Chess Tuning Tools version: **0.6.0b2**
* Python version: **3.8**
* Operating System: vast.ai instance using the **nvidia/cuda:10.1-cudnn7-devel-ubuntu18.04** image.
### Description
I'm trying a tune of three parameters but turning on `warp_input` throws a divide by zero error every iteration and a weird warping for one of the parameters along with weird confidence bound for all parameters. (Everything looks normal without warping.)
**Error:**
```
/root/miniconda/envs/tuning/lib/python3.8/site-packages/sklearn/gaussian_process/kernels.py:255:
RuntimeWarning: divide by zero encountered in log
return np.log(np.hstack(theta))
```
**Without warping:**

**With warping:**

**Warp functions:**
```
Input warping was applied using the following parameters for the beta distributions:
{'ScaleQ': (0.191, 0.849), 'Cpuct': (0.471, 1.083), 'FpuValue': (0.635, 1.421)}
```

This warping for `ScaleQ` doesn't make sense but I suspect that it doesn't know what it's doing because the effect of this parameter is still not yet determined very well. The effect isn't nearly as strong as the other parameters and there is an expected interaction among all of them. I'll keep running non-warped `ts` and `vr` sampling for a while and try again once it starts to understand better.
Files to reproduce:
[DataAndConfig.zip](https://github.com/kiudee/chess-tuning-tools/files/5197274/DataAndConfig.zip)
### What I Did
Changing `"warp_inputs": false` to `"warp_inputs": true` in the `config.json` results in the runtime warning:
```
2020-09-09 18:34:09,833 INFO Got score: 5.6995265465242655 +- 1.9955240492535908
2020-09-09 18:34:09,834 INFO Updating model
/root/miniconda/envs/tuning/lib/python3.8/site-packages/sklearn/gaussian_process/kernels.py:255: RuntimeWarning: divide by zero encountered in log
return np.log(np.hstack(theta))
2020-09-09 18:34:28,397 INFO GP sampling finished (18.563089s)
2020-09-09 18:34:28,398 DEBUG GP kernel: 1.54**2 * Matern(length_scale=[0.522, 0.362, 0.276], nu=2.5) + WhiteKernel(noise_level=3.88e-06)
2020-09-09 18:34:28,399 DEBUG Input warping was applied using the following parameters for the beta distributions:
{'ScaleQ': (0.191, 0.849), 'Cpuct': (0.471, 1.083), 'FpuValue': (0.635, 1.421)}
2020-09-09 18:34:28,408 INFO Starting iteration 281
2020-09-09 18:34:41,482 INFO Current optimum:
{'ScaleQ': 0.3991959134189197, 'Cpuct': 4.700855456794591, 'FpuValue': 1.4722686789734403}
2020-09-09 18:34:41,482 INFO Estimated value: -0.5733 +- 0.3404
2020-09-09 18:34:41,482 INFO 90.0% confidence interval of the value: (-1.1332, -0.0133)
2020-09-09 18:34:41,648 INFO 90.0% confidence intervals of the parameters:
Parameter Lower bound Upper bound
---------------------------------------------------
ScaleQ 11.758489497804575 11.758489497804575
Cpuct 0.9963961100717935 0.9963961100717935
FpuValue -0.9812151210716575 -0.9812151210716575
```
---
Related to #92.
|
1.0
|
Divide by zero error when using input warping - * Chess Tuning Tools version: **0.6.0b2**
* Python version: **3.8**
* Operating System: vast.ai instance using the **nvidia/cuda:10.1-cudnn7-devel-ubuntu18.04** image.
### Description
I'm trying a tune of three parameters but turning on `warp_input` throws a divide by zero error every iteration and a weird warping for one of the parameters along with weird confidence bound for all parameters. (Everything looks normal without warping.)
**Error:**
```
/root/miniconda/envs/tuning/lib/python3.8/site-packages/sklearn/gaussian_process/kernels.py:255:
RuntimeWarning: divide by zero encountered in log
return np.log(np.hstack(theta))
```
**Without warping:**

**With warping:**

**Warp functions:**
```
Input warping was applied using the following parameters for the beta distributions:
{'ScaleQ': (0.191, 0.849), 'Cpuct': (0.471, 1.083), 'FpuValue': (0.635, 1.421)}
```

This warping for `ScaleQ` doesn't make sense but I suspect that it doesn't know what it's doing because the effect of this parameter is still not yet determined very well. The effect isn't nearly as strong as the other parameters and there is an expected interaction among all of them. I'll keep running non-warped `ts` and `vr` sampling for a while and try again once it starts to understand better.
Files to reproduce:
[DataAndConfig.zip](https://github.com/kiudee/chess-tuning-tools/files/5197274/DataAndConfig.zip)
### What I Did
Changing `"warp_inputs": false` to `"warp_inputs": true` in the `config.json` results in the runtime warning:
```
2020-09-09 18:34:09,833 INFO Got score: 5.6995265465242655 +- 1.9955240492535908
2020-09-09 18:34:09,834 INFO Updating model
/root/miniconda/envs/tuning/lib/python3.8/site-packages/sklearn/gaussian_process/kernels.py:255: RuntimeWarning: divide by zero encountered in log
return np.log(np.hstack(theta))
2020-09-09 18:34:28,397 INFO GP sampling finished (18.563089s)
2020-09-09 18:34:28,398 DEBUG GP kernel: 1.54**2 * Matern(length_scale=[0.522, 0.362, 0.276], nu=2.5) + WhiteKernel(noise_level=3.88e-06)
2020-09-09 18:34:28,399 DEBUG Input warping was applied using the following parameters for the beta distributions:
{'ScaleQ': (0.191, 0.849), 'Cpuct': (0.471, 1.083), 'FpuValue': (0.635, 1.421)}
2020-09-09 18:34:28,408 INFO Starting iteration 281
2020-09-09 18:34:41,482 INFO Current optimum:
{'ScaleQ': 0.3991959134189197, 'Cpuct': 4.700855456794591, 'FpuValue': 1.4722686789734403}
2020-09-09 18:34:41,482 INFO Estimated value: -0.5733 +- 0.3404
2020-09-09 18:34:41,482 INFO 90.0% confidence interval of the value: (-1.1332, -0.0133)
2020-09-09 18:34:41,648 INFO 90.0% confidence intervals of the parameters:
Parameter Lower bound Upper bound
---------------------------------------------------
ScaleQ 11.758489497804575 11.758489497804575
Cpuct 0.9963961100717935 0.9963961100717935
FpuValue -0.9812151210716575 -0.9812151210716575
```
---
Related to #92.
|
priority
|
divide by zero error when using input warping chess tuning tools version python version operating system vast ai instance using the nvidia cuda devel image description i m trying a tune of three parameters but turning on warp input throws a divide by zero error every iteration and a weird warping for one of the parameters along with weird confidence bound for all parameters everything looks normal without warping error root miniconda envs tuning lib site packages sklearn gaussian process kernels py runtimewarning divide by zero encountered in log return np log np hstack theta without warping with warping warp functions input warping was applied using the following parameters for the beta distributions scaleq cpuct fpuvalue this warping for scaleq doesn t make sense but i suspect that it doesn t know what it s doing because the effect of this parameter is still not yet determined very well the effect isn t nearly as strong as the other parameters and there is an expected interaction among all of them i ll keep running non warped ts and vr sampling for a while and try again once it starts to understand better files to reproduce what i did changing warp inputs false to warp inputs true in the config json results in the runtime warning info got score info updating model root miniconda envs tuning lib site packages sklearn gaussian process kernels py runtimewarning divide by zero encountered in log return np log np hstack theta info gp sampling finished debug gp kernel matern length scale nu whitekernel noise level debug input warping was applied using the following parameters for the beta distributions scaleq cpuct fpuvalue info starting iteration info current optimum scaleq cpuct fpuvalue info estimated value info confidence interval of the value info confidence intervals of the parameters parameter lower bound upper bound scaleq cpuct fpuvalue related to
| 1
|
452,538
| 13,055,436,003
|
IssuesEvent
|
2020-07-30 01:40:51
|
Azure/vscode-kubernetes-tools
|
https://api.github.com/repos/Azure/vscode-kubernetes-tools
|
closed
|
Terminal shell path broken
|
bug priority-high
|
Since upgrading to latest version I get a
`The terminal shell path ""C:\Program Files (x86)\.vs-kubernetes\tools\kubectl\kubectl".exe" does not exist`
The error is only when right clicking and trying to initiate a terminal session... this used to work well before. All other commands work.
|
1.0
|
Terminal shell path broken - Since upgrading to latest version I get a
`The terminal shell path ""C:\Program Files (x86)\.vs-kubernetes\tools\kubectl\kubectl".exe" does not exist`
The error is only when right clicking and trying to initiate a terminal session... this used to work well before. All other commands work.
|
priority
|
terminal shell path broken since upgrading to latest version i get a the terminal shell path c program files vs kubernetes tools kubectl kubectl exe does not exist the error is only when right clicking and trying to initiate a terminal session this used to work well before all other commands work
| 1
|
481,789
| 13,891,930,819
|
IssuesEvent
|
2020-10-19 11:26:23
|
CLIxIndia-Dev/clixoer
|
https://api.github.com/repos/CLIxIndia-Dev/clixoer
|
closed
|
The COOL PREVIEW Changes
|
enhancement frontend highpriority
|
**The cool preview changes **
This task contains changes suggestion about cool page preview
- The arrange the name before the Button in cool page preview
- The Affordance icon need to responsive clear in mobile view
- change in copy share button ( Twitter )
|
1.0
|
The COOL PREVIEW Changes - **The cool preview changes **
This task contains changes suggestion about cool page preview
- The arrange the name before the Button in cool page preview
- The Affordance icon need to responsive clear in mobile view
- change in copy share button ( Twitter )
|
priority
|
the cool preview changes the cool preview changes this task contains changes suggestion about cool page preview the arrange the name before the button in cool page preview the affordance icon need to responsive clear in mobile view change in copy share button twitter
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.