Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
354,932
| 25,175,238,220
|
IssuesEvent
|
2022-11-11 08:38:57
|
garfield-oo7/pe
|
https://api.github.com/repos/garfield-oo7/pe
|
opened
|
view, parameter issue
|
severity.Low type.DocumentationBug
|
In the user guide it is mentioned that the command view takes in one parameter index, and index must be a positive integer, however when you try to enter the number 0, which can be considered a positive integer, it throws an error saying that please enter a positive integer.
I think it will be nice if you can specify in the user guide that view takes in a paramter index, which takes in integer values greater than or equal to 1.


<!--session: 1668155718992-d5714102-bc2d-493e-8161-1c9ed5f3d971-->
<!--Version: Web v3.4.4-->
|
1.0
|
view, parameter issue - In the user guide it is mentioned that the command view takes in one parameter index, and index must be a positive integer, however when you try to enter the number 0, which can be considered a positive integer, it throws an error saying that please enter a positive integer.
I think it will be nice if you can specify in the user guide that view takes in a paramter index, which takes in integer values greater than or equal to 1.


<!--session: 1668155718992-d5714102-bc2d-493e-8161-1c9ed5f3d971-->
<!--Version: Web v3.4.4-->
|
non_process
|
view parameter issue in the user guide it is mentioned that the command view takes in one parameter index and index must be a positive integer however when you try to enter the number which can be considered a positive integer it throws an error saying that please enter a positive integer i think it will be nice if you can specify in the user guide that view takes in a paramter index which takes in integer values greater than or equal to
| 0
|
6,175
| 9,084,139,203
|
IssuesEvent
|
2019-02-18 01:56:37
|
googleapis/google-cloud-cpp
|
https://api.github.com/repos/googleapis/google-cloud-cpp
|
closed
|
Flaky test: InstanceAdminIntegrationTest.*
|
priority: p2 type: process
|
The integration tests fail from time to time, but they run fine a second time. It must be some kind of race condition, but as they execute against the emulator it is a more subtle one that the typical "used the same resources from two simultaneous builds".
The latest failure was in:
https://travis-ci.org/GoogleCloudPlatform/google-cloud-cpp/jobs/468482203
The relevant log lines are:
```
Running bigtable::InstanceAdmin integration test.
[==========] Running 5 tests from 1 test case.
[----------] Global test environment set-up.
[----------] 5 tests from InstanceAdminIntegrationTest
[ RUN ] InstanceAdminIntegrationTest.ListAllClustersTest
unknown file: Failure
C++ exception with description "InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904709) instance missing: InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904709) instance missing [5=NOT_FOUND] - " thrown in the test body.
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest (32 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile (11 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest (9 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest (9 ms)
[ RUN ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest
[ OK ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest (6 ms)
[----------] 5 tests from InstanceAdminIntegrationTest (67 ms total)
[----------] Global test environment tear-down
[==========] 5 tests from 1 test case ran. (68 ms total)
[ PASSED ] 4 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest
1 FAILED TEST
Killing Bigtable Emulators [27247 27248] .+..+. done.
Failure or timeout in integration test during attempt=1 Sat Dec 15 20:11:49 UTC 2018
Launching Cloud Bigtable emulators in the background
Successfully connected to the Cloud Bigtable emulator.
Successfully connected to the Cloud Bigtable Instance Admin emulator.
Running bigtable::InstanceAdmin integration test.
[==========] Running 5 tests from 1 test case.
[----------] Global test environment set-up.
[----------] 5 tests from InstanceAdminIntegrationTest
[ RUN ] InstanceAdminIntegrationTest.ListAllClustersTest
unknown file: Failure
C++ exception with description "InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904710) instance missing: InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904710) instance missing [5=NOT_FOUND] - " thrown in the test body.
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest (7 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile (8 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest (5 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest (6 ms)
[ RUN ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest
[ OK ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest (5 ms)
[----------] 5 tests from InstanceAdminIntegrationTest (31 ms total)
[----------] Global test environment tear-down
[==========] 5 tests from 1 test case ran. (31 ms total)
[ PASSED ] 4 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest
1 FAILED TEST
Killing Bigtable Emulators [27304 27305] .+..+. done.
Failure or timeout in integration test during attempt=2 Sat Dec 15 20:11:50 UTC 2018
Launching Cloud Bigtable emulators in the background
Successfully connected to the Cloud Bigtable emulator.
Successfully connected to the Cloud Bigtable Instance Admin emulator.
Running bigtable::InstanceAdmin integration test.
[==========] Running 5 tests from 1 test case.
[----------] Global test environment set-up.
[----------] 5 tests from InstanceAdminIntegrationTest
[ RUN ] InstanceAdminIntegrationTest.ListAllClustersTest
unknown file: Failure
C++ exception with description "InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904710) instance missing: InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904710) instance missing [5=NOT_FOUND] - " thrown in the test body.
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest (8 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile (7 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest (5 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest (6 ms)
[ RUN ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest
[ OK ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest (5 ms)
[----------] 5 tests from InstanceAdminIntegrationTest (31 ms total)
[----------] Global test environment tear-down
[==========] 5 tests from 1 test case ran. (31 ms total)
[ PASSED ] 4 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest
1 FAILED TEST
Killing Bigtable Emulators [27361 27362] .+..+. done.
Failure or timeout in integration test during attempt=3 Sat Dec 15 20:11:50 UTC 2018
Integration tests failed multiple times, aborting tests.
The command "ci/travis/build-linux.sh" exited with 1.
```
|
1.0
|
Flaky test: InstanceAdminIntegrationTest.* - The integration tests fail from time to time, but they run fine a second time. It must be some kind of race condition, but as they execute against the emulator it is a more subtle one that the typical "used the same resources from two simultaneous builds".
The latest failure was in:
https://travis-ci.org/GoogleCloudPlatform/google-cloud-cpp/jobs/468482203
The relevant log lines are:
```
Running bigtable::InstanceAdmin integration test.
[==========] Running 5 tests from 1 test case.
[----------] Global test environment set-up.
[----------] 5 tests from InstanceAdminIntegrationTest
[ RUN ] InstanceAdminIntegrationTest.ListAllClustersTest
unknown file: Failure
C++ exception with description "InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904709) instance missing: InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904709) instance missing [5=NOT_FOUND] - " thrown in the test body.
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest (32 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile (11 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest (9 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest (9 ms)
[ RUN ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest
[ OK ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest (6 ms)
[----------] 5 tests from InstanceAdminIntegrationTest (67 ms total)
[----------] Global test environment tear-down
[==========] 5 tests from 1 test case ran. (68 ms total)
[ PASSED ] 4 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest
1 FAILED TEST
Killing Bigtable Emulators [27247 27248] .+..+. done.
Failure or timeout in integration test during attempt=1 Sat Dec 15 20:11:49 UTC 2018
Launching Cloud Bigtable emulators in the background
Successfully connected to the Cloud Bigtable emulator.
Successfully connected to the Cloud Bigtable Instance Admin emulator.
Running bigtable::InstanceAdmin integration test.
[==========] Running 5 tests from 1 test case.
[----------] Global test environment set-up.
[----------] 5 tests from InstanceAdminIntegrationTest
[ RUN ] InstanceAdminIntegrationTest.ListAllClustersTest
unknown file: Failure
C++ exception with description "InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904710) instance missing: InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904710) instance missing [5=NOT_FOUND] - " thrown in the test body.
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest (7 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile (8 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest (5 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest (6 ms)
[ RUN ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest
[ OK ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest (5 ms)
[----------] 5 tests from InstanceAdminIntegrationTest (31 ms total)
[----------] Global test environment tear-down
[==========] 5 tests from 1 test case ran. (31 ms total)
[ PASSED ] 4 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest
1 FAILED TEST
Killing Bigtable Emulators [27304 27305] .+..+. done.
Failure or timeout in integration test during attempt=2 Sat Dec 15 20:11:50 UTC 2018
Launching Cloud Bigtable emulators in the background
Successfully connected to the Cloud Bigtable emulator.
Successfully connected to the Cloud Bigtable Instance Admin emulator.
Running bigtable::InstanceAdmin integration test.
[==========] Running 5 tests from 1 test case.
[----------] Global test environment set-up.
[----------] 5 tests from InstanceAdminIntegrationTest
[ RUN ] InstanceAdminIntegrationTest.ListAllClustersTest
unknown file: Failure
C++ exception with description "InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904710) instance missing: InstanceAdmin::DeleteInstance(parent=projects/emulated-1544904710) instance missing [5=NOT_FOUND] - " thrown in the test body.
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest (8 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteAppProfile (7 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteInstanceTest (5 ms)
[ RUN ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest
[ OK ] InstanceAdminIntegrationTest.CreateListGetDeleteClusterTest (6 ms)
[ RUN ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest
[ OK ] InstanceAdminIntegrationTest.SetGetTestIamAPIsTest (5 ms)
[----------] 5 tests from InstanceAdminIntegrationTest (31 ms total)
[----------] Global test environment tear-down
[==========] 5 tests from 1 test case ran. (31 ms total)
[ PASSED ] 4 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] InstanceAdminIntegrationTest.ListAllClustersTest
1 FAILED TEST
Killing Bigtable Emulators [27361 27362] .+..+. done.
Failure or timeout in integration test during attempt=3 Sat Dec 15 20:11:50 UTC 2018
Integration tests failed multiple times, aborting tests.
The command "ci/travis/build-linux.sh" exited with 1.
```
|
process
|
flaky test instanceadminintegrationtest the integration tests fail from time to time but they run fine a second time it must be some kind of race condition but as they execute against the emulator it is a more subtle one that the typical used the same resources from two simultaneous builds the latest failure was in the relevant log lines are running bigtable instanceadmin integration test running tests from test case global test environment set up tests from instanceadminintegrationtest instanceadminintegrationtest listallclusterstest unknown file failure c exception with description instanceadmin deleteinstance parent projects emulated instance missing instanceadmin deleteinstance parent projects emulated instance missing thrown in the test body instanceadminintegrationtest listallclusterstest ms instanceadminintegrationtest createlistgetdeleteappprofile instanceadminintegrationtest createlistgetdeleteappprofile ms instanceadminintegrationtest createlistgetdeleteinstancetest instanceadminintegrationtest createlistgetdeleteinstancetest ms instanceadminintegrationtest createlistgetdeleteclustertest instanceadminintegrationtest createlistgetdeleteclustertest ms instanceadminintegrationtest setgettestiamapistest instanceadminintegrationtest setgettestiamapistest ms tests from instanceadminintegrationtest ms total global test environment tear down tests from test case ran ms total tests test listed below instanceadminintegrationtest listallclusterstest failed test killing bigtable emulators done failure or timeout in integration test during attempt sat dec utc launching cloud bigtable emulators in the background successfully connected to the cloud bigtable emulator successfully connected to the cloud bigtable instance admin emulator running bigtable instanceadmin integration test running tests from test case global test environment set up tests from instanceadminintegrationtest instanceadminintegrationtest listallclusterstest unknown file failure c exception with description instanceadmin deleteinstance parent projects emulated instance missing instanceadmin deleteinstance parent projects emulated instance missing thrown in the test body instanceadminintegrationtest listallclusterstest ms instanceadminintegrationtest createlistgetdeleteappprofile instanceadminintegrationtest createlistgetdeleteappprofile ms instanceadminintegrationtest createlistgetdeleteinstancetest instanceadminintegrationtest createlistgetdeleteinstancetest ms instanceadminintegrationtest createlistgetdeleteclustertest instanceadminintegrationtest createlistgetdeleteclustertest ms instanceadminintegrationtest setgettestiamapistest instanceadminintegrationtest setgettestiamapistest ms tests from instanceadminintegrationtest ms total global test environment tear down tests from test case ran ms total tests test listed below instanceadminintegrationtest listallclusterstest failed test killing bigtable emulators done failure or timeout in integration test during attempt sat dec utc launching cloud bigtable emulators in the background successfully connected to the cloud bigtable emulator successfully connected to the cloud bigtable instance admin emulator running bigtable instanceadmin integration test running tests from test case global test environment set up tests from instanceadminintegrationtest instanceadminintegrationtest listallclusterstest unknown file failure c exception with description instanceadmin deleteinstance parent projects emulated instance missing instanceadmin deleteinstance parent projects emulated instance missing thrown in the test body instanceadminintegrationtest listallclusterstest ms instanceadminintegrationtest createlistgetdeleteappprofile instanceadminintegrationtest createlistgetdeleteappprofile ms instanceadminintegrationtest createlistgetdeleteinstancetest instanceadminintegrationtest createlistgetdeleteinstancetest ms instanceadminintegrationtest createlistgetdeleteclustertest instanceadminintegrationtest createlistgetdeleteclustertest ms instanceadminintegrationtest setgettestiamapistest instanceadminintegrationtest setgettestiamapistest ms tests from instanceadminintegrationtest ms total global test environment tear down tests from test case ran ms total tests test listed below instanceadminintegrationtest listallclusterstest failed test killing bigtable emulators done failure or timeout in integration test during attempt sat dec utc integration tests failed multiple times aborting tests the command ci travis build linux sh exited with
| 1
|
5,295
| 7,107,028,087
|
IssuesEvent
|
2018-01-16 18:31:43
|
microsoftgraph/msgraph-sdk-dotnet
|
https://api.github.com/repos/microsoftgraph/msgraph-sdk-dotnet
|
closed
|
Values larger than Int32.Max are truncated when sent to Excel sheets
|
service bug under investigation
|
I'm using the Microsoft Graph API to modify cells in Excel spreadsheets. When writing integer values larger than Int32.Max to the spreadsheet via the API, the upper 32 bits of the values appear to be truncated, and only the lower 32 bits seem to be written to the spreadsheet.
Below are the values that I have tested.
Test 1:
Value sent: 1,234,000,000,000 (0x11F5021B400)
Value written to cell: 1344386048 (0x5021B400)
The upper bits appear to have been truncated, leaving behind only the lower 32 bits.
Test 2:
Value sent: 2,147,483,648 (Int32.Max + 1) (0x80000000)
Value written to the cell: -2147483648
If 0x80000000 is interpreted as a signed Int32 number, then it is -2147483648
Test 3:
Value sent: 4294967296 (0x100000000)
Value written to the cell: 0
When the highest bit of 0x100000000 is truncated, the remaining 32 bits become 0 in signed Int32
Test 4:
Value sent: 4294967295 (0xFFFFFFFF)
Value written to the cell: -1
When interpreted as a signed Int32 value, 0xFFFFFFFF becomes -1
All of the values were sent using the .NET Microsoft Graph Client library, and were of type long. Specifically, the values were sent using the Microsoft.Graph.WorkbookRange class. Setting the value type of the cells in the WorkbookRange object to Double did not fix the`problem.
Thanks in advance for any help or suggestion.
|
1.0
|
Values larger than Int32.Max are truncated when sent to Excel sheets - I'm using the Microsoft Graph API to modify cells in Excel spreadsheets. When writing integer values larger than Int32.Max to the spreadsheet via the API, the upper 32 bits of the values appear to be truncated, and only the lower 32 bits seem to be written to the spreadsheet.
Below are the values that I have tested.
Test 1:
Value sent: 1,234,000,000,000 (0x11F5021B400)
Value written to cell: 1344386048 (0x5021B400)
The upper bits appear to have been truncated, leaving behind only the lower 32 bits.
Test 2:
Value sent: 2,147,483,648 (Int32.Max + 1) (0x80000000)
Value written to the cell: -2147483648
If 0x80000000 is interpreted as a signed Int32 number, then it is -2147483648
Test 3:
Value sent: 4294967296 (0x100000000)
Value written to the cell: 0
When the highest bit of 0x100000000 is truncated, the remaining 32 bits become 0 in signed Int32
Test 4:
Value sent: 4294967295 (0xFFFFFFFF)
Value written to the cell: -1
When interpreted as a signed Int32 value, 0xFFFFFFFF becomes -1
All of the values were sent using the .NET Microsoft Graph Client library, and were of type long. Specifically, the values were sent using the Microsoft.Graph.WorkbookRange class. Setting the value type of the cells in the WorkbookRange object to Double did not fix the`problem.
Thanks in advance for any help or suggestion.
|
non_process
|
values larger than max are truncated when sent to excel sheets i m using the microsoft graph api to modify cells in excel spreadsheets when writing integer values larger than max to the spreadsheet via the api the upper bits of the values appear to be truncated and only the lower bits seem to be written to the spreadsheet below are the values that i have tested test value sent value written to cell the upper bits appear to have been truncated leaving behind only the lower bits test value sent max value written to the cell if is interpreted as a signed number then it is test value sent value written to the cell when the highest bit of is truncated the remaining bits become in signed test value sent value written to the cell when interpreted as a signed value becomes all of the values were sent using the net microsoft graph client library and were of type long specifically the values were sent using the microsoft graph workbookrange class setting the value type of the cells in the workbookrange object to double did not fix the problem thanks in advance for any help or suggestion
| 0
|
652
| 3,120,796,281
|
IssuesEvent
|
2015-09-05 02:14:14
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
[Converge] child_process argument type checking
|
child_process
|
Continuing from https://github.com/nodejs/node-convergence-archive/issues/22, I don't believe this has been done yet, @jasnell can you confirm please? That thread seems to have enough agreement to pull these changes in. Marking on the 4.0.0 milestone.
> These commits add argument type checking to methods within child_process. This needs to be reconciled with the current io.js behavior. The change introduces a new throw so there's a potential API compatibility issue.
> /cc @trevnorris
>
> * `a1b2875afd9affd3e0147a10f04a2daf8c598761 lint: fix lint issues` - https://github.com/joyent/node/commit/a1b2875afd9affd3e0147a10f04a2daf8c598761
> * `8032a210250c86877898331ac4474f39b3a573c6 test: test all spawn parameter positions` - https://github.com/joyent/node/commit/8032a210250c86877898331ac4474f39b3a573c6
> * `70dafa7b624abd43432e03304d65cc527fbecc11 child_process: check fork args is an array` - https://github.com/joyent/node/commit/70dafa7b624abd43432e03304d65cc527fbecc11
> * `e17c5a72b23f920f291d61f2780068c18768cb92 child_process: check execFile args is an array` - https://github.com/joyent/node/commit/e17c5a72b23f920f291d61f2780068c18768cb92
|
1.0
|
[Converge] child_process argument type checking - Continuing from https://github.com/nodejs/node-convergence-archive/issues/22, I don't believe this has been done yet, @jasnell can you confirm please? That thread seems to have enough agreement to pull these changes in. Marking on the 4.0.0 milestone.
> These commits add argument type checking to methods within child_process. This needs to be reconciled with the current io.js behavior. The change introduces a new throw so there's a potential API compatibility issue.
> /cc @trevnorris
>
> * `a1b2875afd9affd3e0147a10f04a2daf8c598761 lint: fix lint issues` - https://github.com/joyent/node/commit/a1b2875afd9affd3e0147a10f04a2daf8c598761
> * `8032a210250c86877898331ac4474f39b3a573c6 test: test all spawn parameter positions` - https://github.com/joyent/node/commit/8032a210250c86877898331ac4474f39b3a573c6
> * `70dafa7b624abd43432e03304d65cc527fbecc11 child_process: check fork args is an array` - https://github.com/joyent/node/commit/70dafa7b624abd43432e03304d65cc527fbecc11
> * `e17c5a72b23f920f291d61f2780068c18768cb92 child_process: check execFile args is an array` - https://github.com/joyent/node/commit/e17c5a72b23f920f291d61f2780068c18768cb92
|
process
|
child process argument type checking continuing from i don t believe this has been done yet jasnell can you confirm please that thread seems to have enough agreement to pull these changes in marking on the milestone these commits add argument type checking to methods within child process this needs to be reconciled with the current io js behavior the change introduces a new throw so there s a potential api compatibility issue cc trevnorris lint fix lint issues test test all spawn parameter positions child process check fork args is an array child process check execfile args is an array
| 1
|
134,125
| 29,832,673,700
|
IssuesEvent
|
2023-06-18 13:11:43
|
llvm/llvm-project
|
https://api.github.com/repos/llvm/llvm-project
|
closed
|
clang 3.5 and 3.4 crash in clang::CodeGen::CodeGenTypes::ConvertType(clang::QualType)
|
c++ clang:codegen bugzilla worksforme
|
| | |
| --- | --- |
| Bugzilla Link | [19532](https://llvm.org/bz19532) |
| Version | trunk |
| OS | Linux |
| Reporter | LLVM Bugzilla Contributor |
| CC | @DougGregor,@rnk |
## Extended Description
Compiling https://github.com/beniz/libcmaes with clang 3.4 and clang 3.5 on Linux and on Mac OSX does inevitably crash clang with following error:
0 libLLVM-3.5.so.1 0x00007fc78d6907e2 llvm::sys::PrintStackTrace(_IO_FILE*) + 34
1 libLLVM-3.5.so.1 0x00007fc78d6905d4
2 libpthread.so.0 0x00007fc78c526340
3 clang 0x0000000000863d5e clang::CodeGen::CodeGenTypes::ConvertType(clang::QualType) + 1934
4 clang 0x00000000008a0ade clang::CodeGen::CodeGenTypes::arrangeLLVMFunctionInfo(clang::CanQual<clang::Type>, bool, llvm::ArrayRef<clang::CanQual<clang::Type> >, clang::FunctionType::ExtInfo, clang::CodeGen::RequiredArgs) + 606
5 clang 0x00000000008a7fe1
6 clang 0x00000000008a8810 clang::CodeGen::CodeGenTypes::arrangeCXXMethodType(clang::CXXRecordDecl const*, clang::FunctionProtoType const*) + 192
7 clang 0x0000000000849ed3 clang::CodeGen::CodeGenModule::EmitGlobalFunctionDefinition(clang::GlobalDecl, llvm::GlobalValue*) + 67
8 clang 0x000000000084ac64 clang::CodeGen::CodeGenModule::EmitGlobalDefinition(clang::GlobalDecl, llvm::GlobalValue*) + 244
9 clang 0x000000000084d42d clang::CodeGen::CodeGenModule::EmitGlobal(clang::GlobalDecl) + 1341
10 clang 0x000000000084dc33
11 clang 0x00000000007f5643
12 clang 0x00000000007f30dc
13 clang 0x0000000000cbece5 clang::Sema::InstantiateFunctionDefinition(clang::SourceLocation, clang::FunctionDecl*, bool, bool) + 1813
14 clang 0x0000000000c98fe2 clang::Sema::InstantiateClassMembers(clang::SourceLocation, clang::CXXRecordDecl*, clang::MultiLevelTemplateArgumentList const&, clang::TemplateSpecializationKind) + 898
15 clang 0x0000000000c98ef3 clang::Sema::InstantiateClassMembers(clang::SourceLocation, clang::CXXRecordDecl*, clang::MultiLevelTemplateArgumentList const&, clang::TemplateSpecializationKind) + 659
16 clang 0x0000000000c99196 clang::Sema::InstantiateClassTemplateSpecializationMembers(clang::SourceLocation, clang::ClassTemplateSpecializationDecl*, clang::TemplateSpecializationKind) + 86
17 clang 0x0000000000c40148 clang::Sema::ActOnExplicitInstantiation(clang::Scope*, clang::SourceLocation, clang::SourceLocation, unsigned int, clang::SourceLocation, clang::CXXScopeSpec const&, clang::OpaquePtr<clang::TemplateName>, clang::SourceLocation, clang::SourceLocation, llvm::MutableArrayRef<clang::ParsedTemplateArgument>, clang::SourceLocation, clang::AttributeList*) + 1432
18 clang 0x000000000099094c clang::Parser::ParseClassSpecifier(clang::tok::TokenKind, clang::SourceLocation, clang::DeclSpec&, clang::Parser::ParsedTemplateInfo const&, clang::AccessSpecifier, bool, clang::Parser::DeclSpecContext, clang::Parser::ParsedAttributesWithRange&) + 5180
19 clang 0x0000000000979f0e clang::Parser::ParseDeclarationSpecifiers(clang::DeclSpec&, clang::Parser::ParsedTemplateInfo const&, clang::AccessSpecifier, clang::Parser::DeclSpecContext, clang::Parser::LateParsedAttrList*) + 2542
20 clang 0x00000000009cc59d clang::Parser::ParseSingleDeclarationAfterTemplate(unsigned int, clang::Parser::ParsedTemplateInfo const&, clang::ParsingDeclRAIIObject&, clang::SourceLocation&, clang::AccessSpecifier, clang::AttributeList*) + 909
21 clang 0x00000000009cde19 clang::Parser::ParseExplicitInstantiation(unsigned int, clang::SourceLocation, clang::SourceLocation, clang::SourceLocation&, clang::AccessSpecifier) + 169
22 clang 0x00000000009cdfdc clang::Parser::ParseDeclarationStartingWithTemplate(unsigned int, clang::SourceLocation&, clang::AccessSpecifier, clang::AttributeList*) + 300
23 clang 0x000000000097e9b8 clang::Parser::ParseDeclaration(llvm::SmallVector<clang::Stmt*, 32u>&, unsigned int, clang::SourceLocation&, clang::Parser::ParsedAttributesWithRange&) + 584
24 clang 0x0000000000968544 clang::Parser::ParseExternalDeclaration(clang::Parser::ParsedAttributesWithRange&, clang::ParsingDeclSpec*) + 372
25 clang 0x0000000000989458 clang::Parser::ParseInnerNamespace(std::vector<clang::SourceLocation, std::allocator<clang::SourceLocation> >&, std::vector<clang::IdentifierInfo*, std::allocator<clang::IdentifierInfo*> >&, std::vector<clang::SourceLocation, std::allocator<clang::SourceLocation> >&, unsigned int, clang::SourceLocation&, clang::ParsedAttributes&, clang::BalancedDelimiterTracker&) + 408
26 clang 0x0000000000989d0e clang::Parser::ParseNamespace(unsigned int, clang::SourceLocation&, clang::SourceLocation) + 2014
27 clang 0x000000000097ea10 clang::Parser::ParseDeclaration(llvm::SmallVector<clang::Stmt*, 32u>&, unsigned int, clang::SourceLocation&, clang::Parser::ParsedAttributesWithRange&) + 672
28 clang 0x0000000000968544 clang::Parser::ParseExternalDeclaration(clang::Parser::ParsedAttributesWithRange&, clang::ParsingDeclSpec*) + 372
29 clang 0x0000000000968c8c clang::Parser::ParseTopLevelDecl(clang::OpaquePtr<clang::DeclGroupRef>&) + 428
30 clang 0x000000000096027b clang::ParseAST(clang::Sema&, bool, bool) + 283
31 clang 0x000000000068f409 clang::FrontendAction::Execute() + 169
32 clang 0x000000000066f795 clang::CompilerInstance::ExecuteAction(clang::FrontendAction&) + 293
33 clang 0x000000000065a276 clang::ExecuteCompilerInvocation(clang::CompilerInstance*) + 1654
34 clang 0x0000000000654e48 cc1_main(char const**, char const**, char const*, void*) + 1176
35 clang 0x0000000000653c10 main + 4640
36 libc.so.6 0x00007fc78bc57ec5 __libc_start_main + 245
37 clang 0x0000000000653e7f
Stack dump:
0. Program arguments: /usr/bin/clang -cc1 -triple x86_64-pc-linux-gnu -emit-obj -disable-free -disable-llvm-verifier -main-file-name cmastrategy.cc -mrelocation-model pic -pic-level 2 -fmath-errno -masm-verbose -mconstructor-aliases -munwind-tables -fuse-init-array -target-cpu x86-64 -target-linker-version 2.24 -momit-leaf-frame-pointer -g -sys-header-deps -D HAVE_CONFIG_H -D PIC -O2 -Wall -std=c++11 -fdeprecated-macro -ferror-limit 19 -fmessage-length 112 -mstackrealign -fobjc-runtime=gcc -fcxx-exceptions -fexceptions -fdiagnostics-show-option -fcolor-diagnostics -vectorize-loops -vectorize-slp -x c++ cmastrategy-dac114.cpp
1. cmastrategy.cc:163:47: current parser token ';'
2. cmastrategy.cc:29:1: parsing namespace 'libcmaes'
3. cmastrategy.cc:33:92: LLVM IR generation of declaration 'libcmaes::CMAStrategy<libcmaes::CovarianceUpdate>::<anonymous class>::operator()'
4. cmastrategy.cc:33:92: Generating code for declaration 'libcmaes::CMAStrategy<libcmaes::CovarianceUpdate>::<anonymous class>::operator()'
Segmentation fault (core dumped)
|
1.0
|
clang 3.5 and 3.4 crash in clang::CodeGen::CodeGenTypes::ConvertType(clang::QualType) - | | |
| --- | --- |
| Bugzilla Link | [19532](https://llvm.org/bz19532) |
| Version | trunk |
| OS | Linux |
| Reporter | LLVM Bugzilla Contributor |
| CC | @DougGregor,@rnk |
## Extended Description
Compiling https://github.com/beniz/libcmaes with clang 3.4 and clang 3.5 on Linux and on Mac OSX does inevitably crash clang with following error:
0 libLLVM-3.5.so.1 0x00007fc78d6907e2 llvm::sys::PrintStackTrace(_IO_FILE*) + 34
1 libLLVM-3.5.so.1 0x00007fc78d6905d4
2 libpthread.so.0 0x00007fc78c526340
3 clang 0x0000000000863d5e clang::CodeGen::CodeGenTypes::ConvertType(clang::QualType) + 1934
4 clang 0x00000000008a0ade clang::CodeGen::CodeGenTypes::arrangeLLVMFunctionInfo(clang::CanQual<clang::Type>, bool, llvm::ArrayRef<clang::CanQual<clang::Type> >, clang::FunctionType::ExtInfo, clang::CodeGen::RequiredArgs) + 606
5 clang 0x00000000008a7fe1
6 clang 0x00000000008a8810 clang::CodeGen::CodeGenTypes::arrangeCXXMethodType(clang::CXXRecordDecl const*, clang::FunctionProtoType const*) + 192
7 clang 0x0000000000849ed3 clang::CodeGen::CodeGenModule::EmitGlobalFunctionDefinition(clang::GlobalDecl, llvm::GlobalValue*) + 67
8 clang 0x000000000084ac64 clang::CodeGen::CodeGenModule::EmitGlobalDefinition(clang::GlobalDecl, llvm::GlobalValue*) + 244
9 clang 0x000000000084d42d clang::CodeGen::CodeGenModule::EmitGlobal(clang::GlobalDecl) + 1341
10 clang 0x000000000084dc33
11 clang 0x00000000007f5643
12 clang 0x00000000007f30dc
13 clang 0x0000000000cbece5 clang::Sema::InstantiateFunctionDefinition(clang::SourceLocation, clang::FunctionDecl*, bool, bool) + 1813
14 clang 0x0000000000c98fe2 clang::Sema::InstantiateClassMembers(clang::SourceLocation, clang::CXXRecordDecl*, clang::MultiLevelTemplateArgumentList const&, clang::TemplateSpecializationKind) + 898
15 clang 0x0000000000c98ef3 clang::Sema::InstantiateClassMembers(clang::SourceLocation, clang::CXXRecordDecl*, clang::MultiLevelTemplateArgumentList const&, clang::TemplateSpecializationKind) + 659
16 clang 0x0000000000c99196 clang::Sema::InstantiateClassTemplateSpecializationMembers(clang::SourceLocation, clang::ClassTemplateSpecializationDecl*, clang::TemplateSpecializationKind) + 86
17 clang 0x0000000000c40148 clang::Sema::ActOnExplicitInstantiation(clang::Scope*, clang::SourceLocation, clang::SourceLocation, unsigned int, clang::SourceLocation, clang::CXXScopeSpec const&, clang::OpaquePtr<clang::TemplateName>, clang::SourceLocation, clang::SourceLocation, llvm::MutableArrayRef<clang::ParsedTemplateArgument>, clang::SourceLocation, clang::AttributeList*) + 1432
18 clang 0x000000000099094c clang::Parser::ParseClassSpecifier(clang::tok::TokenKind, clang::SourceLocation, clang::DeclSpec&, clang::Parser::ParsedTemplateInfo const&, clang::AccessSpecifier, bool, clang::Parser::DeclSpecContext, clang::Parser::ParsedAttributesWithRange&) + 5180
19 clang 0x0000000000979f0e clang::Parser::ParseDeclarationSpecifiers(clang::DeclSpec&, clang::Parser::ParsedTemplateInfo const&, clang::AccessSpecifier, clang::Parser::DeclSpecContext, clang::Parser::LateParsedAttrList*) + 2542
20 clang 0x00000000009cc59d clang::Parser::ParseSingleDeclarationAfterTemplate(unsigned int, clang::Parser::ParsedTemplateInfo const&, clang::ParsingDeclRAIIObject&, clang::SourceLocation&, clang::AccessSpecifier, clang::AttributeList*) + 909
21 clang 0x00000000009cde19 clang::Parser::ParseExplicitInstantiation(unsigned int, clang::SourceLocation, clang::SourceLocation, clang::SourceLocation&, clang::AccessSpecifier) + 169
22 clang 0x00000000009cdfdc clang::Parser::ParseDeclarationStartingWithTemplate(unsigned int, clang::SourceLocation&, clang::AccessSpecifier, clang::AttributeList*) + 300
23 clang 0x000000000097e9b8 clang::Parser::ParseDeclaration(llvm::SmallVector<clang::Stmt*, 32u>&, unsigned int, clang::SourceLocation&, clang::Parser::ParsedAttributesWithRange&) + 584
24 clang 0x0000000000968544 clang::Parser::ParseExternalDeclaration(clang::Parser::ParsedAttributesWithRange&, clang::ParsingDeclSpec*) + 372
25 clang 0x0000000000989458 clang::Parser::ParseInnerNamespace(std::vector<clang::SourceLocation, std::allocator<clang::SourceLocation> >&, std::vector<clang::IdentifierInfo*, std::allocator<clang::IdentifierInfo*> >&, std::vector<clang::SourceLocation, std::allocator<clang::SourceLocation> >&, unsigned int, clang::SourceLocation&, clang::ParsedAttributes&, clang::BalancedDelimiterTracker&) + 408
26 clang 0x0000000000989d0e clang::Parser::ParseNamespace(unsigned int, clang::SourceLocation&, clang::SourceLocation) + 2014
27 clang 0x000000000097ea10 clang::Parser::ParseDeclaration(llvm::SmallVector<clang::Stmt*, 32u>&, unsigned int, clang::SourceLocation&, clang::Parser::ParsedAttributesWithRange&) + 672
28 clang 0x0000000000968544 clang::Parser::ParseExternalDeclaration(clang::Parser::ParsedAttributesWithRange&, clang::ParsingDeclSpec*) + 372
29 clang 0x0000000000968c8c clang::Parser::ParseTopLevelDecl(clang::OpaquePtr<clang::DeclGroupRef>&) + 428
30 clang 0x000000000096027b clang::ParseAST(clang::Sema&, bool, bool) + 283
31 clang 0x000000000068f409 clang::FrontendAction::Execute() + 169
32 clang 0x000000000066f795 clang::CompilerInstance::ExecuteAction(clang::FrontendAction&) + 293
33 clang 0x000000000065a276 clang::ExecuteCompilerInvocation(clang::CompilerInstance*) + 1654
34 clang 0x0000000000654e48 cc1_main(char const**, char const**, char const*, void*) + 1176
35 clang 0x0000000000653c10 main + 4640
36 libc.so.6 0x00007fc78bc57ec5 __libc_start_main + 245
37 clang 0x0000000000653e7f
Stack dump:
0. Program arguments: /usr/bin/clang -cc1 -triple x86_64-pc-linux-gnu -emit-obj -disable-free -disable-llvm-verifier -main-file-name cmastrategy.cc -mrelocation-model pic -pic-level 2 -fmath-errno -masm-verbose -mconstructor-aliases -munwind-tables -fuse-init-array -target-cpu x86-64 -target-linker-version 2.24 -momit-leaf-frame-pointer -g -sys-header-deps -D HAVE_CONFIG_H -D PIC -O2 -Wall -std=c++11 -fdeprecated-macro -ferror-limit 19 -fmessage-length 112 -mstackrealign -fobjc-runtime=gcc -fcxx-exceptions -fexceptions -fdiagnostics-show-option -fcolor-diagnostics -vectorize-loops -vectorize-slp -x c++ cmastrategy-dac114.cpp
1. cmastrategy.cc:163:47: current parser token ';'
2. cmastrategy.cc:29:1: parsing namespace 'libcmaes'
3. cmastrategy.cc:33:92: LLVM IR generation of declaration 'libcmaes::CMAStrategy<libcmaes::CovarianceUpdate>::<anonymous class>::operator()'
4. cmastrategy.cc:33:92: Generating code for declaration 'libcmaes::CMAStrategy<libcmaes::CovarianceUpdate>::<anonymous class>::operator()'
Segmentation fault (core dumped)
|
non_process
|
clang and crash in clang codegen codegentypes converttype clang qualtype bugzilla link version trunk os linux reporter llvm bugzilla contributor cc douggregor rnk extended description compiling with clang and clang on linux and on mac osx does inevitably crash clang with following error libllvm so llvm sys printstacktrace io file libllvm so libpthread so clang clang codegen codegentypes converttype clang qualtype clang clang codegen codegentypes arrangellvmfunctioninfo clang canqual bool llvm arrayref clang functiontype extinfo clang codegen requiredargs clang clang clang codegen codegentypes arrangecxxmethodtype clang cxxrecorddecl const clang functionprototype const clang clang codegen codegenmodule emitglobalfunctiondefinition clang globaldecl llvm globalvalue clang clang codegen codegenmodule emitglobaldefinition clang globaldecl llvm globalvalue clang clang codegen codegenmodule emitglobal clang globaldecl clang clang clang clang clang sema instantiatefunctiondefinition clang sourcelocation clang functiondecl bool bool clang clang sema instantiateclassmembers clang sourcelocation clang cxxrecorddecl clang multileveltemplateargumentlist const clang templatespecializationkind clang clang sema instantiateclassmembers clang sourcelocation clang cxxrecorddecl clang multileveltemplateargumentlist const clang templatespecializationkind clang clang sema instantiateclasstemplatespecializationmembers clang sourcelocation clang classtemplatespecializationdecl clang templatespecializationkind clang clang sema actonexplicitinstantiation clang scope clang sourcelocation clang sourcelocation unsigned int clang sourcelocation clang cxxscopespec const clang opaqueptr clang sourcelocation clang sourcelocation llvm mutablearrayref clang sourcelocation clang attributelist clang clang parser parseclassspecifier clang tok tokenkind clang sourcelocation clang declspec clang parser parsedtemplateinfo const clang accessspecifier bool clang parser declspeccontext clang parser parsedattributeswithrange clang clang parser parsedeclarationspecifiers clang declspec clang parser parsedtemplateinfo const clang accessspecifier clang parser declspeccontext clang parser lateparsedattrlist clang clang parser parsesingledeclarationaftertemplate unsigned int clang parser parsedtemplateinfo const clang parsingdeclraiiobject clang sourcelocation clang accessspecifier clang attributelist clang clang parser parseexplicitinstantiation unsigned int clang sourcelocation clang sourcelocation clang sourcelocation clang accessspecifier clang clang parser parsedeclarationstartingwithtemplate unsigned int clang sourcelocation clang accessspecifier clang attributelist clang clang parser parsedeclaration llvm smallvector unsigned int clang sourcelocation clang parser parsedattributeswithrange clang clang parser parseexternaldeclaration clang parser parsedattributeswithrange clang parsingdeclspec clang clang parser parseinnernamespace std vector std vector std vector unsigned int clang sourcelocation clang parsedattributes clang balanceddelimitertracker clang clang parser parsenamespace unsigned int clang sourcelocation clang sourcelocation clang clang parser parsedeclaration llvm smallvector unsigned int clang sourcelocation clang parser parsedattributeswithrange clang clang parser parseexternaldeclaration clang parser parsedattributeswithrange clang parsingdeclspec clang clang parser parsetopleveldecl clang opaqueptr clang clang parseast clang sema bool bool clang clang frontendaction execute clang clang compilerinstance executeaction clang frontendaction clang clang executecompilerinvocation clang compilerinstance clang main char const char const char const void clang main libc so libc start main clang stack dump program arguments usr bin clang triple pc linux gnu emit obj disable free disable llvm verifier main file name cmastrategy cc mrelocation model pic pic level fmath errno masm verbose mconstructor aliases munwind tables fuse init array target cpu target linker version momit leaf frame pointer g sys header deps d have config h d pic wall std c fdeprecated macro ferror limit fmessage length mstackrealign fobjc runtime gcc fcxx exceptions fexceptions fdiagnostics show option fcolor diagnostics vectorize loops vectorize slp x c cmastrategy cpp cmastrategy cc current parser token cmastrategy cc parsing namespace libcmaes cmastrategy cc llvm ir generation of declaration libcmaes cmastrategy operator cmastrategy cc generating code for declaration libcmaes cmastrategy operator segmentation fault core dumped
| 0
|
349,541
| 10,470,820,365
|
IssuesEvent
|
2019-09-23 05:44:57
|
r888888888/danbooru
|
https://api.github.com/repos/r888888888/danbooru
|
closed
|
Orphaned comments
|
Bug Low Priority
|
There are 37 comments with invalid post ids:

Since these are all old low-quality comments belonging to expunged posts, they should be expunged too.
|
1.0
|
Orphaned comments - There are 37 comments with invalid post ids:

Since these are all old low-quality comments belonging to expunged posts, they should be expunged too.
|
non_process
|
orphaned comments there are comments with invalid post ids since these are all old low quality comments belonging to expunged posts they should be expunged too
| 0
|
69,077
| 17,567,202,094
|
IssuesEvent
|
2021-08-14 00:47:20
|
processing/processing-android
|
https://api.github.com/repos/processing/processing-android
|
opened
|
Query renderer from sketch's code
|
mode building
|
The renderer is needed during the sketch build process because selecting the correct template to build a watch face project [depends](https://github.com/processing/processing-android/blob/master/mode/src/processing/mode/android/AndroidBuild.java#L424) on whether the sketch uses the default renderer or the OpenGL renderer.
There was [a function](https://github.com/processing/processing-android/commit/897fb8e75b9937324830f31f5f487454690eb6a9#diff-1e0e4c73d892cb9dab69e207040f098dca246922cbdaf3186544c227513e8167L262) in an early version of the preprocessor in processing 4 that allowed to retrieve the renderer, but its now gone. The parsed render string [is private](https://github.com/processing/processing4/blob/master/java/src/processing/mode/java/preproc/PdeParseTreeListener.java#L79) in the PdeParseTreeListener and not available outside the class, so this requires either re-introducing an method to get the string, or implementing a custom parsing of the renderer in the mode itself.
@sampottinger @benfry let me know if you have any comments about this. Thank you!
|
1.0
|
Query renderer from sketch's code - The renderer is needed during the sketch build process because selecting the correct template to build a watch face project [depends](https://github.com/processing/processing-android/blob/master/mode/src/processing/mode/android/AndroidBuild.java#L424) on whether the sketch uses the default renderer or the OpenGL renderer.
There was [a function](https://github.com/processing/processing-android/commit/897fb8e75b9937324830f31f5f487454690eb6a9#diff-1e0e4c73d892cb9dab69e207040f098dca246922cbdaf3186544c227513e8167L262) in an early version of the preprocessor in processing 4 that allowed to retrieve the renderer, but its now gone. The parsed render string [is private](https://github.com/processing/processing4/blob/master/java/src/processing/mode/java/preproc/PdeParseTreeListener.java#L79) in the PdeParseTreeListener and not available outside the class, so this requires either re-introducing an method to get the string, or implementing a custom parsing of the renderer in the mode itself.
@sampottinger @benfry let me know if you have any comments about this. Thank you!
|
non_process
|
query renderer from sketch s code the renderer is needed during the sketch build process because selecting the correct template to build a watch face project on whether the sketch uses the default renderer or the opengl renderer there was in an early version of the preprocessor in processing that allowed to retrieve the renderer but its now gone the parsed render string in the pdeparsetreelistener and not available outside the class so this requires either re introducing an method to get the string or implementing a custom parsing of the renderer in the mode itself sampottinger benfry let me know if you have any comments about this thank you
| 0
|
189,382
| 22,047,028,824
|
IssuesEvent
|
2022-05-30 03:44:41
|
panasalap/linux-4.1.15
|
https://api.github.com/repos/panasalap/linux-4.1.15
|
closed
|
CVE-2017-16525 (Medium) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246 - autoclosed
|
security vulnerability
|
## CVE-2017-16525 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b></p></summary>
<p>
<p>Linux kernel stable tree mirror</p>
<p>Library home page: <a href=https://github.com/gregkh/linux.git>https://github.com/gregkh/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/aae4c2fa46027fd4c477372871df090c6b94f3f1">aae4c2fa46027fd4c477372871df090c6b94f3f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The usb_serial_console_disconnect function in drivers/usb/serial/console.c in the Linux kernel before 4.13.8 allows local users to cause a denial of service (use-after-free and system crash) or possibly have unspecified other impact via a crafted USB device, related to disconnection and failed setup.
<p>Publish Date: 2017-11-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16525>CVE-2017-16525</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16525">https://nvd.nist.gov/vuln/detail/CVE-2017-16525</a></p>
<p>Release Date: 2017-11-04</p>
<p>Fix Resolution: 4.13.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-16525 (Medium) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246 - autoclosed - ## CVE-2017-16525 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b></p></summary>
<p>
<p>Linux kernel stable tree mirror</p>
<p>Library home page: <a href=https://github.com/gregkh/linux.git>https://github.com/gregkh/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/aae4c2fa46027fd4c477372871df090c6b94f3f1">aae4c2fa46027fd4c477372871df090c6b94f3f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The usb_serial_console_disconnect function in drivers/usb/serial/console.c in the Linux kernel before 4.13.8 allows local users to cause a denial of service (use-after-free and system crash) or possibly have unspecified other impact via a crafted USB device, related to disconnection and failed setup.
<p>Publish Date: 2017-11-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16525>CVE-2017-16525</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16525">https://nvd.nist.gov/vuln/detail/CVE-2017-16525</a></p>
<p>Release Date: 2017-11-04</p>
<p>Fix Resolution: 4.13.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in autoclosed cve medium severity vulnerability vulnerable library linux kernel stable tree mirror library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the usb serial console disconnect function in drivers usb serial console c in the linux kernel before allows local users to cause a denial of service use after free and system crash or possibly have unspecified other impact via a crafted usb device related to disconnection and failed setup publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
187,374
| 15,097,310,233
|
IssuesEvent
|
2021-02-07 18:14:47
|
NFIBrokerage/slipstream
|
https://api.github.com/repos/NFIBrokerage/slipstream
|
closed
|
example of a repeater-style client
|
t:Documentation
|
some clients simply re-broadcast a message after joining a topic
I've been calling them a Repeater in the Haste implementation
it probably makes sense to solidify this with an example+tutorial
|
1.0
|
example of a repeater-style client - some clients simply re-broadcast a message after joining a topic
I've been calling them a Repeater in the Haste implementation
it probably makes sense to solidify this with an example+tutorial
|
non_process
|
example of a repeater style client some clients simply re broadcast a message after joining a topic i ve been calling them a repeater in the haste implementation it probably makes sense to solidify this with an example tutorial
| 0
|
14,845
| 18,239,434,307
|
IssuesEvent
|
2021-10-01 11:03:33
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[feature] [mesh] Mesh Calculator
|
User Manual Automatic new feature Processing Alg 3.6 Mesh
|
Original commit: https://github.com/qgis/QGIS/commit/cd9a84e11c2649f3af9b4cee08652bdfcd340134 by PeterPetrik
Similarly to raster calculator, mesh calculator can take dataset groups from current mesh layer and
combine them with various aritmentic/logical operators to new dataset group.
|
1.0
|
[feature] [mesh] Mesh Calculator - Original commit: https://github.com/qgis/QGIS/commit/cd9a84e11c2649f3af9b4cee08652bdfcd340134 by PeterPetrik
Similarly to raster calculator, mesh calculator can take dataset groups from current mesh layer and
combine them with various aritmentic/logical operators to new dataset group.
|
process
|
mesh calculator original commit by peterpetrik similarly to raster calculator mesh calculator can take dataset groups from current mesh layer and combine them with various aritmentic logical operators to new dataset group
| 1
|
210,017
| 16,079,712,672
|
IssuesEvent
|
2021-04-26 00:50:59
|
backend-br/vagas
|
https://api.github.com/repos/backend-br/vagas
|
closed
|
[Porto Alegre] Desenvolvedor (a) Back-End Júnior @ Grupo Dream Work
|
AWS Express Git Junior MySQL PJ Rest Stale Testes automatizados startup
|
## Descrição da vaga
**Selecionamos para startup de tecnologia sediada em Porto Alegre:**
**Estamos procurando uma pessoa proativa, apaixonada e com muita vontade de trabalhar! Os principais conhecimentos necessários para a função são em Node.js, Express.js, API REST e MySQL, juntamente com um perfil questionador e motivado por desafios. Também trabalhamos com tecnologias como Git, AWS, etc. Seria muito legal ter conhecimento sobre elas, ok? Um grande diferencial para a vaga é a fluência na língua inglesa.**
## Local
**Híbrido, Porto Alegre**
## Requisitos
**Obrigatórios:**
- Experiência prévia com Node.js e MySQL;
- Proatividade e capacidade de pesquisar e buscar soluções para os problemas e desafios;
- Vivência com desenvolvimento em ambientes e times ágeis;
- Inglês intermediário / avançado.
**Desejáveis:**
- Experiência com teste unitário e testes automatizados;
- Conhecimento dobre conceitos do "mundo de startups";
- Perfil empreendedor para ajudar no desenvolvimento da empresa.
## A empresa oferece:
- Salário à combinar - Enviar CV COM PRETENSÃO SALARIAL;
- Horários flexíveis;
- Ambiente informal e com constante mudança;
- Total liberdade para colaborar no caminho que a empresa irá seguir;
- Possibilidade de participação em eventos de tecnologia e/ou empreendedorismo.
## Contratação
**PJ a combinar pretensão salarial**
## Como se candidatar
**Por favor envie um email informando sua pretensão salarial para rh@grupodreamwork.com.br com seu CV anexado - enviar no assunto: Vaga Desenvolvedor (a) Back-End Júnior**
**Ou candidatar-se através do link:** https://jobs.recrutei.com.br/grupo-dream-work/vacancy/6116-dw1028-desenvolvedor-a-back-end-junior?fbclid=IwAR2x8pV3OmdJqBKwo2gWMJYpA4J6MVUyFkwCVKmwPPMjgWoj4nWQj2FTPdk
## Labels
Júnior
Híbrido
PJ
#### Nível
- Júnior
|
1.0
|
[Porto Alegre] Desenvolvedor (a) Back-End Júnior @ Grupo Dream Work - ## Descrição da vaga
**Selecionamos para startup de tecnologia sediada em Porto Alegre:**
**Estamos procurando uma pessoa proativa, apaixonada e com muita vontade de trabalhar! Os principais conhecimentos necessários para a função são em Node.js, Express.js, API REST e MySQL, juntamente com um perfil questionador e motivado por desafios. Também trabalhamos com tecnologias como Git, AWS, etc. Seria muito legal ter conhecimento sobre elas, ok? Um grande diferencial para a vaga é a fluência na língua inglesa.**
## Local
**Híbrido, Porto Alegre**
## Requisitos
**Obrigatórios:**
- Experiência prévia com Node.js e MySQL;
- Proatividade e capacidade de pesquisar e buscar soluções para os problemas e desafios;
- Vivência com desenvolvimento em ambientes e times ágeis;
- Inglês intermediário / avançado.
**Desejáveis:**
- Experiência com teste unitário e testes automatizados;
- Conhecimento dobre conceitos do "mundo de startups";
- Perfil empreendedor para ajudar no desenvolvimento da empresa.
## A empresa oferece:
- Salário à combinar - Enviar CV COM PRETENSÃO SALARIAL;
- Horários flexíveis;
- Ambiente informal e com constante mudança;
- Total liberdade para colaborar no caminho que a empresa irá seguir;
- Possibilidade de participação em eventos de tecnologia e/ou empreendedorismo.
## Contratação
**PJ a combinar pretensão salarial**
## Como se candidatar
**Por favor envie um email informando sua pretensão salarial para rh@grupodreamwork.com.br com seu CV anexado - enviar no assunto: Vaga Desenvolvedor (a) Back-End Júnior**
**Ou candidatar-se através do link:** https://jobs.recrutei.com.br/grupo-dream-work/vacancy/6116-dw1028-desenvolvedor-a-back-end-junior?fbclid=IwAR2x8pV3OmdJqBKwo2gWMJYpA4J6MVUyFkwCVKmwPPMjgWoj4nWQj2FTPdk
## Labels
Júnior
Híbrido
PJ
#### Nível
- Júnior
|
non_process
|
desenvolvedor a back end júnior grupo dream work descrição da vaga selecionamos para startup de tecnologia sediada em porto alegre estamos procurando uma pessoa proativa apaixonada e com muita vontade de trabalhar os principais conhecimentos necessários para a função são em node js express js api rest e mysql juntamente com um perfil questionador e motivado por desafios também trabalhamos com tecnologias como git aws etc seria muito legal ter conhecimento sobre elas ok um grande diferencial para a vaga é a fluência na língua inglesa local híbrido porto alegre requisitos obrigatórios experiência prévia com node js e mysql proatividade e capacidade de pesquisar e buscar soluções para os problemas e desafios vivência com desenvolvimento em ambientes e times ágeis inglês intermediário avançado desejáveis experiência com teste unitário e testes automatizados conhecimento dobre conceitos do mundo de startups perfil empreendedor para ajudar no desenvolvimento da empresa a empresa oferece salário à combinar enviar cv com pretensão salarial horários flexíveis ambiente informal e com constante mudança total liberdade para colaborar no caminho que a empresa irá seguir possibilidade de participação em eventos de tecnologia e ou empreendedorismo contratação pj a combinar pretensão salarial como se candidatar por favor envie um email informando sua pretensão salarial para rh grupodreamwork com br com seu cv anexado enviar no assunto vaga desenvolvedor a back end júnior ou candidatar se através do link labels júnior híbrido pj nível júnior
| 0
|
10,648
| 13,446,738,036
|
IssuesEvent
|
2020-09-08 13:22:31
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
Process Component add support for option "create_new_console"
|
Feature Process
|
**Description**
I saw that PHP has added in version 7.4.4 the option "create_new_console" for the function _proc-open.php_ (https://www.php.net/manual/en/function.proc-open.php), "create_new_console" can totally change the behavior of _proc-open.php_ on windows so it may be important for some users.
This is the description of the option:
> create_new_console (windows only): the new process has a new console, instead of inheriting its parent's console
From what I've seen The symfony Process component doesn't support this option.
And I think it can be implemented very easily.
In fact I think Process::start should have an additional parameter [] in order to pass (other_options) to proc-open.
|
1.0
|
Process Component add support for option "create_new_console" - **Description**
I saw that PHP has added in version 7.4.4 the option "create_new_console" for the function _proc-open.php_ (https://www.php.net/manual/en/function.proc-open.php), "create_new_console" can totally change the behavior of _proc-open.php_ on windows so it may be important for some users.
This is the description of the option:
> create_new_console (windows only): the new process has a new console, instead of inheriting its parent's console
From what I've seen The symfony Process component doesn't support this option.
And I think it can be implemented very easily.
In fact I think Process::start should have an additional parameter [] in order to pass (other_options) to proc-open.
|
process
|
process component add support for option create new console description i saw that php has added in version the option create new console for the function proc open php create new console can totally change the behavior of proc open php on windows so it may be important for some users this is the description of the option create new console windows only the new process has a new console instead of inheriting its parent s console from what i ve seen the symfony process component doesn t support this option and i think it can be implemented very easily in fact i think process start should have an additional parameter in order to pass other options to proc open
| 1
|
5,243
| 8,038,843,255
|
IssuesEvent
|
2018-07-30 16:30:06
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
opened
|
Improve deploy process of releases
|
process: release process: tests type: chore
|
### Current behavior:
- Need more visibility on if there are failing tests in 'production' env.
### Desired behavior:
<!-- A clear and concise description of what you want to happen -->
|
2.0
|
Improve deploy process of releases - ### Current behavior:
- Need more visibility on if there are failing tests in 'production' env.
### Desired behavior:
<!-- A clear and concise description of what you want to happen -->
|
process
|
improve deploy process of releases current behavior need more visibility on if there are failing tests in production env desired behavior
| 1
|
16,686
| 21,791,068,428
|
IssuesEvent
|
2022-05-14 22:53:25
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Way Too Complicated
|
suggested title in process
|
Please add as much of the following info as you can:
Title: Way Too Complicated
Type (film/tv show): TV show
Film or show in which it appears: Dinosaurs (Disney)
Is the parent film/show streaming anywhere? Disney+ I think
About when in the parent film/show does it appear? S02E04
Actual footage of the film/show can be seen (yes/no)? yes
https://www.youtube.com/watch?v=s834kh7MiSw

|
1.0
|
Add Way Too Complicated - Please add as much of the following info as you can:
Title: Way Too Complicated
Type (film/tv show): TV show
Film or show in which it appears: Dinosaurs (Disney)
Is the parent film/show streaming anywhere? Disney+ I think
About when in the parent film/show does it appear? S02E04
Actual footage of the film/show can be seen (yes/no)? yes
https://www.youtube.com/watch?v=s834kh7MiSw

|
process
|
add way too complicated please add as much of the following info as you can title way too complicated type film tv show tv show film or show in which it appears dinosaurs disney is the parent film show streaming anywhere disney i think about when in the parent film show does it appear actual footage of the film show can be seen yes no yes
| 1
|
171,667
| 6,492,970,033
|
IssuesEvent
|
2017-08-21 15:14:52
|
openshift/origin-web-console
|
https://api.github.com/repos/openshift/origin-web-console
|
closed
|
Secret in binding confirmation should be a link
|
area/usability kind/bug priority/P3
|
The secret named in the results step of the bind wizard should be a link to that secret, like in the mockup below.

|
1.0
|
Secret in binding confirmation should be a link - The secret named in the results step of the bind wizard should be a link to that secret, like in the mockup below.

|
non_process
|
secret in binding confirmation should be a link the secret named in the results step of the bind wizard should be a link to that secret like in the mockup below
| 0
|
77,134
| 9,543,752,438
|
IssuesEvent
|
2019-05-01 11:36:48
|
med-material/BakingTrayTaskVR
|
https://api.github.com/repos/med-material/BakingTrayTaskVR
|
opened
|
Improve Visuals
|
design
|
- [ ] The scene currently doesn't have nice lighting, and lack more or less shadows completely.
- [ ] The current model does not look like something eatable, so it would be nice to have a proper cookie model instead.
|
1.0
|
Improve Visuals - - [ ] The scene currently doesn't have nice lighting, and lack more or less shadows completely.
- [ ] The current model does not look like something eatable, so it would be nice to have a proper cookie model instead.
|
non_process
|
improve visuals the scene currently doesn t have nice lighting and lack more or less shadows completely the current model does not look like something eatable so it would be nice to have a proper cookie model instead
| 0
|
398,202
| 27,188,034,597
|
IssuesEvent
|
2023-02-19 13:09:19
|
aiken-lang/aiken
|
https://api.github.com/repos/aiken-lang/aiken
|
opened
|
Documentation gaps
|
need for documentation help welcomed
|
- [ ] Extend the documentation on constants in the user manual.
Cover what types can be used as constants and give examples.
- [ ] An end-to-end minting/burning policy example would be beneficial.
We only have spending validator. Ideally, we could show an example of a minting policy that is parameterized by a UTXO reference.
- [ ] A "crash course" on eUTXO, presenting the core concept of validators, datums and redeemers.
Highlight the key difference between the accounting model and the eUTXO model. And finally, present the Aiken's syntax for declaring validators.
- [ ] Add `Ordering` to [aiken-lang/prelude](https://github.com/aiken-lang/prelude)
- [ ] Document Aiken's builtins, possibly in [aiken-lang/prelude](https://github.com/aiken-lang/prelude).
We do have the [UPLC builtins](https://aiken-lang.org/uplc/builtins) page in the user manual but, the name of some builtins are different in Aiken (e.g. ByteString vs ByteArray), and not all builtins are supported in Aiken (because they have a direct syntax equivalent in the language).
- [ ] A slower / more gentle "introduction to Aiken" piece meant for people with little programming experience.
Especially FP. This could use an approach similar to _The Little Schemer_ with a 2-column layout showing a dialogue between an all-knowing-entity and a freshman asking questions; slowly diving into the language.
|
1.0
|
Documentation gaps - - [ ] Extend the documentation on constants in the user manual.
Cover what types can be used as constants and give examples.
- [ ] An end-to-end minting/burning policy example would be beneficial.
We only have spending validator. Ideally, we could show an example of a minting policy that is parameterized by a UTXO reference.
- [ ] A "crash course" on eUTXO, presenting the core concept of validators, datums and redeemers.
Highlight the key difference between the accounting model and the eUTXO model. And finally, present the Aiken's syntax for declaring validators.
- [ ] Add `Ordering` to [aiken-lang/prelude](https://github.com/aiken-lang/prelude)
- [ ] Document Aiken's builtins, possibly in [aiken-lang/prelude](https://github.com/aiken-lang/prelude).
We do have the [UPLC builtins](https://aiken-lang.org/uplc/builtins) page in the user manual but, the name of some builtins are different in Aiken (e.g. ByteString vs ByteArray), and not all builtins are supported in Aiken (because they have a direct syntax equivalent in the language).
- [ ] A slower / more gentle "introduction to Aiken" piece meant for people with little programming experience.
Especially FP. This could use an approach similar to _The Little Schemer_ with a 2-column layout showing a dialogue between an all-knowing-entity and a freshman asking questions; slowly diving into the language.
|
non_process
|
documentation gaps extend the documentation on constants in the user manual cover what types can be used as constants and give examples an end to end minting burning policy example would be beneficial we only have spending validator ideally we could show an example of a minting policy that is parameterized by a utxo reference a crash course on eutxo presenting the core concept of validators datums and redeemers highlight the key difference between the accounting model and the eutxo model and finally present the aiken s syntax for declaring validators add ordering to document aiken s builtins possibly in we do have the page in the user manual but the name of some builtins are different in aiken e g bytestring vs bytearray and not all builtins are supported in aiken because they have a direct syntax equivalent in the language a slower more gentle introduction to aiken piece meant for people with little programming experience especially fp this could use an approach similar to the little schemer with a column layout showing a dialogue between an all knowing entity and a freshman asking questions slowly diving into the language
| 0
|
607,863
| 18,792,839,744
|
IssuesEvent
|
2021-11-08 18:37:47
|
azerothcore/azerothcore-wotlk
|
https://api.github.com/repos/azerothcore/azerothcore-wotlk
|
opened
|
Server crash
|
ChromieCraft Generic Priority-Critical
|
```
Thread 3 "worldserver" received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x14c7833fe700 (LWP 20404)]
0x000000000155a2fe in Acore::Abort (file=0x16325b7 "/root/env/chromiecraft/azerothcore-wotlk/src/server/worldserver/Main.cpp", line=646, function=0x1891455 "Handler") at /root/env/chromiecraft/azerothcore-wotlk/src/common/Debugging/Errors.cpp:115
115 Crash(formattedMessage.c_str());
#0 0x000000000155a2fe in Acore::Abort (file=0x16325b7 "/root/env/chromiecraft/azerothcore-wotlk/src/server/worldserver/Main.cpp", line=646, function=0x1891455 "Handler") at /root/env/chromiecraft/azerothcore-wotlk/src/common/Debugging/Errors.cpp:115
#1 0x0000000000435861 in FreezeDetector::Handler (freezeDetectorRef=..., error=...) at /root/env/chromiecraft/azerothcore-wotlk/src/server/worldserver/Main.cpp:646
#2 0x00000000004500f9 in std::__invoke_impl<void, void (*&)(std::weak_ptr<FreezeDetector>, boost::system::error_code const&), std::weak_ptr<FreezeDetector>&, boost::system::error_code const&> (__f=<optimized out>, __args=..., __args=...) at /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/invoke.h:60
#3 std::__invoke<void (*&)(std::weak_ptr<FreezeDetector>, boost::system::error_code const&), std::weak_ptr<FreezeDetector>&, boost::system::error_code const&> (__fn=<optimized out>, __args=..., __args=...) at /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/invoke.h:95
#4 std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>::__call<void, boost::system::error_code const&, 0ul, 1ul>(std::tuple<boost::system::error_code const&>&&, std::_Index_tuple<0ul, 1ul>) (this=0x14c7833fc410, __args=...) at /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/functional:400
#5 std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>::operator()<boost::system::error_code const&, void>(boost::system::error_code const&) (this=0x14c7833fc410, __args=...) at /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/functional:482
#6 boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>::operator()() (this=0x14c7833fc410) at /usr/include/boost/asio/detail/bind_handler.hpp:65
#7 boost::asio::asio_handler_invoke<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, ...) (function=...) at /usr/include/boost/asio/handler_invoke_hook.hpp:69
#8 boost_asio_handler_invoke_helpers::invoke<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>, std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>&) (function=..., context=...) at /usr/include/boost/asio/detail/handler_invoke_helpers.hpp:37
#9 boost::asio::detail::asio_handler_invoke<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>, std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>*) (function=..., this_handler=0x14c7833fc410) at /usr/include/boost/asio/detail/bind_handler.hpp:106
#10 boost_asio_handler_invoke_helpers::invoke<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>, boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&) (function=..., context=...) at /usr/include/boost/asio/detail/handler_invoke_helpers.hpp:37
#11 boost::asio::detail::io_object_executor<boost::asio::io_context::executor_type>::dispatch<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>, std::allocator<void> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&&, std::allocator<void> const&) const (this=<optimized out>, f=..., a=...) at /usr/include/boost/asio/detail/io_object_executor.hpp:119
#12 boost::asio::detail::handler_work<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::asio::detail::io_object_executor<boost::asio::io_context::executor_type>, boost::asio::detail::io_object_executor<boost::asio::io_context::executor_type> >::complete<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>&) (this=<optimized out>, function=..., handler=...) at /usr/include/boost/asio/detail/handler_work.hpp:72
#13 boost::asio::detail::wait_handler<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::asio::detail::io_object_executor<boost::asio::io_context::executor_type> >::do_complete(void*, boost::asio::detail::scheduler_operation*, boost::system::error_code const&, unsigned long) (owner=0x14c783a7e100, base=<optimized out>) at /usr/include/boost/asio/detail/wait_handler.hpp:73
#14 0x000000000043ca27 in boost::asio::detail::scheduler_operation::complete (this=0x14c567f81a60, owner=0x14c783a7e100, ec=..., bytes_transferred=22830303576176) at /usr/include/boost/asio/detail/scheduler_operation.hpp:40
#15 boost::asio::detail::scheduler::do_run_one (this=0x14c783a7e100, lock=..., this_thread=..., ec=...) at /usr/include/boost/asio/detail/impl/scheduler.ipp:447
#16 0x000000000043c551 in boost::asio::detail::scheduler::run (this=0x14c783a7e100, ec=...) at /usr/include/boost/asio/detail/impl/scheduler.ipp:200
#17 0x0000000000455a18 in boost::asio::io_context::run (this=<optimized out>) at /usr/include/boost/asio/impl/io_context.ipp:63
#18 Acore::Asio::IoContext::run (this=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/common/Asio/IoContext.h:45
#19 0x000014c7863e0de4 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6
#20 0x000014c7864fb609 in start_thread (arg=<optimized out>) at pthread_create.c:477
#21 0x000014c7860d0293 in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
3129135.124827 [Thread debugging using libthread_db enabled]
3129135.124835 Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
3129135.279625 [New Thread 0x146106fff700 (LWP 32096)]
3129135.279707 [New Thread 0x146106dfe700 (LWP 32097)]
3129135.280356 [New Thread 0x146106bfd700 (LWP 32098)]
3129135.280513 [New Thread 0x1461063ff700 (LWP 32099)]
3129135.284903 [New Thread 0x146105bff700 (LWP 32100)]
3129135.286931 [New Thread 0x1461059fe700 (LWP 32101)]
3129135.288365 [New Thread 0x1461057fd700 (LWP 32102)]
3129135.289783 [New Thread 0x1461055fc700 (LWP 32103)]
3129135.291235 [New Thread 0x1461053fb700 (LWP 32104)]
3129135.299775 [New Thread 0x146104f7a700 (LWP 32105)]
3129135.301227 [New Thread 0x146104d79700 (LWP 32106)]
3129135.302702 [New Thread 0x146104b78700 (LWP 32107)]
3129135.304063 [New Thread 0x146104977700 (LWP 32108)]
3129135.305538 [New Thread 0x146104776700 (LWP 32109)]
3129135.307072 [New Thread 0x146104575700 (LWP 32110)]
3129135.308548 [New Thread 0x146104374700 (LWP 32111)]
3129135.310073 [New Thread 0x146104173700 (LWP 32112)]
3129135.322658 [New Thread 0x146103f72700 (LWP 32113)]
3129135.324166 [New Thread 0x146103d71700 (LWP 32114)]
3129135.325637 [New Thread 0x146103b70700 (LWP 32115)]
3129135.327169 [New Thread 0x14610396f700 (LWP 32116)]
3129135.328900 [New Thread 0x14610376e700 (LWP 32117)]
3129135.634706 [New Thread 0x1461019ff700 (LWP 32118)]
3129144.464561 [New Thread 0x1460a91ff700 (LWP 32161)]
3129144.464658 [New Thread 0x1460a8dfe700 (LWP 32162)]
3129144.464766 [New Thread 0x1460a8bfd700 (LWP 32163)]
3129144.464857 [New Thread 0x1460a89fc700 (LWP 32164)]
3129144.464931 [New Thread 0x1460a87fb700 (LWP 32165)]
3129144.464999 [New Thread 0x1460a85fa700 (LWP 32166)]
3129144.465070 [New Thread 0x1460a83f9700 (LWP 32167)]
3129144.465143 [New Thread 0x1460a81f8700 (LWP 32168)]
3129155.905077 [New Thread 0x145da63ff700 (LWP 32174)]
3129155.943396 [New Thread 0x145da5bff700 (LWP 32175)]
3129155.969088 [New Thread 0x145da59fe700 (LWP 32176)]
Thread 33 "worldserver" received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x145da63ff700 (LWP 32174)]
0x0000000001559e47 in Acore::Assert (file=0x1892428 "/root/env/chromiecraft/azerothcore-wotlk/src/server/shared/Network/SocketMgr.h", line=35, function=0x1892477 "~SocketMgr", debugInfo=Python Exception <class 'gdb.error'> There is no member named _M_dataplus.:
, message=0x1892482 "!_threads && !_acceptor && !_threadCount", format=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/common/Debugging/Errors.cpp:78
78 Crash(formattedMessage.c_str());
#0 0x0000000001559e47 in Acore::Assert (file=0x1892428 "/root/env/chromiecraft/azerothcore-wotlk/src/server/shared/Network/SocketMgr.h", line=35, function=0x1892477 "~SocketMgr", debugInfo=Python Exception <class 'gdb.error'> There is no member named _M_dataplus.:
, message=0x1892482 "!_threads && !_acceptor && !_threadCount", format=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/common/Debugging/Errors.cpp:78
#1 0x000000000114de83 in SocketMgr<WorldSocket>::~SocketMgr (this=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/server/shared/Network/SocketMgr.h:35
#2 0x0000146109a1ba27 in __run_exit_handlers (status=-1, listp=0x146109bbd718 <__exit_funcs>, run_list_atexit=run_list_atexit@entry=true, run_dtors=run_dtors@entry=true) at exit.c:108
#3 0x0000146109a1bbe0 in __GI_exit (status=<optimized out>) at exit.c:139
#4 0x000000000045e22b in ACSoapThread (host=..., port=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/server/worldserver/ACSoap/ACSoap.cpp:40
#5 0x0000146109e04de4 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6
#6 0x0000146109f1f609 in start_thread (arg=<optimized out>) at pthread_create.c:477
#7 0x0000146109af4293 in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
3129177.758451 [Thread debugging using libthread_db enabled]
3129177.758459 Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
3129177.912977 [New Thread 0x154629fff700 (LWP 32196)]
3129177.913055 [New Thread 0x154629dfe700 (LWP 32197)]
3129177.913724 [New Thread 0x154629bfd700 (LWP 32198)]
3129177.913884 [New Thread 0x1546293ff700 (LWP 32199)]
3129177.918199 [New Thread 0x154628bff700 (LWP 32200)]
3129177.920319 [New Thread 0x1546289fe700 (LWP 32201)]
3129177.921864 [New Thread 0x1546287fd700 (LWP 32202)]
3129177.923304 [New Thread 0x1546285fc700 (LWP 32203)]
3129177.924744 [New Thread 0x1546283fb700 (LWP 32204)]
3129177.933043 [New Thread 0x154627f7a700 (LWP 32205)]
3129177.934403 [New Thread 0x154627d79700 (LWP 32206)]
3129177.935877 [New Thread 0x154627b78700 (LWP 32207)]
3129177.937273 [New Thread 0x154627977700 (LWP 32208)]
3129177.938752 [New Thread 0x154627776700 (LWP 32209)]
3129177.940155 [New Thread 0x154627575700 (LWP 32210)]
3129177.941634 [New Thread 0x154627374700 (LWP 32211)]
3129177.943152 [New Thread 0x154627173700 (LWP 32212)]
3129177.955998 [New Thread 0x154626f72700 (LWP 32213)]
3129177.957359 [New Thread 0x154626d71700 (LWP 32214)]
3129177.958826 [New Thread 0x154626b70700 (LWP 32215)]
3129177.960208 [New Thread 0x15462696f700 (LWP 32216)]
3129177.961675 [New Thread 0x15462676e700 (LWP 32217)]
3129178.233959 [New Thread 0x1546249ff700 (LWP 32218)]
3129186.684418 [New Thread 0x1545ccbff700 (LWP 32220)]
3129186.684566 [New Thread 0x1545cc7fe700 (LWP 32221)]
3129186.684680 [New Thread 0x1545cc5fd700 (LWP 32222)]
3129186.684755 [New Thread 0x1545cc3fc700 (LWP 32223)]
3129186.684827 [New Thread 0x1545cc1fb700 (LWP 32224)]
3129186.684895 [New Thread 0x1545cbffa700 (LWP 32225)]
3129186.684951 [New Thread 0x1545cbdf9700 (LWP 32226)]
3129186.685008 [New Thread 0x1545cbbf8700 (LWP 32227)]
3129197.792383 [New Thread 0x1542cabff700 (LWP 32240)]
3129197.839468 [New Thread 0x1542ca3ff700 (LWP 32241)]
3129197.839562 [New Thread 0x1542ca1fe700 (LWP 32242)]
3129197.839659 [New Thread 0x1542c99ff700 (LWP 32243)]
3129197.839742 [New Thread 0x1542c91ff700 (LWP 32244)]
3129197.839838 [New Thread 0x1542c89ff700 (LWP 32245)]
3129197.839937 [New Thread 0x1542c81ff700 (LWP 32246)]
3129197.840021 [New Thread 0x1542c79ff700 (LWP 32247)]
3129197.840085 [New Thread 0x1542c71ff700 (LWP 32248)]
3129198.031107 [New Thread 0x1542c55ff700 (LWP 32249)]
3129198.031200 [New Thread 0x1542c53fe700 (LWP 32250)]
```
### AC Version
https://github.com/chromiecraft/azerothcore-wotlk/commit/27b880876dd411b2433ec227d69f6f79dc0c3ce4
##### OPERATING SYSTEM:
Ubuntu 20.04
##### MODULES:
- [mod-ah-bot](https://github.com/azerothcore/mod-ah-bot)
- [mod-cfbg](https://github.com/azerothcore/mod-cfbg)
- [mod-chat-transmitter](https://github.com/azerothcore/mod-chat-transmitter)
- [mod-chromie-xp](https://github.com/azerothcore/mod-chromie-xp)
- [mod-desertion-warnings](https://github.com/azerothcore/mod-desertion-warnings)
- [mod-duel-reset](https://github.com/azerothcore/mod-duel-reset)
- [mod-eluna-lua-engine](https://github.com/azerothcore/mod-eluna-lua-engine)
- [mod-ip-tracker](https://github.com/azerothcore/mod-ip-tracker)
- [mod-low-level-arena](https://github.com/azerothcore/mod-low-level-arena)
- [mod-multi-client-check](https://github.com/azerothcore/mod-multi-client-check)
- [mod-rdf-expansion](https://github.com/azerothcore/mod-rdf-expansion)
- [mod-pvp-titles](https://github.com/azerothcore/mod-pvp-titles)
- [mod-pvpstats-announcer](https://github.com/azerothcore/mod-pvpstats-announcer)
- [mod-queue-list-cache](https://github.com/azerothcore/mod-queue-list-cache)
- [mod-server-auto-shutdown](https://github.com/azerothcore/mod-server-auto-shutdown)
- [lua-CarbonCopy](https://github.com/55Honey/Acore_CarbonCopy)
- [lua-LevelUpReward](https://github.com/55Honey/Acore_LevelUpReward)
- [lua-send-and-bind](https://github.com/55Honey/Acore_SendAndBind)
- [lua-Zonecheck](https://github.com/55Honey/acore_Zonecheck)
##### OTHER CUSTOMIZATIONS:
None.
##### SERVER:
ChromieCraft
|
1.0
|
Server crash - ```
Thread 3 "worldserver" received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x14c7833fe700 (LWP 20404)]
0x000000000155a2fe in Acore::Abort (file=0x16325b7 "/root/env/chromiecraft/azerothcore-wotlk/src/server/worldserver/Main.cpp", line=646, function=0x1891455 "Handler") at /root/env/chromiecraft/azerothcore-wotlk/src/common/Debugging/Errors.cpp:115
115 Crash(formattedMessage.c_str());
#0 0x000000000155a2fe in Acore::Abort (file=0x16325b7 "/root/env/chromiecraft/azerothcore-wotlk/src/server/worldserver/Main.cpp", line=646, function=0x1891455 "Handler") at /root/env/chromiecraft/azerothcore-wotlk/src/common/Debugging/Errors.cpp:115
#1 0x0000000000435861 in FreezeDetector::Handler (freezeDetectorRef=..., error=...) at /root/env/chromiecraft/azerothcore-wotlk/src/server/worldserver/Main.cpp:646
#2 0x00000000004500f9 in std::__invoke_impl<void, void (*&)(std::weak_ptr<FreezeDetector>, boost::system::error_code const&), std::weak_ptr<FreezeDetector>&, boost::system::error_code const&> (__f=<optimized out>, __args=..., __args=...) at /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/invoke.h:60
#3 std::__invoke<void (*&)(std::weak_ptr<FreezeDetector>, boost::system::error_code const&), std::weak_ptr<FreezeDetector>&, boost::system::error_code const&> (__fn=<optimized out>, __args=..., __args=...) at /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/invoke.h:95
#4 std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>::__call<void, boost::system::error_code const&, 0ul, 1ul>(std::tuple<boost::system::error_code const&>&&, std::_Index_tuple<0ul, 1ul>) (this=0x14c7833fc410, __args=...) at /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/functional:400
#5 std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>::operator()<boost::system::error_code const&, void>(boost::system::error_code const&) (this=0x14c7833fc410, __args=...) at /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/functional:482
#6 boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>::operator()() (this=0x14c7833fc410) at /usr/include/boost/asio/detail/bind_handler.hpp:65
#7 boost::asio::asio_handler_invoke<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, ...) (function=...) at /usr/include/boost/asio/handler_invoke_hook.hpp:69
#8 boost_asio_handler_invoke_helpers::invoke<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>, std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>&) (function=..., context=...) at /usr/include/boost/asio/detail/handler_invoke_helpers.hpp:37
#9 boost::asio::detail::asio_handler_invoke<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>, std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>*) (function=..., this_handler=0x14c7833fc410) at /usr/include/boost/asio/detail/bind_handler.hpp:106
#10 boost_asio_handler_invoke_helpers::invoke<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>, boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&) (function=..., context=...) at /usr/include/boost/asio/detail/handler_invoke_helpers.hpp:37
#11 boost::asio::detail::io_object_executor<boost::asio::io_context::executor_type>::dispatch<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>, std::allocator<void> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&&, std::allocator<void> const&) const (this=<optimized out>, f=..., a=...) at /usr/include/boost/asio/detail/io_object_executor.hpp:119
#12 boost::asio::detail::handler_work<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::asio::detail::io_object_executor<boost::asio::io_context::executor_type>, boost::asio::detail::io_object_executor<boost::asio::io_context::executor_type> >::complete<boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code> >(boost::asio::detail::binder1<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::system::error_code>&, std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>&) (this=<optimized out>, function=..., handler=...) at /usr/include/boost/asio/detail/handler_work.hpp:72
#13 boost::asio::detail::wait_handler<std::_Bind<void (*(std::weak_ptr<FreezeDetector>, std::_Placeholder<1>))(std::weak_ptr<FreezeDetector>, boost::system::error_code const&)>, boost::asio::detail::io_object_executor<boost::asio::io_context::executor_type> >::do_complete(void*, boost::asio::detail::scheduler_operation*, boost::system::error_code const&, unsigned long) (owner=0x14c783a7e100, base=<optimized out>) at /usr/include/boost/asio/detail/wait_handler.hpp:73
#14 0x000000000043ca27 in boost::asio::detail::scheduler_operation::complete (this=0x14c567f81a60, owner=0x14c783a7e100, ec=..., bytes_transferred=22830303576176) at /usr/include/boost/asio/detail/scheduler_operation.hpp:40
#15 boost::asio::detail::scheduler::do_run_one (this=0x14c783a7e100, lock=..., this_thread=..., ec=...) at /usr/include/boost/asio/detail/impl/scheduler.ipp:447
#16 0x000000000043c551 in boost::asio::detail::scheduler::run (this=0x14c783a7e100, ec=...) at /usr/include/boost/asio/detail/impl/scheduler.ipp:200
#17 0x0000000000455a18 in boost::asio::io_context::run (this=<optimized out>) at /usr/include/boost/asio/impl/io_context.ipp:63
#18 Acore::Asio::IoContext::run (this=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/common/Asio/IoContext.h:45
#19 0x000014c7863e0de4 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6
#20 0x000014c7864fb609 in start_thread (arg=<optimized out>) at pthread_create.c:477
#21 0x000014c7860d0293 in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
3129135.124827 [Thread debugging using libthread_db enabled]
3129135.124835 Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
3129135.279625 [New Thread 0x146106fff700 (LWP 32096)]
3129135.279707 [New Thread 0x146106dfe700 (LWP 32097)]
3129135.280356 [New Thread 0x146106bfd700 (LWP 32098)]
3129135.280513 [New Thread 0x1461063ff700 (LWP 32099)]
3129135.284903 [New Thread 0x146105bff700 (LWP 32100)]
3129135.286931 [New Thread 0x1461059fe700 (LWP 32101)]
3129135.288365 [New Thread 0x1461057fd700 (LWP 32102)]
3129135.289783 [New Thread 0x1461055fc700 (LWP 32103)]
3129135.291235 [New Thread 0x1461053fb700 (LWP 32104)]
3129135.299775 [New Thread 0x146104f7a700 (LWP 32105)]
3129135.301227 [New Thread 0x146104d79700 (LWP 32106)]
3129135.302702 [New Thread 0x146104b78700 (LWP 32107)]
3129135.304063 [New Thread 0x146104977700 (LWP 32108)]
3129135.305538 [New Thread 0x146104776700 (LWP 32109)]
3129135.307072 [New Thread 0x146104575700 (LWP 32110)]
3129135.308548 [New Thread 0x146104374700 (LWP 32111)]
3129135.310073 [New Thread 0x146104173700 (LWP 32112)]
3129135.322658 [New Thread 0x146103f72700 (LWP 32113)]
3129135.324166 [New Thread 0x146103d71700 (LWP 32114)]
3129135.325637 [New Thread 0x146103b70700 (LWP 32115)]
3129135.327169 [New Thread 0x14610396f700 (LWP 32116)]
3129135.328900 [New Thread 0x14610376e700 (LWP 32117)]
3129135.634706 [New Thread 0x1461019ff700 (LWP 32118)]
3129144.464561 [New Thread 0x1460a91ff700 (LWP 32161)]
3129144.464658 [New Thread 0x1460a8dfe700 (LWP 32162)]
3129144.464766 [New Thread 0x1460a8bfd700 (LWP 32163)]
3129144.464857 [New Thread 0x1460a89fc700 (LWP 32164)]
3129144.464931 [New Thread 0x1460a87fb700 (LWP 32165)]
3129144.464999 [New Thread 0x1460a85fa700 (LWP 32166)]
3129144.465070 [New Thread 0x1460a83f9700 (LWP 32167)]
3129144.465143 [New Thread 0x1460a81f8700 (LWP 32168)]
3129155.905077 [New Thread 0x145da63ff700 (LWP 32174)]
3129155.943396 [New Thread 0x145da5bff700 (LWP 32175)]
3129155.969088 [New Thread 0x145da59fe700 (LWP 32176)]
Thread 33 "worldserver" received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x145da63ff700 (LWP 32174)]
0x0000000001559e47 in Acore::Assert (file=0x1892428 "/root/env/chromiecraft/azerothcore-wotlk/src/server/shared/Network/SocketMgr.h", line=35, function=0x1892477 "~SocketMgr", debugInfo=Python Exception <class 'gdb.error'> There is no member named _M_dataplus.:
, message=0x1892482 "!_threads && !_acceptor && !_threadCount", format=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/common/Debugging/Errors.cpp:78
78 Crash(formattedMessage.c_str());
#0 0x0000000001559e47 in Acore::Assert (file=0x1892428 "/root/env/chromiecraft/azerothcore-wotlk/src/server/shared/Network/SocketMgr.h", line=35, function=0x1892477 "~SocketMgr", debugInfo=Python Exception <class 'gdb.error'> There is no member named _M_dataplus.:
, message=0x1892482 "!_threads && !_acceptor && !_threadCount", format=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/common/Debugging/Errors.cpp:78
#1 0x000000000114de83 in SocketMgr<WorldSocket>::~SocketMgr (this=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/server/shared/Network/SocketMgr.h:35
#2 0x0000146109a1ba27 in __run_exit_handlers (status=-1, listp=0x146109bbd718 <__exit_funcs>, run_list_atexit=run_list_atexit@entry=true, run_dtors=run_dtors@entry=true) at exit.c:108
#3 0x0000146109a1bbe0 in __GI_exit (status=<optimized out>) at exit.c:139
#4 0x000000000045e22b in ACSoapThread (host=..., port=<optimized out>) at /root/env/chromiecraft/azerothcore-wotlk/src/server/worldserver/ACSoap/ACSoap.cpp:40
#5 0x0000146109e04de4 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6
#6 0x0000146109f1f609 in start_thread (arg=<optimized out>) at pthread_create.c:477
#7 0x0000146109af4293 in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
3129177.758451 [Thread debugging using libthread_db enabled]
3129177.758459 Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
3129177.912977 [New Thread 0x154629fff700 (LWP 32196)]
3129177.913055 [New Thread 0x154629dfe700 (LWP 32197)]
3129177.913724 [New Thread 0x154629bfd700 (LWP 32198)]
3129177.913884 [New Thread 0x1546293ff700 (LWP 32199)]
3129177.918199 [New Thread 0x154628bff700 (LWP 32200)]
3129177.920319 [New Thread 0x1546289fe700 (LWP 32201)]
3129177.921864 [New Thread 0x1546287fd700 (LWP 32202)]
3129177.923304 [New Thread 0x1546285fc700 (LWP 32203)]
3129177.924744 [New Thread 0x1546283fb700 (LWP 32204)]
3129177.933043 [New Thread 0x154627f7a700 (LWP 32205)]
3129177.934403 [New Thread 0x154627d79700 (LWP 32206)]
3129177.935877 [New Thread 0x154627b78700 (LWP 32207)]
3129177.937273 [New Thread 0x154627977700 (LWP 32208)]
3129177.938752 [New Thread 0x154627776700 (LWP 32209)]
3129177.940155 [New Thread 0x154627575700 (LWP 32210)]
3129177.941634 [New Thread 0x154627374700 (LWP 32211)]
3129177.943152 [New Thread 0x154627173700 (LWP 32212)]
3129177.955998 [New Thread 0x154626f72700 (LWP 32213)]
3129177.957359 [New Thread 0x154626d71700 (LWP 32214)]
3129177.958826 [New Thread 0x154626b70700 (LWP 32215)]
3129177.960208 [New Thread 0x15462696f700 (LWP 32216)]
3129177.961675 [New Thread 0x15462676e700 (LWP 32217)]
3129178.233959 [New Thread 0x1546249ff700 (LWP 32218)]
3129186.684418 [New Thread 0x1545ccbff700 (LWP 32220)]
3129186.684566 [New Thread 0x1545cc7fe700 (LWP 32221)]
3129186.684680 [New Thread 0x1545cc5fd700 (LWP 32222)]
3129186.684755 [New Thread 0x1545cc3fc700 (LWP 32223)]
3129186.684827 [New Thread 0x1545cc1fb700 (LWP 32224)]
3129186.684895 [New Thread 0x1545cbffa700 (LWP 32225)]
3129186.684951 [New Thread 0x1545cbdf9700 (LWP 32226)]
3129186.685008 [New Thread 0x1545cbbf8700 (LWP 32227)]
3129197.792383 [New Thread 0x1542cabff700 (LWP 32240)]
3129197.839468 [New Thread 0x1542ca3ff700 (LWP 32241)]
3129197.839562 [New Thread 0x1542ca1fe700 (LWP 32242)]
3129197.839659 [New Thread 0x1542c99ff700 (LWP 32243)]
3129197.839742 [New Thread 0x1542c91ff700 (LWP 32244)]
3129197.839838 [New Thread 0x1542c89ff700 (LWP 32245)]
3129197.839937 [New Thread 0x1542c81ff700 (LWP 32246)]
3129197.840021 [New Thread 0x1542c79ff700 (LWP 32247)]
3129197.840085 [New Thread 0x1542c71ff700 (LWP 32248)]
3129198.031107 [New Thread 0x1542c55ff700 (LWP 32249)]
3129198.031200 [New Thread 0x1542c53fe700 (LWP 32250)]
```
### AC Version
https://github.com/chromiecraft/azerothcore-wotlk/commit/27b880876dd411b2433ec227d69f6f79dc0c3ce4
##### OPERATING SYSTEM:
Ubuntu 20.04
##### MODULES:
- [mod-ah-bot](https://github.com/azerothcore/mod-ah-bot)
- [mod-cfbg](https://github.com/azerothcore/mod-cfbg)
- [mod-chat-transmitter](https://github.com/azerothcore/mod-chat-transmitter)
- [mod-chromie-xp](https://github.com/azerothcore/mod-chromie-xp)
- [mod-desertion-warnings](https://github.com/azerothcore/mod-desertion-warnings)
- [mod-duel-reset](https://github.com/azerothcore/mod-duel-reset)
- [mod-eluna-lua-engine](https://github.com/azerothcore/mod-eluna-lua-engine)
- [mod-ip-tracker](https://github.com/azerothcore/mod-ip-tracker)
- [mod-low-level-arena](https://github.com/azerothcore/mod-low-level-arena)
- [mod-multi-client-check](https://github.com/azerothcore/mod-multi-client-check)
- [mod-rdf-expansion](https://github.com/azerothcore/mod-rdf-expansion)
- [mod-pvp-titles](https://github.com/azerothcore/mod-pvp-titles)
- [mod-pvpstats-announcer](https://github.com/azerothcore/mod-pvpstats-announcer)
- [mod-queue-list-cache](https://github.com/azerothcore/mod-queue-list-cache)
- [mod-server-auto-shutdown](https://github.com/azerothcore/mod-server-auto-shutdown)
- [lua-CarbonCopy](https://github.com/55Honey/Acore_CarbonCopy)
- [lua-LevelUpReward](https://github.com/55Honey/Acore_LevelUpReward)
- [lua-send-and-bind](https://github.com/55Honey/Acore_SendAndBind)
- [lua-Zonecheck](https://github.com/55Honey/acore_Zonecheck)
##### OTHER CUSTOMIZATIONS:
None.
##### SERVER:
ChromieCraft
|
non_process
|
server crash thread worldserver received signal sigsegv segmentation fault in acore abort file root env chromiecraft azerothcore wotlk src server worldserver main cpp line function handler at root env chromiecraft azerothcore wotlk src common debugging errors cpp crash formattedmessage c str in acore abort file root env chromiecraft azerothcore wotlk src server worldserver main cpp line function handler at root env chromiecraft azerothcore wotlk src common debugging errors cpp in freezedetector handler freezedetectorref error at root env chromiecraft azerothcore wotlk src server worldserver main cpp in std invoke impl boost system error code const std weak ptr boost system error code const f args args at usr bin lib gcc linux gnu include c bits invoke h std invoke boost system error code const std weak ptr boost system error code const fn args args at usr bin lib gcc linux gnu include c bits invoke h std bind std placeholder std weak ptr boost system error code const call std tuple std index tuple this args at usr bin lib gcc linux gnu include c functional std bind std placeholder std weak ptr boost system error code const operator boost system error code const this args at usr bin lib gcc linux gnu include c functional boost asio detail std placeholder std weak ptr boost system error code const boost system error code operator this at usr include boost asio detail bind handler hpp boost asio asio handler invoke std placeholder std weak ptr boost system error code const boost system error code boost asio detail std placeholder std weak ptr boost system error code const boost system error code function at usr include boost asio handler invoke hook hpp boost asio handler invoke helpers invoke std placeholder std weak ptr boost system error code const boost system error code std bind std placeholder std weak ptr boost system error code const boost asio detail std placeholder std weak ptr boost system error code const boost system error code std bind std placeholder std weak ptr boost system error code const function context at usr include boost asio detail handler invoke helpers hpp boost asio detail asio handler invoke std placeholder std weak ptr boost system error code const boost system error code std bind std placeholder std weak ptr boost system error code const boost system error code boost asio detail std placeholder std weak ptr boost system error code const boost system error code boost asio detail std placeholder std weak ptr boost system error code const boost system error code function this handler at usr include boost asio detail bind handler hpp boost asio handler invoke helpers invoke std placeholder std weak ptr boost system error code const boost system error code boost asio detail std placeholder std weak ptr boost system error code const boost system error code boost asio detail std placeholder std weak ptr boost system error code const boost system error code boost asio detail std placeholder std weak ptr boost system error code const boost system error code function context at usr include boost asio detail handler invoke helpers hpp boost asio detail io object executor dispatch std placeholder std weak ptr boost system error code const boost system error code std allocator boost asio detail std placeholder std weak ptr boost system error code const boost system error code std allocator const const this f a at usr include boost asio detail io object executor hpp boost asio detail handler work std placeholder std weak ptr boost system error code const boost asio detail io object executor boost asio detail io object executor complete std placeholder std weak ptr boost system error code const boost system error code boost asio detail std placeholder std weak ptr boost system error code const boost system error code std bind std placeholder std weak ptr boost system error code const this function handler at usr include boost asio detail handler work hpp boost asio detail wait handler std placeholder std weak ptr boost system error code const boost asio detail io object executor do complete void boost asio detail scheduler operation boost system error code const unsigned long owner base at usr include boost asio detail wait handler hpp in boost asio detail scheduler operation complete this owner ec bytes transferred at usr include boost asio detail scheduler operation hpp boost asio detail scheduler do run one this lock this thread ec at usr include boost asio detail impl scheduler ipp in boost asio detail scheduler run this ec at usr include boost asio detail impl scheduler ipp in boost asio io context run this at usr include boost asio impl io context ipp acore asio iocontext run this at root env chromiecraft azerothcore wotlk src common asio iocontext h in from usr lib linux gnu libstdc so in start thread arg at pthread create c in clone at sysdeps unix sysv linux clone s using host libthread db library lib linux gnu libthread db so thread worldserver received signal sigsegv segmentation fault in acore assert file root env chromiecraft azerothcore wotlk src server shared network socketmgr h line function socketmgr debuginfo python exception there is no member named m dataplus message threads acceptor threadcount format at root env chromiecraft azerothcore wotlk src common debugging errors cpp crash formattedmessage c str in acore assert file root env chromiecraft azerothcore wotlk src server shared network socketmgr h line function socketmgr debuginfo python exception there is no member named m dataplus message threads acceptor threadcount format at root env chromiecraft azerothcore wotlk src common debugging errors cpp in socketmgr socketmgr this at root env chromiecraft azerothcore wotlk src server shared network socketmgr h in run exit handlers status listp run list atexit run list atexit entry true run dtors run dtors entry true at exit c in gi exit status at exit c in acsoapthread host port at root env chromiecraft azerothcore wotlk src server worldserver acsoap acsoap cpp in from usr lib linux gnu libstdc so in start thread arg at pthread create c in clone at sysdeps unix sysv linux clone s using host libthread db library lib linux gnu libthread db so ac version operating system ubuntu modules other customizations none server chromiecraft
| 0
|
21,787
| 30,296,854,005
|
IssuesEvent
|
2023-07-09 23:48:33
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
dbretina 2.2.10 has 1 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/dbretina
https://inspector.pypi.io/project/dbretina
```{
"dependency": "dbretina",
"version": "2.2.10",
"result": {
"issues": 1,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "kSpider2/ks_filter.py/kSpider2/ks_filter.py:15",
"code": " subprocess.run([\"awk\"], stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpm3pk021t/dbretina"
}
}```
|
1.0
|
dbretina 2.2.10 has 1 GuardDog issues - https://pypi.org/project/dbretina
https://inspector.pypi.io/project/dbretina
```{
"dependency": "dbretina",
"version": "2.2.10",
"result": {
"issues": 1,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "kSpider2/ks_filter.py/kSpider2/ks_filter.py:15",
"code": " subprocess.run([\"awk\"], stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpm3pk021t/dbretina"
}
}```
|
process
|
dbretina has guarddog issues dependency dbretina version result issues errors results silent process execution location ks filter py ks filter py code subprocess run stdin subprocess devnull n stdout subprocess devnull stderr subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp dbretina
| 1
|
7,076
| 10,224,829,959
|
IssuesEvent
|
2019-08-16 13:46:12
|
Ultimate-Hosts-Blacklist/whitelist
|
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
|
closed
|
ytmnd.com
|
whitelisting process
|
*@maxgoldberg commented on Aug 12, 2019, 2:39 AM UTC:*
**Domains or links**
ytmnd.com, *.ytmnd.com
**More Information**
How did you discover your web site or domain was listed here?
1. Installed pi-hole and was no longer able to access site.
**Have you requested removal from other sources?**
Not necessary, this appears to be the only place blocking it.
**Additional context**
N/A
*This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Badd-Boyz-Hosts#67](https://github.com/mitchellkrogza/Badd-Boyz-Hosts/issues/67).*
|
1.0
|
ytmnd.com - *@maxgoldberg commented on Aug 12, 2019, 2:39 AM UTC:*
**Domains or links**
ytmnd.com, *.ytmnd.com
**More Information**
How did you discover your web site or domain was listed here?
1. Installed pi-hole and was no longer able to access site.
**Have you requested removal from other sources?**
Not necessary, this appears to be the only place blocking it.
**Additional context**
N/A
*This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Badd-Boyz-Hosts#67](https://github.com/mitchellkrogza/Badd-Boyz-Hosts/issues/67).*
|
process
|
ytmnd com maxgoldberg commented on aug am utc domains or links ytmnd com ytmnd com more information how did you discover your web site or domain was listed here installed pi hole and was no longer able to access site have you requested removal from other sources not necessary this appears to be the only place blocking it additional context n a this issue was moved by from
| 1
|
255,521
| 19,306,203,927
|
IssuesEvent
|
2021-12-13 11:47:51
|
o3de/o3de
|
https://api.github.com/repos/o3de/o3de
|
opened
|
Update physx docs to mention the new LY_PHYSX_PROFILE_USE_CHECKED_LIBS macro for cmake configuration
|
feature/physics kind/documentation needs-triage sig/core priority/minor
|
This change lives in development branch at the moment, once it hits release "docs/user-guide/interactivity/physics/debugging/" page needs to be modified to include that users can use the macro `LY_PHYSX_PROFILE_USE_CHECKED_LIBS ` in cmake configuration to make o3de profile configuration use the checked version of PhysX libraries, which will do extra checks and log more detailed errors of what's going on inside PhysX. This is useful to figure out if something is going wrong inside PhysX.
|
1.0
|
Update physx docs to mention the new LY_PHYSX_PROFILE_USE_CHECKED_LIBS macro for cmake configuration - This change lives in development branch at the moment, once it hits release "docs/user-guide/interactivity/physics/debugging/" page needs to be modified to include that users can use the macro `LY_PHYSX_PROFILE_USE_CHECKED_LIBS ` in cmake configuration to make o3de profile configuration use the checked version of PhysX libraries, which will do extra checks and log more detailed errors of what's going on inside PhysX. This is useful to figure out if something is going wrong inside PhysX.
|
non_process
|
update physx docs to mention the new ly physx profile use checked libs macro for cmake configuration this change lives in development branch at the moment once it hits release docs user guide interactivity physics debugging page needs to be modified to include that users can use the macro ly physx profile use checked libs in cmake configuration to make profile configuration use the checked version of physx libraries which will do extra checks and log more detailed errors of what s going on inside physx this is useful to figure out if something is going wrong inside physx
| 0
|
11,301
| 14,105,771,126
|
IssuesEvent
|
2020-11-06 14:00:00
|
paul-buerkner/brms
|
https://api.github.com/repos/paul-buerkner/brms
|
closed
|
Estimate and plot survival functions
|
feature post-processing
|
I realized that in survival regression it is pretty common to compute and plot the survival functions for different conditions. Using the internal structure of `marginal_effects` this should be relatively straightforward to implement.
|
1.0
|
Estimate and plot survival functions - I realized that in survival regression it is pretty common to compute and plot the survival functions for different conditions. Using the internal structure of `marginal_effects` this should be relatively straightforward to implement.
|
process
|
estimate and plot survival functions i realized that in survival regression it is pretty common to compute and plot the survival functions for different conditions using the internal structure of marginal effects this should be relatively straightforward to implement
| 1
|
1,508
| 4,102,792,593
|
IssuesEvent
|
2016-06-04 07:25:58
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
opened
|
The href of a link processsed incorrectly in an iframe without src if it is set from the main window
|
!IMPORTANT! AREA: client SYSTEM: URL processing TYPE: bug
|
There is an iframe
```html
<iframe id="iframe"></iframe>
```
If I create a link inside the iframe in the following way:
```js
// This code executes in the context of the top window
var iframe = document.getElementById('iframe');
var body = iframe.contentDocument.body;
body.innerHTML = '<a href="somePage.html">link</a>';
```
the link has processed href without the `iframe` flag. As I result, when I click on it, the new page in the iframe has `task.js` script instead of `iframe-task.js`.
If I do it in the other way:
```js
// This code executes in the context of the top window
var iframeDocument = document.getElementById('iframe').contentWindow.document;
var link = iframeDocument.createElement('a');
link.textContent = 'link';
link.href = 'somePage.html';
iframeDocument.body.appendChild(link);
```
it works ok in chrome and ie, but still broken in firefox.
|
1.0
|
The href of a link processsed incorrectly in an iframe without src if it is set from the main window - There is an iframe
```html
<iframe id="iframe"></iframe>
```
If I create a link inside the iframe in the following way:
```js
// This code executes in the context of the top window
var iframe = document.getElementById('iframe');
var body = iframe.contentDocument.body;
body.innerHTML = '<a href="somePage.html">link</a>';
```
the link has processed href without the `iframe` flag. As I result, when I click on it, the new page in the iframe has `task.js` script instead of `iframe-task.js`.
If I do it in the other way:
```js
// This code executes in the context of the top window
var iframeDocument = document.getElementById('iframe').contentWindow.document;
var link = iframeDocument.createElement('a');
link.textContent = 'link';
link.href = 'somePage.html';
iframeDocument.body.appendChild(link);
```
it works ok in chrome and ie, but still broken in firefox.
|
process
|
the href of a link processsed incorrectly in an iframe without src if it is set from the main window there is an iframe html if i create a link inside the iframe in the following way js this code executes in the context of the top window var iframe document getelementbyid iframe var body iframe contentdocument body body innerhtml link the link has processed href without the iframe flag as i result when i click on it the new page in the iframe has task js script instead of iframe task js if i do it in the other way js this code executes in the context of the top window var iframedocument document getelementbyid iframe contentwindow document var link iframedocument createelement a link textcontent link link href somepage html iframedocument body appendchild link it works ok in chrome and ie but still broken in firefox
| 1
|
5,907
| 8,725,552,096
|
IssuesEvent
|
2018-12-10 09:44:27
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
opened
|
discussion deletion bug
|
2.0.6 Process bug
|
When deleting a discussion, it doesnt get deleted, until moving to another tab(tasks, projects, etc.) or refreshing the page, and all the changes before doing so are saved and remain after deletion.
|
1.0
|
discussion deletion bug - When deleting a discussion, it doesnt get deleted, until moving to another tab(tasks, projects, etc.) or refreshing the page, and all the changes before doing so are saved and remain after deletion.
|
process
|
discussion deletion bug when deleting a discussion it doesnt get deleted until moving to another tab tasks projects etc or refreshing the page and all the changes before doing so are saved and remain after deletion
| 1
|
764,266
| 26,792,443,832
|
IssuesEvent
|
2023-02-01 09:30:23
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.reddit.com - see bug description
|
browser-firefox priority-critical os-linux engine-gecko
|
<!-- @browser: Firefox 109.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/109.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/117501 -->
**URL**: https://www.reddit.com/
**Browser / Version**: Firefox 109.0
**Operating System**: Linux
**Tested Another Browser**: Yes Edge
**Problem type**: Something else
**Description**: Cannot use scrollbar after click on a post
**Steps to Reproduce**:
After click on a post and open, it appear two scrollbars, the reddit site have some kind of overlay so i can get to the main page just clicking outside of the post area, i cannot use the bar while seeing the post(i think it is moving the main page that is behind the post), i can still use mouse tho.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/1/f2e61738-d1e0-4db5-8852-ea2ea92457be.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.reddit.com - see bug description - <!-- @browser: Firefox 109.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/109.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/117501 -->
**URL**: https://www.reddit.com/
**Browser / Version**: Firefox 109.0
**Operating System**: Linux
**Tested Another Browser**: Yes Edge
**Problem type**: Something else
**Description**: Cannot use scrollbar after click on a post
**Steps to Reproduce**:
After click on a post and open, it appear two scrollbars, the reddit site have some kind of overlay so i can get to the main page just clicking outside of the post area, i cannot use the bar while seeing the post(i think it is moving the main page that is behind the post), i can still use mouse tho.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/1/f2e61738-d1e0-4db5-8852-ea2ea92457be.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
see bug description url browser version firefox operating system linux tested another browser yes edge problem type something else description cannot use scrollbar after click on a post steps to reproduce after click on a post and open it appear two scrollbars the reddit site have some kind of overlay so i can get to the main page just clicking outside of the post area i cannot use the bar while seeing the post i think it is moving the main page that is behind the post i can still use mouse tho view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
290,321
| 21,875,768,264
|
IssuesEvent
|
2022-05-19 09:59:14
|
appsmithorg/appsmith
|
https://api.github.com/repos/appsmithorg/appsmith
|
closed
|
[Docs] #12531 [Bug]-[1200]:Unable to select previously selected option from dropdown if value outside the options is set - unless other option is selected at-least once
|
Documentation User Education Pod
|
> TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | App Viewers Pod
**Parent Ticket** | #12531
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment |
|
1.0
|
[Docs] #12531 [Bug]-[1200]:Unable to select previously selected option from dropdown if value outside the options is set - unless other option is selected at-least once - > TODO
- [ ] Evaluate if this task is needed. If not add the "Skip Docs" label on the parent ticket
- [ ] Fill these fields
- [ ] Prepare first draft
- [ ] Add label: "Ready for Docs Team"
Field | Details
-----|-----
**POD** | App Viewers Pod
**Parent Ticket** | #12531
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment |
|
non_process
|
unable to select previously selected option from dropdown if value outside the options is set unless other option is selected at least once todo evaluate if this task is needed if not add the skip docs label on the parent ticket fill these fields prepare first draft add label ready for docs team field details pod app viewers pod parent ticket engineer release date live date first draft auto assign priority environment
| 0
|
19,266
| 25,455,925,088
|
IssuesEvent
|
2022-11-24 14:11:22
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[IDP] [PM] Last name is getting displayed in the phone number field
|
Bug Blocker P0 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
**Steps:**
1. Login to PM
2. Click on 'Admins' tab
3. Edit admin in the list
4. Update phone number and Verify
**AR:** Last name is getting displayed in the phone number field
**ER:** Updated phone number should get displayed
|
3.0
|
[IDP] [PM] Last name is getting displayed in the phone number field - **Steps:**
1. Login to PM
2. Click on 'Admins' tab
3. Edit admin in the list
4. Update phone number and Verify
**AR:** Last name is getting displayed in the phone number field
**ER:** Updated phone number should get displayed
|
process
|
last name is getting displayed in the phone number field steps login to pm click on admins tab edit admin in the list update phone number and verify ar last name is getting displayed in the phone number field er updated phone number should get displayed
| 1
|
2,054
| 4,862,864,124
|
IssuesEvent
|
2016-11-14 13:53:31
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
closed
|
Upgrade framework does not continue after `read timeout`
|
priority_critical process_wontfix type_bug
|
As seen in this error, upgrade framework stops after it cannot reach 1 sdm node. It should skip it:
```
2016-09-21 17:21:26 04800 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 0 - INFO - +++ Starting framework update +++
2016-09-21 17:21:26 04900 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 1 - INFO - Generating SSH client connections for each storage router
2016-09-21 17:21:26 36700 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 2 - INFO - 10.106.2.34 : Creating lock files
2016-09-21 17:21:27 23700 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 3 - INFO - 10.106.2.34 : Executing function get_metadata_framework
2016-09-21 17:21:27 38900 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 4 - INFO - 10.106.2.34 : Executing function get_metadata_volumedriver
2016-09-21 17:21:27 54800 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 5 - INFO - 10.106.2.34 : Executing function get_metadata_alba
2016-09-21 17:21:27 62100 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 6 - INFO - 10.106.2.34 : Executing function get_metadata_sdm
2016-09-21 17:23:27 73300 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 7 - ERROR - Error during framework update: HTTPSConnectionPool(host='10.106.2.11', port=8500): Read timed out. (read timeout=120)
```
|
1.0
|
Upgrade framework does not continue after `read timeout` - As seen in this error, upgrade framework stops after it cannot reach 1 sdm node. It should skip it:
```
2016-09-21 17:21:26 04800 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 0 - INFO - +++ Starting framework update +++
2016-09-21 17:21:26 04900 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 1 - INFO - Generating SSH client connections for each storage router
2016-09-21 17:21:26 36700 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 2 - INFO - 10.106.2.34 : Creating lock files
2016-09-21 17:21:27 23700 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 3 - INFO - 10.106.2.34 : Executing function get_metadata_framework
2016-09-21 17:21:27 38900 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 4 - INFO - 10.106.2.34 : Executing function get_metadata_volumedriver
2016-09-21 17:21:27 54800 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 5 - INFO - 10.106.2.34 : Executing function get_metadata_alba
2016-09-21 17:21:27 62100 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 6 - INFO - 10.106.2.34 : Executing function get_metadata_sdm
2016-09-21 17:23:27 73300 +0200 - stor-202.be-gen8-1 - 10809/140546189371200 - lib/update - 7 - ERROR - Error during framework update: HTTPSConnectionPool(host='10.106.2.11', port=8500): Read timed out. (read timeout=120)
```
|
process
|
upgrade framework does not continue after read timeout as seen in this error upgrade framework stops after it cannot reach sdm node it should skip it stor be lib update info starting framework update stor be lib update info generating ssh client connections for each storage router stor be lib update info creating lock files stor be lib update info executing function get metadata framework stor be lib update info executing function get metadata volumedriver stor be lib update info executing function get metadata alba stor be lib update info executing function get metadata sdm stor be lib update error error during framework update httpsconnectionpool host port read timed out read timeout
| 1
|
42,949
| 23,052,000,950
|
IssuesEvent
|
2022-07-24 19:20:28
|
maxbachmann/RapidFuzz
|
https://api.github.com/repos/maxbachmann/RapidFuzz
|
closed
|
implement string_metric.levenshtein_editops using Hirschbergs algorithm
|
performance
|
[Hirschbergs algorithm](https://en.wikipedia.org/wiki/Hirschberg%27s_algorithm) could be used to split the alignment calculation into multiple subproblems which are solved using the current algorithm. This would significantly reduce the memory consumption for long sequences. However it should be checked starting at which memory usage this should be used, so it does not pessimize performance.
|
True
|
implement string_metric.levenshtein_editops using Hirschbergs algorithm - [Hirschbergs algorithm](https://en.wikipedia.org/wiki/Hirschberg%27s_algorithm) could be used to split the alignment calculation into multiple subproblems which are solved using the current algorithm. This would significantly reduce the memory consumption for long sequences. However it should be checked starting at which memory usage this should be used, so it does not pessimize performance.
|
non_process
|
implement string metric levenshtein editops using hirschbergs algorithm could be used to split the alignment calculation into multiple subproblems which are solved using the current algorithm this would significantly reduce the memory consumption for long sequences however it should be checked starting at which memory usage this should be used so it does not pessimize performance
| 0
|
19,188
| 25,309,963,595
|
IssuesEvent
|
2022-11-17 16:42:00
|
ESMValGroup/ESMValCore
|
https://api.github.com/repos/ESMValGroup/ESMValCore
|
opened
|
Generalized mas: apply a mask built from a dataset to all other datasets
|
enhancement preprocessor
|
Hi @alistairsellar @eleanorgb @ehogan - this is a follow-up to our discussion we had at the MO during the ESMValTool workshop (my apologies for the slightly tardy time to open this, been busy with the release that is now done).
Let me try and recap in broad lines what the requirements of the functionality are, but by all means, please add to this in here so we can get the full specifications together, so I can do the implementation dance :smiley: :+1:
- a mask can be built from an existing dataset (in the datasets list in the recipe, whether it be model data or OBS/obs4mips etc data)
- generation of that mask should be based on user input in the form of kwargs of the mask preprocessor
- that mask should be applied to all datasets that need to run in a diagnostic
I have started working on a prototype in #1803 - that does the following:
- a preprocessor called [`mask_generalized`](https://github.com/ESMValGroup/ESMValCore/blob/7e318cb6a1505c0cce662fa8d3693a8307a6a733/esmvalcore/preprocessor/_mask.py#L706) ingests two cubes - one to be masked with, the other that the mask will be built from, and a dictionary holding the masking operation - first use case here is above threshold masking; then it builds the mask from `mask_cube` and applies it to the generic `cube` that denotes all other datasets; you can see this prototype tested in its [unit test](https://github.com/ESMValGroup/ESMValCore/blob/7e318cb6a1505c0cce662fa8d3693a8307a6a733/tests/unit/preprocessor/_mask/test_mask.py#L131). The actual API implementation and frontend via recipe is to follow, but I thought I'd first touch base with you to draft the specifications of the backend, namely:
- am I going down the right path?
- could you please provide me with a few actual real world use cases that I can implement?
BTW am CC-ing @bouweandela and @schlunma (will possibly CC the rest of the dev community as well, in the near future, after a few preliminary tosses of ideas among us)
Cheers :beer:
|
1.0
|
Generalized mas: apply a mask built from a dataset to all other datasets - Hi @alistairsellar @eleanorgb @ehogan - this is a follow-up to our discussion we had at the MO during the ESMValTool workshop (my apologies for the slightly tardy time to open this, been busy with the release that is now done).
Let me try and recap in broad lines what the requirements of the functionality are, but by all means, please add to this in here so we can get the full specifications together, so I can do the implementation dance :smiley: :+1:
- a mask can be built from an existing dataset (in the datasets list in the recipe, whether it be model data or OBS/obs4mips etc data)
- generation of that mask should be based on user input in the form of kwargs of the mask preprocessor
- that mask should be applied to all datasets that need to run in a diagnostic
I have started working on a prototype in #1803 - that does the following:
- a preprocessor called [`mask_generalized`](https://github.com/ESMValGroup/ESMValCore/blob/7e318cb6a1505c0cce662fa8d3693a8307a6a733/esmvalcore/preprocessor/_mask.py#L706) ingests two cubes - one to be masked with, the other that the mask will be built from, and a dictionary holding the masking operation - first use case here is above threshold masking; then it builds the mask from `mask_cube` and applies it to the generic `cube` that denotes all other datasets; you can see this prototype tested in its [unit test](https://github.com/ESMValGroup/ESMValCore/blob/7e318cb6a1505c0cce662fa8d3693a8307a6a733/tests/unit/preprocessor/_mask/test_mask.py#L131). The actual API implementation and frontend via recipe is to follow, but I thought I'd first touch base with you to draft the specifications of the backend, namely:
- am I going down the right path?
- could you please provide me with a few actual real world use cases that I can implement?
BTW am CC-ing @bouweandela and @schlunma (will possibly CC the rest of the dev community as well, in the near future, after a few preliminary tosses of ideas among us)
Cheers :beer:
|
process
|
generalized mas apply a mask built from a dataset to all other datasets hi alistairsellar eleanorgb ehogan this is a follow up to our discussion we had at the mo during the esmvaltool workshop my apologies for the slightly tardy time to open this been busy with the release that is now done let me try and recap in broad lines what the requirements of the functionality are but by all means please add to this in here so we can get the full specifications together so i can do the implementation dance smiley a mask can be built from an existing dataset in the datasets list in the recipe whether it be model data or obs etc data generation of that mask should be based on user input in the form of kwargs of the mask preprocessor that mask should be applied to all datasets that need to run in a diagnostic i have started working on a prototype in that does the following a preprocessor called ingests two cubes one to be masked with the other that the mask will be built from and a dictionary holding the masking operation first use case here is above threshold masking then it builds the mask from mask cube and applies it to the generic cube that denotes all other datasets you can see this prototype tested in its the actual api implementation and frontend via recipe is to follow but i thought i d first touch base with you to draft the specifications of the backend namely am i going down the right path could you please provide me with a few actual real world use cases that i can implement btw am cc ing bouweandela and schlunma will possibly cc the rest of the dev community as well in the near future after a few preliminary tosses of ideas among us cheers beer
| 1
|
307,183
| 26,518,571,817
|
IssuesEvent
|
2023-01-18 23:21:05
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_cond_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass)
|
module: flaky-tests skipped module: unknown
|
Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_cond_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10713308472).
Over the past 72 hours, it has flakily failed in 2 workflow(s).
**Debugging instructions (after clicking on the recent samples link):**
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Grep for `test_cond_dynamic_shapes`
Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
|
1.0
|
DISABLED test_cond_dynamic_shapes (torch._dynamo.testing.make_test_cls_with_patches.<locals>.DummyTestClass) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_cond_dynamic_shapes) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/10713308472).
Over the past 72 hours, it has flakily failed in 2 workflow(s).
**Debugging instructions (after clicking on the recent samples link):**
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Grep for `test_cond_dynamic_shapes`
Error retrieving /opt/conda/lib/python3.10/site-packages/torch/_dynamo/testing.py: Error: Statuscode 301
|
non_process
|
disabled test cond dynamic shapes torch dynamo testing make test cls with patches dummytestclass platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has flakily failed in workflow s debugging instructions after clicking on the recent samples link to find relevant log snippets click on the workflow logs linked above grep for test cond dynamic shapes error retrieving opt conda lib site packages torch dynamo testing py error statuscode
| 0
|
13,365
| 15,831,158,465
|
IssuesEvent
|
2021-04-06 13:20:41
|
ESE-Peasy/PosturePerfection
|
https://api.github.com/repos/ESE-Peasy/PosturePerfection
|
closed
|
Incorrect arrow indicated when passing the north line
|
bug post-processing
|
# Goal
The arrow is drawn in the incorrect direction if you set your ideal pose on one side of the north line, and then move into a posture which is on the other side
# Plan
1. Revise the logic to calculate angles
1. Determine why the incorrect direction is being indicated
1. Fix it 😄
|
1.0
|
Incorrect arrow indicated when passing the north line - # Goal
The arrow is drawn in the incorrect direction if you set your ideal pose on one side of the north line, and then move into a posture which is on the other side
# Plan
1. Revise the logic to calculate angles
1. Determine why the incorrect direction is being indicated
1. Fix it 😄
|
process
|
incorrect arrow indicated when passing the north line goal the arrow is drawn in the incorrect direction if you set your ideal pose on one side of the north line and then move into a posture which is on the other side plan revise the logic to calculate angles determine why the incorrect direction is being indicated fix it 😄
| 1
|
5,903
| 8,722,366,751
|
IssuesEvent
|
2018-12-09 11:47:28
|
kerubistan/kerub
|
https://api.github.com/repos/kerubistan/kerub
|
opened
|
measure and track cpu temperature information
|
component:data processing enhancement priority: normal
|
probably lm-sensors, or proc filesystem, or whatever is available on the OS
|
1.0
|
measure and track cpu temperature information - probably lm-sensors, or proc filesystem, or whatever is available on the OS
|
process
|
measure and track cpu temperature information probably lm sensors or proc filesystem or whatever is available on the os
| 1
|
70,841
| 3,344,097,617
|
IssuesEvent
|
2015-11-16 00:21:35
|
oakesville/mythling
|
https://api.github.com/repos/oakesville/mythling
|
closed
|
Back Button Problems with EPG Activity
|
priority:medium type:usability
|
In EPG Activity, Close Calendar, Search and Details Dialog on Back Button (see comment below from TQ). Probably should use same javascript interaction from FireTV EPG.
|
1.0
|
Back Button Problems with EPG Activity - In EPG Activity, Close Calendar, Search and Details Dialog on Back Button (see comment below from TQ). Probably should use same javascript interaction from FireTV EPG.
|
non_process
|
back button problems with epg activity in epg activity close calendar search and details dialog on back button see comment below from tq probably should use same javascript interaction from firetv epg
| 0
|
13,636
| 16,255,155,182
|
IssuesEvent
|
2021-05-08 03:07:28
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
Coprocessor support ENUM / SET
|
difficulty/medium sig/coprocessor type/enhancement
|
## Development Task
Currently TiKV Coprocessor does not support ENUM / SET data types, so that expressions related to these data types cannot be pushed down, which greatly affects performance in some scenarios. We have met this case several times.
This task is to add ENUM / SET data type support to the TiKV Coprocessor.
There are some previous discussions in https://github.com/tikv/tikv/issues/5738 . Some content is outdated. According to our latest studies, the idea "Always treat ENUM / SET as string" raised in https://github.com/tikv/tikv/issues/5738#issuecomment-568288021 will not work in some edge cases, so that I would like to recommend to implement this feature by introducing two brand new `EvalType`s.
@skyzh will attach a detailed RFC about how to implement this feature later.
|
1.0
|
Coprocessor support ENUM / SET - ## Development Task
Currently TiKV Coprocessor does not support ENUM / SET data types, so that expressions related to these data types cannot be pushed down, which greatly affects performance in some scenarios. We have met this case several times.
This task is to add ENUM / SET data type support to the TiKV Coprocessor.
There are some previous discussions in https://github.com/tikv/tikv/issues/5738 . Some content is outdated. According to our latest studies, the idea "Always treat ENUM / SET as string" raised in https://github.com/tikv/tikv/issues/5738#issuecomment-568288021 will not work in some edge cases, so that I would like to recommend to implement this feature by introducing two brand new `EvalType`s.
@skyzh will attach a detailed RFC about how to implement this feature later.
|
process
|
coprocessor support enum set development task currently tikv coprocessor does not support enum set data types so that expressions related to these data types cannot be pushed down which greatly affects performance in some scenarios we have met this case several times this task is to add enum set data type support to the tikv coprocessor there are some previous discussions in some content is outdated according to our latest studies the idea always treat enum set as string raised in will not work in some edge cases so that i would like to recommend to implement this feature by introducing two brand new evaltype s skyzh will attach a detailed rfc about how to implement this feature later
| 1
|
21,469
| 29,503,167,125
|
IssuesEvent
|
2023-06-03 02:03:37
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Capture test artifacts (movies and screenshots) when running AppVeyor tests
|
process: tests CI: appveyor type: chore stale
|
Because trying to update kitchensink fails (Chrome 81 browser)
- https://www.appveyor.com/docs/packaging-artifacts/
- failing build https://github.com/cypress-io/cypress/pull/7282
```
Running: examples\assertions.spec.js (3 of 19)
Assertions
Implicit Assertions
√ .should() - make an assertion about the current subject (1317ms)
√ .and() - chain multiple assertions together (472ms)
Explicit Assertions
√ expect - make an assertion about a specified subject (486ms)
√ pass your own callback function to should() (516ms)
√ finds element by class name regex (586ms)
√ can throw any error (428ms)
√ matches unknown text between two elements (374ms)
√ assert - assert shape of an object (356ms)
1) retries the should callback until assertions pass
8 passing (13s)
1 failing
1) Assertions
Explicit Assertions
retries the should callback until assertions pass:
AssertionError: Timed out retrying: Expected to find element: `#random-number`, but never found it.
at Context.eval (https://example.cypress.io/__cypress/tests?p=cypress\integration\examples\assertions.spec.js:146:18)
```
|
1.0
|
Capture test artifacts (movies and screenshots) when running AppVeyor tests - Because trying to update kitchensink fails (Chrome 81 browser)
- https://www.appveyor.com/docs/packaging-artifacts/
- failing build https://github.com/cypress-io/cypress/pull/7282
```
Running: examples\assertions.spec.js (3 of 19)
Assertions
Implicit Assertions
√ .should() - make an assertion about the current subject (1317ms)
√ .and() - chain multiple assertions together (472ms)
Explicit Assertions
√ expect - make an assertion about a specified subject (486ms)
√ pass your own callback function to should() (516ms)
√ finds element by class name regex (586ms)
√ can throw any error (428ms)
√ matches unknown text between two elements (374ms)
√ assert - assert shape of an object (356ms)
1) retries the should callback until assertions pass
8 passing (13s)
1 failing
1) Assertions
Explicit Assertions
retries the should callback until assertions pass:
AssertionError: Timed out retrying: Expected to find element: `#random-number`, but never found it.
at Context.eval (https://example.cypress.io/__cypress/tests?p=cypress\integration\examples\assertions.spec.js:146:18)
```
|
process
|
capture test artifacts movies and screenshots when running appveyor tests because trying to update kitchensink fails chrome browser failing build running examples assertions spec js of assertions implicit assertions √ should make an assertion about the current subject √ and chain multiple assertions together explicit assertions √ expect make an assertion about a specified subject √ pass your own callback function to should √ finds element by class name regex √ can throw any error √ matches unknown text between two elements √ assert assert shape of an object retries the should callback until assertions pass passing failing assertions explicit assertions retries the should callback until assertions pass assertionerror timed out retrying expected to find element random number but never found it at context eval
| 1
|
13,319
| 15,786,563,581
|
IssuesEvent
|
2021-04-01 17:55:58
|
hasura/ask-me-anything
|
https://api.github.com/repos/hasura/ask-me-anything
|
closed
|
What was your (Adron's) favorite database ... ?
|
processing-for-shortvid question
|
I have opinions about all sorts of things, and that being the case, I have multiple answers for this question!
|
1.0
|
What was your (Adron's) favorite database ... ? - I have opinions about all sorts of things, and that being the case, I have multiple answers for this question!
|
process
|
what was your adron s favorite database i have opinions about all sorts of things and that being the case i have multiple answers for this question
| 1
|
10,364
| 13,186,015,272
|
IssuesEvent
|
2020-08-12 22:50:43
|
googleapis/nodejs-speech
|
https://api.github.com/repos/googleapis/nodejs-speech
|
closed
|
Streaming wav file doesn't know the sample rate in advance
|
api: speech type: process
|
Going through the example for streaming a wav file, though speech to text it requires the programer to set the sampleRateHertz in advance. But different wav files can have different sample rates. A program would have to read in the wav file before piping to recognizeStream in order to figure out the sampleRateHertz of the wav file - it seems to contradict the idea of using a stream.
The example code should deal with this - show us how to set the sampleRateHertz in the config based on what's in the wav file while using the stream api.
Or - I wish the product would figure out the sampleRateHertz based on the incoming wav file - like it does for client.recognize and client.longRunningRecognize
|
1.0
|
Streaming wav file doesn't know the sample rate in advance - Going through the example for streaming a wav file, though speech to text it requires the programer to set the sampleRateHertz in advance. But different wav files can have different sample rates. A program would have to read in the wav file before piping to recognizeStream in order to figure out the sampleRateHertz of the wav file - it seems to contradict the idea of using a stream.
The example code should deal with this - show us how to set the sampleRateHertz in the config based on what's in the wav file while using the stream api.
Or - I wish the product would figure out the sampleRateHertz based on the incoming wav file - like it does for client.recognize and client.longRunningRecognize
|
process
|
streaming wav file doesn t know the sample rate in advance going through the example for streaming a wav file though speech to text it requires the programer to set the sampleratehertz in advance but different wav files can have different sample rates a program would have to read in the wav file before piping to recognizestream in order to figure out the sampleratehertz of the wav file it seems to contradict the idea of using a stream the example code should deal with this show us how to set the sampleratehertz in the config based on what s in the wav file while using the stream api or i wish the product would figure out the sampleratehertz based on the incoming wav file like it does for client recognize and client longrunningrecognize
| 1
|
585,359
| 17,485,855,359
|
IssuesEvent
|
2021-08-09 10:57:36
|
slynch8/10x
|
https://api.github.com/repos/slynch8/10x
|
opened
|
Disable adding folders if folder sync is true
|
feature Priority 3
|
Can be confusing if you add files or folders to a sync'd folder, as the ones that have been added will disappear when re-opening. Need to disallow adding stuff to a sync'd folder, and maybe disable setting sync to true if there are manually added files/folders.
|
1.0
|
Disable adding folders if folder sync is true - Can be confusing if you add files or folders to a sync'd folder, as the ones that have been added will disappear when re-opening. Need to disallow adding stuff to a sync'd folder, and maybe disable setting sync to true if there are manually added files/folders.
|
non_process
|
disable adding folders if folder sync is true can be confusing if you add files or folders to a sync d folder as the ones that have been added will disappear when re opening need to disallow adding stuff to a sync d folder and maybe disable setting sync to true if there are manually added files folders
| 0
|
19,464
| 25,758,413,222
|
IssuesEvent
|
2022-12-08 18:15:51
|
mdsreq-fga-unb/2022.2-Receitalista
|
https://api.github.com/repos/mdsreq-fga-unb/2022.2-Receitalista
|
closed
|
Processo de Desenvolvimento de Software
|
Processo
|
- [x] em todas as atividades não há entrega/produto informado, e sim, o que será feito. Corrigir.
- [x] nenhuma das atividades está inserida no ciclo de vida que será utilizado.
- [x] em que atividade, por exemplo, será feito o backlog do produto?
- [x] em que atividade, por exemplo, será feito o backlog da sprint?
- [x] pair programming não é uma atividade.
|
1.0
|
Processo de Desenvolvimento de Software - - [x] em todas as atividades não há entrega/produto informado, e sim, o que será feito. Corrigir.
- [x] nenhuma das atividades está inserida no ciclo de vida que será utilizado.
- [x] em que atividade, por exemplo, será feito o backlog do produto?
- [x] em que atividade, por exemplo, será feito o backlog da sprint?
- [x] pair programming não é uma atividade.
|
process
|
processo de desenvolvimento de software em todas as atividades não há entrega produto informado e sim o que será feito corrigir nenhuma das atividades está inserida no ciclo de vida que será utilizado em que atividade por exemplo será feito o backlog do produto em que atividade por exemplo será feito o backlog da sprint pair programming não é uma atividade
| 1
|
14,253
| 17,189,038,873
|
IssuesEvent
|
2021-07-16 08:18:15
|
googleapis/python-bigquery
|
https://api.github.com/repos/googleapis/python-bigquery
|
closed
|
don't pass `read_session` to BQ Storage API `rows` from `to_dataframe` and `to_arrow`
|
api: bigquery type: process
|
https://github.com/googleapis/python-bigquery-storage/pull/228 makes `read_session` optional and quietly deprecates it.
If a new enough `google-cloud-bigquery-storage` client is installed, I'd like to avoid passing the `read_session` to `rows`, so that we can more loudly deprecate it (https://github.com/googleapis/python-bigquery-storage/issues/229)
|
1.0
|
don't pass `read_session` to BQ Storage API `rows` from `to_dataframe` and `to_arrow` - https://github.com/googleapis/python-bigquery-storage/pull/228 makes `read_session` optional and quietly deprecates it.
If a new enough `google-cloud-bigquery-storage` client is installed, I'd like to avoid passing the `read_session` to `rows`, so that we can more loudly deprecate it (https://github.com/googleapis/python-bigquery-storage/issues/229)
|
process
|
don t pass read session to bq storage api rows from to dataframe and to arrow makes read session optional and quietly deprecates it if a new enough google cloud bigquery storage client is installed i d like to avoid passing the read session to rows so that we can more loudly deprecate it
| 1
|
5,849
| 8,674,507,911
|
IssuesEvent
|
2018-11-30 07:54:01
|
facebook/osquery
|
https://api.github.com/repos/facebook/osquery
|
closed
|
osquery can't take over audit
|
process auditing
|
# Bug report
### What operating system and version are you using?
```
# cat /etc/*-release
CentOS Linux release 7.3.1611 (Core)
NAME="CentOS Linux"
VERSION="7 (Core)"
ID="centos"
ID_LIKE="rhel fedora"
VERSION_ID="7"
PRETTY_NAME="CentOS Linux 7 (Core)"
ANSI_COLOR="0;31"
CPE_NAME="cpe:/o:centos:centos:7"
HOME_URL="https://www.centos.org/"
BUG_REPORT_URL="https://bugs.centos.org/"
CENTOS_MANTISBT_PROJECT="CentOS-7"
CENTOS_MANTISBT_PROJECT_VERSION="7"
REDHAT_SUPPORT_PRODUCT="centos"
REDHAT_SUPPORT_PRODUCT_VERSION="7"
CentOS Linux release 7.3.1611 (Core)
CentOS Linux release 7.3.1611 (Core)
```
```
# osqueryi --line "SELECT version, build, platform FROM os_version;"
version = CentOS Linux release 7.3.1611 (Core)
build =
platform = rhel
```
### What version of osquery are you using?
```
# osqueryi --line "SELECT version from osquery_info;"
version = 3.3.0-1-g478a7b39
```
### What steps did you take to reproduce the issue?
When i use `systemctl start osqueryd` to start osquery, it can't take over audit.

But when start directly by `/usr/bin/osqueryd --flagfile /etc/osquery/osquery.flags --config_path /etc/osquery/osquery.conf`, audit is normal.

### Others
We now have 50 machines, this happens for several times. In Sometimes it can't reappearance, but most of the time, it'll keep appear.
Thanks
|
1.0
|
osquery can't take over audit - # Bug report
### What operating system and version are you using?
```
# cat /etc/*-release
CentOS Linux release 7.3.1611 (Core)
NAME="CentOS Linux"
VERSION="7 (Core)"
ID="centos"
ID_LIKE="rhel fedora"
VERSION_ID="7"
PRETTY_NAME="CentOS Linux 7 (Core)"
ANSI_COLOR="0;31"
CPE_NAME="cpe:/o:centos:centos:7"
HOME_URL="https://www.centos.org/"
BUG_REPORT_URL="https://bugs.centos.org/"
CENTOS_MANTISBT_PROJECT="CentOS-7"
CENTOS_MANTISBT_PROJECT_VERSION="7"
REDHAT_SUPPORT_PRODUCT="centos"
REDHAT_SUPPORT_PRODUCT_VERSION="7"
CentOS Linux release 7.3.1611 (Core)
CentOS Linux release 7.3.1611 (Core)
```
```
# osqueryi --line "SELECT version, build, platform FROM os_version;"
version = CentOS Linux release 7.3.1611 (Core)
build =
platform = rhel
```
### What version of osquery are you using?
```
# osqueryi --line "SELECT version from osquery_info;"
version = 3.3.0-1-g478a7b39
```
### What steps did you take to reproduce the issue?
When i use `systemctl start osqueryd` to start osquery, it can't take over audit.

But when start directly by `/usr/bin/osqueryd --flagfile /etc/osquery/osquery.flags --config_path /etc/osquery/osquery.conf`, audit is normal.

### Others
We now have 50 machines, this happens for several times. In Sometimes it can't reappearance, but most of the time, it'll keep appear.
Thanks
|
process
|
osquery can t take over audit bug report what operating system and version are you using cat etc release centos linux release core name centos linux version core id centos id like rhel fedora version id pretty name centos linux core ansi color cpe name cpe o centos centos home url bug report url centos mantisbt project centos centos mantisbt project version redhat support product centos redhat support product version centos linux release core centos linux release core osqueryi line select version build platform from os version version centos linux release core build platform rhel what version of osquery are you using osqueryi line select version from osquery info version what steps did you take to reproduce the issue when i use systemctl start osqueryd to start osquery it can t take over audit but when start directly by usr bin osqueryd flagfile etc osquery osquery flags config path etc osquery osquery conf audit is normal others we now have machines this happens for several times in sometimes it can t reappearance but most of the time it ll keep appear thanks
| 1
|
17,724
| 23,625,621,372
|
IssuesEvent
|
2022-08-25 03:22:29
|
MPMG-DCC-UFMG/C01
|
https://api.github.com/repos/MPMG-DCC-UFMG/C01
|
closed
|
Problema na visualização de passos de coleta dinâmica
|
[1] Bug [2] Alta Prioridade [E] Externa [0] Desenvolvimento [3] Processamento Dinâmico
|
## Comportamento Esperado
Visualizar os parâmetros inseridos nos passos de coleta dinâmica assim como especificado no arquivo de configuração da coleta.
## Comportamento Atual
Alguns passos não exibem os parâmetros inseridos previamente, como se não estivessem especificados no arquivo de configuração. No entanto, a coleta ainda é executada como o esperado, mesmo que os parâmetros não apareçam na aba de edição.
## Passos para reproduzir o erro
1. Importar o arquivo de configuração abaixo
2. Acessar a aba "Processamento dinâmico"
3. Verirficar que alguns passos não exibem os parâmetros
## Especificações da Coleta
`{
"source_name": "[Realiza\u00e7\u00e3o-F01] Coleta de Servidores P\u00fablicos - Servidores Ativos de Itapeva",
"base_url": "https://e-gov.betha.com.br/transparencia/01037-108/recursos.faces?mun=5QecS5u6hOnx7T0IfpzfTXDJDYqFSDUS",
"obey_robots": false,
"data_path": "home/isabel/itapeva/servidores_empregados_ativos",
"request_type": "GET",
"form_request_type": "POST",
"antiblock_download_delay": 2,
"antiblock_autothrottle_enabled": false,
"antiblock_autothrottle_start_delay": 2,
"antiblock_autothrottle_max_delay": 10,
"antiblock_ip_rotation_enabled": false,
"antiblock_ip_rotation_type": "tor",
"antiblock_max_reqs_per_ip": 10,
"antiblock_max_reuse_rounds": 10,
"antiblock_proxy_list": "",
"antiblock_user_agent_rotation_enabled": false,
"antiblock_reqs_per_user_agent": 100,
"antiblock_user_agents_list": "",
"antiblock_insert_cookies_enabled": false,
"antiblock_cookies_list": "",
"captcha": "none",
"has_webdriver": false,
"webdriver_path": "",
"img_xpath": "",
"sound_xpath": "",
"dynamic_processing": true,
"skip_iter_errors": false,
"explore_links": false,
"link_extractor_max_depth": null,
"link_extractor_allow_url": "",
"link_extractor_allow_domains": "",
"link_extractor_tags": "",
"link_extractor_attrs": "",
"link_extractor_check_type": false,
"link_extractor_process_value": "",
"download_files": false,
"download_files_allow_url": "",
"download_files_allow_extensions": "",
"download_files_allow_domains": "",
"download_files_tags": "",
"download_files_attrs": "",
"download_files_process_value": "",
"download_files_check_large_content": true,
"download_imgs": false,
"steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"espere\",\"depth\":1,\"arguments\":{\"segundos\":\"4\"}},{\"step\":\"screenshot\",\"depth\":1,\"arguments\":{}},{\"step\":\"clique\",\"depth\":1,\"arguments\":{\"elemento\":\"\\\"//a[text()='Servidores/empregados ativos']\\\"\"}},{\"step\":\"espere\",\"depth\":1,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"screenshot\",\"depth\":1,\"arguments\":{}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"ano\",\"children\":[{\"step\":\"selecione\",\"depth\":2,\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[1]/div[1]/span/div/span/div[1]/div[1]/label/span/select\\\"\",\"opcao\":\"ano\"}},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"se\",\"depth\":2,\"children\":[{\"step\":\"para_cada\",\"depth\":3,\"iterator\":\"mes\",\"children\":[{\"step\":\"selecione\",\"depth\":4,\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[1]/div[1]/span/div/span/div[1]/div[2]/label/span/select\\\"\",\"opcao\":\"mes\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"imprime\",\"depth\":4,\"arguments\":{\"texto\":\"f\\\"Coletando dados de {mes} de {ano}\\\"\"}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='mainForm:consultar']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"6\"}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='btExpandirTodos']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"salva_pagina\",\"depth\":4,\"arguments\":{}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='btn_export_0']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='mainForm:exportType:3']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@title='Exportar consulta para o tipo de arquivo selecionado']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"20\"}},{\"step\":\"screenshot\",\"depth\":4,\"arguments\":{}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[4]/div/div/div/div/div/div/span[3]/span/a\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"4\"}},{\"step\":\"para_cada\",\"depth\":4,\"iterator\":\"dado_servidor\",\"children\":[{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"dado_servidor\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"4\"}},{\"step\":\"screenshot\",\"depth\":5,\"arguments\":{}},{\"step\":\"executar_em_iframe\",\"depth\":5,\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[4]/div/div/div/div/iframe\\\"\"}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[3]/div/div/div/form/div[1]/div/span[1]/div[2]/span[2]/input\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[3]/div/div/div/form/div[4]/div/div/div/div/div/table/tbody/tr[4]/td/input\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[3]/div/div/div/form/div[4]/div/div/div/div/div/div/span[1]/span/input\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"4\"}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[3]/div/div/div/form/div[4]/div/div/div/div/div/div/span[3]/span/a\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"screenshot\",\"depth\":5,\"arguments\":{}},{\"step\":\"sair_de_iframe\",\"depth\":5,\"arguments\":{}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[4]/div/div/a\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"2\"}}],\"iterable\":{\"call\":{\"step\":\"localiza_elementos\",\"arguments\":{\"xpath\":\"\\\"//tbody[*]//img\\\"\"}}}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='mainForm:novaConsulta']\\\"\"}},{\"step\":\"screenshot\",\"depth\":4,\"arguments\":{}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"4\"}}],\"iterable\":{\"call\":{\"step\":\"opcoes\",\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[1]/div[1]/span/div/span/div[1]/div[2]/label/span/select\\\"\"}}}}],\"condition\":{\"call\":{\"step\":\"objeto\",\"arguments\":{\"objeto\":\"int(ano) >= 2021\"}}}}],\"iterable\":{\"call\":{\"step\":\"opcoes\",\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[1]/div[1]/span/div/span/div[1]/div[1]/label/span/select\\\"\"}}}}]}",
"encoding_detection_method": 1,
"expected_runtime_category": "fast",
"templated_url_parameter_handlers": [],
"static_form_parameter_handlers": [],
"templated_url_response_handlers": [],
"static_form_response_handlers": [],
"crawler_id": 310,
"instance_id": "165512008958773"
}`
## Screenshots
Exemplo de passos que não apresentam os elementos especificados:

|
1.0
|
Problema na visualização de passos de coleta dinâmica - ## Comportamento Esperado
Visualizar os parâmetros inseridos nos passos de coleta dinâmica assim como especificado no arquivo de configuração da coleta.
## Comportamento Atual
Alguns passos não exibem os parâmetros inseridos previamente, como se não estivessem especificados no arquivo de configuração. No entanto, a coleta ainda é executada como o esperado, mesmo que os parâmetros não apareçam na aba de edição.
## Passos para reproduzir o erro
1. Importar o arquivo de configuração abaixo
2. Acessar a aba "Processamento dinâmico"
3. Verirficar que alguns passos não exibem os parâmetros
## Especificações da Coleta
`{
"source_name": "[Realiza\u00e7\u00e3o-F01] Coleta de Servidores P\u00fablicos - Servidores Ativos de Itapeva",
"base_url": "https://e-gov.betha.com.br/transparencia/01037-108/recursos.faces?mun=5QecS5u6hOnx7T0IfpzfTXDJDYqFSDUS",
"obey_robots": false,
"data_path": "home/isabel/itapeva/servidores_empregados_ativos",
"request_type": "GET",
"form_request_type": "POST",
"antiblock_download_delay": 2,
"antiblock_autothrottle_enabled": false,
"antiblock_autothrottle_start_delay": 2,
"antiblock_autothrottle_max_delay": 10,
"antiblock_ip_rotation_enabled": false,
"antiblock_ip_rotation_type": "tor",
"antiblock_max_reqs_per_ip": 10,
"antiblock_max_reuse_rounds": 10,
"antiblock_proxy_list": "",
"antiblock_user_agent_rotation_enabled": false,
"antiblock_reqs_per_user_agent": 100,
"antiblock_user_agents_list": "",
"antiblock_insert_cookies_enabled": false,
"antiblock_cookies_list": "",
"captcha": "none",
"has_webdriver": false,
"webdriver_path": "",
"img_xpath": "",
"sound_xpath": "",
"dynamic_processing": true,
"skip_iter_errors": false,
"explore_links": false,
"link_extractor_max_depth": null,
"link_extractor_allow_url": "",
"link_extractor_allow_domains": "",
"link_extractor_tags": "",
"link_extractor_attrs": "",
"link_extractor_check_type": false,
"link_extractor_process_value": "",
"download_files": false,
"download_files_allow_url": "",
"download_files_allow_extensions": "",
"download_files_allow_domains": "",
"download_files_tags": "",
"download_files_attrs": "",
"download_files_process_value": "",
"download_files_check_large_content": true,
"download_imgs": false,
"steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"espere\",\"depth\":1,\"arguments\":{\"segundos\":\"4\"}},{\"step\":\"screenshot\",\"depth\":1,\"arguments\":{}},{\"step\":\"clique\",\"depth\":1,\"arguments\":{\"elemento\":\"\\\"//a[text()='Servidores/empregados ativos']\\\"\"}},{\"step\":\"espere\",\"depth\":1,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"screenshot\",\"depth\":1,\"arguments\":{}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"ano\",\"children\":[{\"step\":\"selecione\",\"depth\":2,\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[1]/div[1]/span/div/span/div[1]/div[1]/label/span/select\\\"\",\"opcao\":\"ano\"}},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"se\",\"depth\":2,\"children\":[{\"step\":\"para_cada\",\"depth\":3,\"iterator\":\"mes\",\"children\":[{\"step\":\"selecione\",\"depth\":4,\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[1]/div[1]/span/div/span/div[1]/div[2]/label/span/select\\\"\",\"opcao\":\"mes\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"imprime\",\"depth\":4,\"arguments\":{\"texto\":\"f\\\"Coletando dados de {mes} de {ano}\\\"\"}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='mainForm:consultar']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"6\"}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='btExpandirTodos']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"salva_pagina\",\"depth\":4,\"arguments\":{}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='btn_export_0']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='mainForm:exportType:3']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@title='Exportar consulta para o tipo de arquivo selecionado']\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"20\"}},{\"step\":\"screenshot\",\"depth\":4,\"arguments\":{}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[4]/div/div/div/div/div/div/span[3]/span/a\\\"\"}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"4\"}},{\"step\":\"para_cada\",\"depth\":4,\"iterator\":\"dado_servidor\",\"children\":[{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"dado_servidor\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"4\"}},{\"step\":\"screenshot\",\"depth\":5,\"arguments\":{}},{\"step\":\"executar_em_iframe\",\"depth\":5,\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[4]/div/div/div/div/iframe\\\"\"}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[3]/div/div/div/form/div[1]/div/span[1]/div[2]/span[2]/input\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[3]/div/div/div/form/div[4]/div/div/div/div/div/table/tbody/tr[4]/td/input\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"1\"}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[3]/div/div/div/form/div[4]/div/div/div/div/div/div/span[1]/span/input\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"4\"}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[3]/div/div/div/form/div[4]/div/div/div/div/div/div/span[3]/span/a\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"screenshot\",\"depth\":5,\"arguments\":{}},{\"step\":\"sair_de_iframe\",\"depth\":5,\"arguments\":{}},{\"step\":\"clique\",\"depth\":5,\"arguments\":{\"elemento\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[4]/div/div/a\\\"\"}},{\"step\":\"espere\",\"depth\":5,\"arguments\":{\"segundos\":\"2\"}}],\"iterable\":{\"call\":{\"step\":\"localiza_elementos\",\"arguments\":{\"xpath\":\"\\\"//tbody[*]//img\\\"\"}}}},{\"step\":\"clique\",\"depth\":4,\"arguments\":{\"elemento\":\"\\\"//input[@id='mainForm:novaConsulta']\\\"\"}},{\"step\":\"screenshot\",\"depth\":4,\"arguments\":{}},{\"step\":\"espere\",\"depth\":4,\"arguments\":{\"segundos\":\"4\"}}],\"iterable\":{\"call\":{\"step\":\"opcoes\",\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[1]/div[1]/span/div/span/div[1]/div[2]/label/span/select\\\"\"}}}}],\"condition\":{\"call\":{\"step\":\"objeto\",\"arguments\":{\"objeto\":\"int(ano) >= 2021\"}}}}],\"iterable\":{\"call\":{\"step\":\"opcoes\",\"arguments\":{\"xpath\":\"\\\"/html/body/div[1]/div[1]/div/div/div/form/div[1]/div[1]/span/div/span/div[1]/div[1]/label/span/select\\\"\"}}}}]}",
"encoding_detection_method": 1,
"expected_runtime_category": "fast",
"templated_url_parameter_handlers": [],
"static_form_parameter_handlers": [],
"templated_url_response_handlers": [],
"static_form_response_handlers": [],
"crawler_id": 310,
"instance_id": "165512008958773"
}`
## Screenshots
Exemplo de passos que não apresentam os elementos especificados:

|
process
|
problema na visualização de passos de coleta dinâmica comportamento esperado visualizar os parâmetros inseridos nos passos de coleta dinâmica assim como especificado no arquivo de configuração da coleta comportamento atual alguns passos não exibem os parâmetros inseridos previamente como se não estivessem especificados no arquivo de configuração no entanto a coleta ainda é executada como o esperado mesmo que os parâmetros não apareçam na aba de edição passos para reproduzir o erro importar o arquivo de configuração abaixo acessar a aba processamento dinâmico verirficar que alguns passos não exibem os parâmetros especificações da coleta source name coleta de servidores p servidores ativos de itapeva base url obey robots false data path home isabel itapeva servidores empregados ativos request type get form request type post antiblock download delay antiblock autothrottle enabled false antiblock autothrottle start delay antiblock autothrottle max delay antiblock ip rotation enabled false antiblock ip rotation type tor antiblock max reqs per ip antiblock max reuse rounds antiblock proxy list antiblock user agent rotation enabled false antiblock reqs per user agent antiblock user agents list antiblock insert cookies enabled false antiblock cookies list captcha none has webdriver false webdriver path img xpath sound xpath dynamic processing true skip iter errors false explore links false link extractor max depth null link extractor allow url link extractor allow domains link extractor tags link extractor attrs link extractor check type false link extractor process value download files false download files allow url download files allow extensions download files allow domains download files tags download files attrs download files process value download files check large content true download imgs false steps step root depth children step espere depth arguments segundos step screenshot depth arguments step para cada depth iterator ano children div div div div form div div span div span div div label span select opcao ano step espere depth arguments segundos step se depth children div div div div form div div span div span div div label span select opcao mes step espere depth arguments segundos step imprime depth arguments texto f coletando dados de mes de ano step clique depth arguments elemento input step espere depth arguments segundos step clique depth arguments elemento input step espere depth arguments segundos step salva pagina depth arguments step clique depth arguments elemento input step espere depth arguments segundos step clique depth arguments elemento input step espere depth arguments segundos step clique depth arguments elemento input step espere depth arguments segundos step screenshot depth arguments step clique depth arguments elemento html body div div div div div form div div div div div div div span span a step espere depth arguments segundos step para cada depth iterator dado servidor children div div div div form div div div div div iframe step clique depth arguments elemento html body div div div div div form div div span div span input step espere depth arguments segundos step clique depth arguments elemento html body div div div div div form div div div div div div table tbody tr td input step espere depth arguments segundos step clique depth arguments elemento html body div div div div div form div div div div div div div span span input step espere depth arguments segundos step clique depth arguments elemento html body div div div div div form div div div div div div div span span a step espere depth arguments segundos step screenshot depth arguments step sair de iframe depth arguments step clique depth arguments elemento html body div div div div div form div div div a step espere depth arguments segundos iterable call step localiza elementos arguments xpath tbody img step clique depth arguments elemento input step screenshot depth arguments step espere depth arguments segundos iterable call step opcoes arguments xpath html body div div div div div form div div span div span div div label span select condition call step objeto arguments objeto int ano iterable call step opcoes arguments xpath html body div div div div div form div div span div span div div label span select encoding detection method expected runtime category fast templated url parameter handlers static form parameter handlers templated url response handlers static form response handlers crawler id instance id screenshots exemplo de passos que não apresentam os elementos especificados
| 1
|
394,579
| 27,033,762,652
|
IssuesEvent
|
2023-02-12 14:27:13
|
hasadna/TreeBase
|
https://api.github.com/repos/hasadna/TreeBase
|
closed
|
פתיחת מייל לפרויקט trees@hasadna.org.il
|
documentation Admin
|
האם יש עלות? המייל גם ככה יפנה אליי. לטובת מראה טוב יותר של עמוד הנחיתה
|
1.0
|
פתיחת מייל לפרויקט trees@hasadna.org.il - האם יש עלות? המייל גם ככה יפנה אליי. לטובת מראה טוב יותר של עמוד הנחיתה
|
non_process
|
פתיחת מייל לפרויקט trees hasadna org il האם יש עלות המייל גם ככה יפנה אליי לטובת מראה טוב יותר של עמוד הנחיתה
| 0
|
6,063
| 8,901,155,078
|
IssuesEvent
|
2019-01-17 01:01:01
|
rubberduck-vba/Rubberduck
|
https://api.github.com/repos/rubberduck-vba/Rubberduck
|
closed
|
Handle Procedure Attributes
|
difficulty-03-duck enhancement module-and-procedure-attributes parse-tree-processing
|
There are a number of opened issues that have been around since 0.x:
- [x] #52 toggle default instance attribute
- [x] #53 toggle default member attribute
- [x] #54 edit procedure description attribute
- [x] #55 toggle iterator getter attribute
2.0 is going to have to make them happen. I'm tagging with [code-parsing], but really this isn't about the parser - it's _purely_ about _parsing_: the VBE doesn't support member attributes, but honors them - this will be dirty string parsing work, without a net*; the VBE API can't help here.
*technically we could use the parser here, although I'm not sure how/whether it handles procedure attributes.
This feature will work off the current selection in the active code pane, and work when the selection is within a procedure:
- Export the code module to a temp file (don't remove it from the project just yet)
- Find all procedure attributes
- Present the user with a modal UI to manage procedure attributes; allow user to cancel, or confirm changes.
- **If changes were made**:
- Remove the code module from the project
- Insert the appropriate attributes and values where applicable
- Save the temp file
- Import the temp file back into the project
---
Default member attribute: there can only be one, implementation must validate that. Sets value `0` to attribute `VB_UserMemId`, where `[MemberName]` is the name of the member/procedure:
```
Attribute [MemberName].VB_UserMemId = 0
```
A default member should be a property getter. Given a `Text` property:
```
Dim foo As New Bar
foo.Text = "123"
```
Can turn into:
```
Dim foo As New Bar
foo = "123"
```
---
Iterator getter attribute: there can only be one, implementation must validate that. Sets value `-4` to attribute `VB_UserMemId`, on a `Property Get` member named `NewEnum`:
```
Attribute NewEnum.VB_UserMemId = -4
```
The feature could create the member if it doesn't exist, and prompt whether to remove it when it does and user is un-checking the box.
This member makes a type usable with a `For Each` loop, e.g. useful for a custom `List` class implementation.
---
Procedure description attribute: one per method, implementation must validate that. Displays a **short** description for the member when the module is viewed in the _Object Browser_:
```
Attribute [MemberName].VB_Description = "Short description of what the procedure does."
```
---
Default instance attribute is toggled in a class module's header:
```
Attribute VB_PredeclaredId = False
```
When set to `True`, VBA creates a _default instance_ of the class, which can then be accessed in a similar way to static classes in C# - instead of this:
```
Dim foo As New Bar
foo.DoSomething
```
You can now have that:
```
Bar.DoSomething
```
|
1.0
|
Handle Procedure Attributes - There are a number of opened issues that have been around since 0.x:
- [x] #52 toggle default instance attribute
- [x] #53 toggle default member attribute
- [x] #54 edit procedure description attribute
- [x] #55 toggle iterator getter attribute
2.0 is going to have to make them happen. I'm tagging with [code-parsing], but really this isn't about the parser - it's _purely_ about _parsing_: the VBE doesn't support member attributes, but honors them - this will be dirty string parsing work, without a net*; the VBE API can't help here.
*technically we could use the parser here, although I'm not sure how/whether it handles procedure attributes.
This feature will work off the current selection in the active code pane, and work when the selection is within a procedure:
- Export the code module to a temp file (don't remove it from the project just yet)
- Find all procedure attributes
- Present the user with a modal UI to manage procedure attributes; allow user to cancel, or confirm changes.
- **If changes were made**:
- Remove the code module from the project
- Insert the appropriate attributes and values where applicable
- Save the temp file
- Import the temp file back into the project
---
Default member attribute: there can only be one, implementation must validate that. Sets value `0` to attribute `VB_UserMemId`, where `[MemberName]` is the name of the member/procedure:
```
Attribute [MemberName].VB_UserMemId = 0
```
A default member should be a property getter. Given a `Text` property:
```
Dim foo As New Bar
foo.Text = "123"
```
Can turn into:
```
Dim foo As New Bar
foo = "123"
```
---
Iterator getter attribute: there can only be one, implementation must validate that. Sets value `-4` to attribute `VB_UserMemId`, on a `Property Get` member named `NewEnum`:
```
Attribute NewEnum.VB_UserMemId = -4
```
The feature could create the member if it doesn't exist, and prompt whether to remove it when it does and user is un-checking the box.
This member makes a type usable with a `For Each` loop, e.g. useful for a custom `List` class implementation.
---
Procedure description attribute: one per method, implementation must validate that. Displays a **short** description for the member when the module is viewed in the _Object Browser_:
```
Attribute [MemberName].VB_Description = "Short description of what the procedure does."
```
---
Default instance attribute is toggled in a class module's header:
```
Attribute VB_PredeclaredId = False
```
When set to `True`, VBA creates a _default instance_ of the class, which can then be accessed in a similar way to static classes in C# - instead of this:
```
Dim foo As New Bar
foo.DoSomething
```
You can now have that:
```
Bar.DoSomething
```
|
process
|
handle procedure attributes there are a number of opened issues that have been around since x toggle default instance attribute toggle default member attribute edit procedure description attribute toggle iterator getter attribute is going to have to make them happen i m tagging with but really this isn t about the parser it s purely about parsing the vbe doesn t support member attributes but honors them this will be dirty string parsing work without a net the vbe api can t help here technically we could use the parser here although i m not sure how whether it handles procedure attributes this feature will work off the current selection in the active code pane and work when the selection is within a procedure export the code module to a temp file don t remove it from the project just yet find all procedure attributes present the user with a modal ui to manage procedure attributes allow user to cancel or confirm changes if changes were made remove the code module from the project insert the appropriate attributes and values where applicable save the temp file import the temp file back into the project default member attribute there can only be one implementation must validate that sets value to attribute vb usermemid where is the name of the member procedure attribute vb usermemid a default member should be a property getter given a text property dim foo as new bar foo text can turn into dim foo as new bar foo iterator getter attribute there can only be one implementation must validate that sets value to attribute vb usermemid on a property get member named newenum attribute newenum vb usermemid the feature could create the member if it doesn t exist and prompt whether to remove it when it does and user is un checking the box this member makes a type usable with a for each loop e g useful for a custom list class implementation procedure description attribute one per method implementation must validate that displays a short description for the member when the module is viewed in the object browser attribute vb description short description of what the procedure does default instance attribute is toggled in a class module s header attribute vb predeclaredid false when set to true vba creates a default instance of the class which can then be accessed in a similar way to static classes in c instead of this dim foo as new bar foo dosomething you can now have that bar dosomething
| 1
|
8,332
| 11,493,161,101
|
IssuesEvent
|
2020-02-11 22:27:13
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
opened
|
Remove old acronyms
|
Process Heating
|
I think this is mostly an issue with PHAST - in Report - Energy Used, System Setup - Design Energy Use & System Setup - Metered Energy
Except when referring to the older software (the home page and the about page), the old acronyms should not be used
Old name New name
PSAT PA
PHAST PHA
FSAT FA
SSMT SA
|
1.0
|
Remove old acronyms - I think this is mostly an issue with PHAST - in Report - Energy Used, System Setup - Design Energy Use & System Setup - Metered Energy
Except when referring to the older software (the home page and the about page), the old acronyms should not be used
Old name New name
PSAT PA
PHAST PHA
FSAT FA
SSMT SA
|
process
|
remove old acronyms i think this is mostly an issue with phast in report energy used system setup design energy use system setup metered energy except when referring to the older software the home page and the about page the old acronyms should not be used old name new name psat pa phast pha fsat fa ssmt sa
| 1
|
8,093
| 11,271,315,781
|
IssuesEvent
|
2020-01-14 12:48:26
|
spring-projects/spring-hateoas
|
https://api.github.com/repos/spring-projects/spring-hateoas
|
closed
|
Refine factory methods and nullability checks for RepresentationModel types
|
in: core in: mediatypes process: in progress type: bug
|
Sonar issued us a curious warning which made us look into the code why this warning was issued. Sonar reported a resource's content should be checked against null because the third constructor in EntityModel has a `@Nullable` annotation for content. However, directly below that, there's an `Assert.notNull` for content. Also, the documentation says that content should not be nullable...
**Constructor**
```
/**
* Creates a new {@link EntityModel} with the given content and {@link Link}s.
*
* @param content must not be {@literal null}.
* @param links the links to add to the {@link EntityModel}.
*/
public EntityModel(@Nullable T content, Iterable<Link> links) {
Assert.notNull(content, "Content must not be null!");
Assert.isTrue(!(content instanceof Collection), "Content must not be a collection! Use Resources instead!");
this.content = content;
this.add(links);
}
```
|
1.0
|
Refine factory methods and nullability checks for RepresentationModel types - Sonar issued us a curious warning which made us look into the code why this warning was issued. Sonar reported a resource's content should be checked against null because the third constructor in EntityModel has a `@Nullable` annotation for content. However, directly below that, there's an `Assert.notNull` for content. Also, the documentation says that content should not be nullable...
**Constructor**
```
/**
* Creates a new {@link EntityModel} with the given content and {@link Link}s.
*
* @param content must not be {@literal null}.
* @param links the links to add to the {@link EntityModel}.
*/
public EntityModel(@Nullable T content, Iterable<Link> links) {
Assert.notNull(content, "Content must not be null!");
Assert.isTrue(!(content instanceof Collection), "Content must not be a collection! Use Resources instead!");
this.content = content;
this.add(links);
}
```
|
process
|
refine factory methods and nullability checks for representationmodel types sonar issued us a curious warning which made us look into the code why this warning was issued sonar reported a resource s content should be checked against null because the third constructor in entitymodel has a nullable annotation for content however directly below that there s an assert notnull for content also the documentation says that content should not be nullable constructor creates a new link entitymodel with the given content and link link s param content must not be literal null param links the links to add to the link entitymodel public entitymodel nullable t content iterable links assert notnull content content must not be null assert istrue content instanceof collection content must not be a collection use resources instead this content content this add links
| 1
|
523,332
| 15,178,376,288
|
IssuesEvent
|
2021-02-14 15:13:27
|
TrinityCore/TrinityCore
|
https://api.github.com/repos/TrinityCore/TrinityCore
|
closed
|
Core/SpawnGroup: Phase 2: SPAWN_FLAG_NO_RESPAWN
|
Comp-Core Priority-FutureFeatureRequest
|
**Description:**
For example we want to temp spawn creature only once without allowing it to respawn. Currently to achieve it we need to spawn SpawnGroup and despawn it after creature is killed. Each time for each creature.
Or we need to change respawn time to prevent respawn. But it's something evil
**Expected behaviour:**
We need new flag SPAWN_FLAG_NO_RESPAWN to spawn creature only once without trying to make it respawn after kill. That flag should be used for 95% of temp spawned creatures or objects
**Branch(es):** both
**TC rev. hash/commit:** 4c0a8ad3b3f9960fa15504cf32a84218807ac9d5
|
1.0
|
Core/SpawnGroup: Phase 2: SPAWN_FLAG_NO_RESPAWN - **Description:**
For example we want to temp spawn creature only once without allowing it to respawn. Currently to achieve it we need to spawn SpawnGroup and despawn it after creature is killed. Each time for each creature.
Or we need to change respawn time to prevent respawn. But it's something evil
**Expected behaviour:**
We need new flag SPAWN_FLAG_NO_RESPAWN to spawn creature only once without trying to make it respawn after kill. That flag should be used for 95% of temp spawned creatures or objects
**Branch(es):** both
**TC rev. hash/commit:** 4c0a8ad3b3f9960fa15504cf32a84218807ac9d5
|
non_process
|
core spawngroup phase spawn flag no respawn description for example we want to temp spawn creature only once without allowing it to respawn currently to achieve it we need to spawn spawngroup and despawn it after creature is killed each time for each creature or we need to change respawn time to prevent respawn but it s something evil expected behaviour we need new flag spawn flag no respawn to spawn creature only once without trying to make it respawn after kill that flag should be used for of temp spawned creatures or objects branch es both tc rev hash commit
| 0
|
39,921
| 6,782,240,883
|
IssuesEvent
|
2017-10-30 06:59:58
|
cyberFund/cyber-search
|
https://api.github.com/repos/cyberFund/cyber-search
|
closed
|
Bitcoin multisig transaction
|
discussion documentation improvement
|
Bitcoin transaction output could contains multiple addresses.
Define model for this
See https://blockchain.info/tx/56214420a7c4dcc4832944298d169a75e93acf9721f00656b2ee0e4d194f9970
|
1.0
|
Bitcoin multisig transaction - Bitcoin transaction output could contains multiple addresses.
Define model for this
See https://blockchain.info/tx/56214420a7c4dcc4832944298d169a75e93acf9721f00656b2ee0e4d194f9970
|
non_process
|
bitcoin multisig transaction bitcoin transaction output could contains multiple addresses define model for this see
| 0
|
19,719
| 26,073,816,742
|
IssuesEvent
|
2022-12-24 07:03:51
|
pyanodon/pybugreports
|
https://api.github.com/repos/pyanodon/pybugreports
|
closed
|
pypostprocessing now incompatible with True Nukes
|
postprocess-fail compatibility
|
### Mod source
PyAE Beta
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [ ] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [X] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [X] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [ ] Pypostprocessing failure
- [ ] Other
### What is the problem?
No suitable lab found to research tech: atomic-bomb
stack traceback:
[C]: in function 'error'
__pypostprocessing__/prototypes/functions/data_parser.lua:324: in function 'parse_tech'
__pypostprocessing__/prototypes/functions/data_parser.lua:133: in function 'run'
__pypostprocessing__/prototypes/functions/auto_tech.lua:34: in function 'run'
__pypostprocessing__/data-final-fixes.lua:144: in main chunk
Mods to be disabled:
• pypostprocessing (0.1.0)
### Steps to reproduce
1. Install https://mods.factorio.com/mod/True-Nukes
2. Profit
### Additional context
It used to work but broke after a recent update to True Nukes
### Log file
_No response_
|
1.0
|
pypostprocessing now incompatible with True Nukes - ### Mod source
PyAE Beta
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [ ] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [X] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [X] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [ ] Pypostprocessing failure
- [ ] Other
### What is the problem?
No suitable lab found to research tech: atomic-bomb
stack traceback:
[C]: in function 'error'
__pypostprocessing__/prototypes/functions/data_parser.lua:324: in function 'parse_tech'
__pypostprocessing__/prototypes/functions/data_parser.lua:133: in function 'run'
__pypostprocessing__/prototypes/functions/auto_tech.lua:34: in function 'run'
__pypostprocessing__/data-final-fixes.lua:144: in main chunk
Mods to be disabled:
• pypostprocessing (0.1.0)
### Steps to reproduce
1. Install https://mods.factorio.com/mod/True-Nukes
2. Profit
### Additional context
It used to work but broke after a recent update to True Nukes
### Log file
_No response_
|
process
|
pypostprocessing now incompatible with true nukes mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem no suitable lab found to research tech atomic bomb stack traceback in function error pypostprocessing prototypes functions data parser lua in function parse tech pypostprocessing prototypes functions data parser lua in function run pypostprocessing prototypes functions auto tech lua in function run pypostprocessing data final fixes lua in main chunk mods to be disabled • pypostprocessing steps to reproduce install profit additional context it used to work but broke after a recent update to true nukes log file no response
| 1
|
15,271
| 19,250,569,633
|
IssuesEvent
|
2021-12-09 04:22:54
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
closed
|
Filebeat decode_cef parser error with \n\r chars in message
|
bug Filebeat :Processors Team:Security-External Integrations
|
The filebeat decode_cef function doesn't skip the new lines chars if are present into message field.
es.
`event.original: "CEF:0|Trend Micro|Deep Security Manager|12.0.393|771|Contact by Unrecognized Client|6|src=10.41.128.199 suser=System msg=A connection to Deep Security Manager was initiated by a client not identifiable as a managed computer: 10.41.130.142:63944. Either the client is not a computer managed by Deep Security, or a secure communication channel could not be established.\n\nAgent GUID or virtual ID did not map to any Hosts. The following endpoint information was provided.\r\n\r\nHTTP Headers:\r\nx-dsa-time: 1584092734\r\ncontent-length: 202\r\nx-dsa-capabilities: 1092777293\r\nx-dsm-minimum-dsm-version: 9.5.0\r\ncontent-encoding: deflate\r\nx-dsm-notifications: GetInterfaces\r\ncontent-type: text/plain; charset\="UTF-8"\r\nplatform: Windows;10;0;14393;;x86_64;400;3\r\nx-dsa-listenport: 4118\r\nuser-agent: Trend Micro Deep Security Agent (10.0.0.2888)\r\n\r\nHost Info:\r\nhostname: test02.local\r\nbios-uuid: 44164530-FAF8-734C-81DF-1D5AC071065B\r\ndomain: LOCAL\r\nhostname-nb: TEST02\r\nhostname-id: {b70482c0-b44c-4221-96df-4dc55f04177e}\r\nhostname-canonical: local/LOCAL-Dienste/TGW/TEST02\r\n TrendMicroDsTenant=Primary TrendMicroDsTenantId=0" ,
error.message: "malformed value for msg at pos 369"
`
For confirmed bugs, please report:
- Version: 7.5.X
- Operating System: Windows Server 2019
- Discuss Forum URL: https://discuss.elastic.co/t/filebeat-7-4-decode-cef-getting-error-malformed-value/202469
- Steps to Reproduce:
when a device send a message with \n or \r chars the decode function rise-up 'malformed value for xx at pos xxx'
|
1.0
|
Filebeat decode_cef parser error with \n\r chars in message - The filebeat decode_cef function doesn't skip the new lines chars if are present into message field.
es.
`event.original: "CEF:0|Trend Micro|Deep Security Manager|12.0.393|771|Contact by Unrecognized Client|6|src=10.41.128.199 suser=System msg=A connection to Deep Security Manager was initiated by a client not identifiable as a managed computer: 10.41.130.142:63944. Either the client is not a computer managed by Deep Security, or a secure communication channel could not be established.\n\nAgent GUID or virtual ID did not map to any Hosts. The following endpoint information was provided.\r\n\r\nHTTP Headers:\r\nx-dsa-time: 1584092734\r\ncontent-length: 202\r\nx-dsa-capabilities: 1092777293\r\nx-dsm-minimum-dsm-version: 9.5.0\r\ncontent-encoding: deflate\r\nx-dsm-notifications: GetInterfaces\r\ncontent-type: text/plain; charset\="UTF-8"\r\nplatform: Windows;10;0;14393;;x86_64;400;3\r\nx-dsa-listenport: 4118\r\nuser-agent: Trend Micro Deep Security Agent (10.0.0.2888)\r\n\r\nHost Info:\r\nhostname: test02.local\r\nbios-uuid: 44164530-FAF8-734C-81DF-1D5AC071065B\r\ndomain: LOCAL\r\nhostname-nb: TEST02\r\nhostname-id: {b70482c0-b44c-4221-96df-4dc55f04177e}\r\nhostname-canonical: local/LOCAL-Dienste/TGW/TEST02\r\n TrendMicroDsTenant=Primary TrendMicroDsTenantId=0" ,
error.message: "malformed value for msg at pos 369"
`
For confirmed bugs, please report:
- Version: 7.5.X
- Operating System: Windows Server 2019
- Discuss Forum URL: https://discuss.elastic.co/t/filebeat-7-4-decode-cef-getting-error-malformed-value/202469
- Steps to Reproduce:
when a device send a message with \n or \r chars the decode function rise-up 'malformed value for xx at pos xxx'
|
process
|
filebeat decode cef parser error with n r chars in message the filebeat decode cef function doesn t skip the new lines chars if are present into message field es event original cef trend micro deep security manager contact by unrecognized client src suser system msg a connection to deep security manager was initiated by a client not identifiable as a managed computer either the client is not a computer managed by deep security or a secure communication channel could not be established n nagent guid or virtual id did not map to any hosts the following endpoint information was provided r n r nhttp headers r nx dsa time r ncontent length r nx dsa capabilities r nx dsm minimum dsm version r ncontent encoding deflate r nx dsm notifications getinterfaces r ncontent type text plain charset utf r nplatform windows r nx dsa listenport r nuser agent trend micro deep security agent r n r nhost info r nhostname local r nbios uuid r ndomain local r nhostname nb r nhostname id r nhostname canonical local local dienste tgw r n trendmicrodstenant primary trendmicrodstenantid error message malformed value for msg at pos for confirmed bugs please report version x operating system windows server discuss forum url steps to reproduce when a device send a message with n or r chars the decode function rise up malformed value for xx at pos xxx
| 1
|
245,846
| 20,799,081,366
|
IssuesEvent
|
2022-03-17 12:15:00
|
Uuvana-Studios/longvinter-windows-client
|
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
|
closed
|
Cant place new tent anywhere
|
bug Not Tested
|
receiving too close to base error no matter how far away from original base.
shared a house code now i cant place down my own house :(
|
1.0
|
Cant place new tent anywhere - receiving too close to base error no matter how far away from original base.
shared a house code now i cant place down my own house :(
|
non_process
|
cant place new tent anywhere receiving too close to base error no matter how far away from original base shared a house code now i cant place down my own house
| 0
|
283,407
| 21,316,554,320
|
IssuesEvent
|
2022-04-16 11:33:14
|
jr-mojito/pe
|
https://api.github.com/repos/jr-mojito/pe
|
opened
|
Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function
|
severity.Low type.DocumentationBug
|
Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function

<!--session: 1650108417268-dccfef75-89da-4d4a-87bc-8b7c9ca8a778-->
<!--Version: Web v3.4.2-->
|
1.0
|
Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function - Option field for [rt/RECEIVE_TYPE] found on UG but not available for Edit function

<!--session: 1650108417268-dccfef75-89da-4d4a-87bc-8b7c9ca8a778-->
<!--Version: Web v3.4.2-->
|
non_process
|
option field for found on ug but not available for edit function option field for found on ug but not available for edit function
| 0
|
4,708
| 7,547,918,735
|
IssuesEvent
|
2018-04-18 09:30:26
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Very slow child_process.spawn stdout stream performance on macOS
|
child_process
|
The `child_process.spawn` stdout stream seems to be taking ~4x longer than other "similar" methods of I/O streaming with similarly sized data streams.
Running on Node.js v4.2.1, Mac OS X 10.11, Macbook Air 1.7GHz
As a baseline, the following program, where `file` is a 472MB gzip file, and `stdout` is set to `ignore` takes 6.5 secs with the gzip process using ~100% CPU:
``` js
var dataLength = 0
var time = process.hrtime()
var gzip = spawn('gzip', ['-cd', file], {stdio: ['ignore', 'ignore', 'ignore']})
if (gzip.stdout != null) {
gzip.stdout.on('data', function(data) {
if (!dataLength) dataLength = data.length
})
}
gzip.on('close', function(code) {
var diff = process.hrtime(time)
console.log('stream took %d seconds', diff[0] + diff[1] / 1e9)
if (dataLength) console.log('buffer length was %d', dataLength)
})
```
``` console
$ node spawn-bench.js
stream took 6.497701762 seconds
```
If I set the `stdout` option from `ignore` to `pipe`, it suddenly takes 27 secs, with gzip only using ~35% CPU and node using ~75%:
``` js
var dataLength = 0
var time = process.hrtime()
var gzip = spawn('gzip', ['-cd', file], {stdio: ['ignore', 'pipe', 'ignore']})
if (gzip.stdout != null) {
gzip.stdout.on('data', function(data) {
if (!dataLength) dataLength = data.length
})
}
gzip.on('close', function(code) {
var diff = process.hrtime(time)
console.log('stream took %d seconds', diff[0] + diff[1] / 1e9)
if (dataLength) console.log('buffer length was %d', dataLength)
})
```
``` console
$ node spawn-bench.js
stream took 27.406851714 seconds
buffer length was 8192
```
Fine, so that 4x overhead could just be standard pipe/stream overhead, but if I pipe in from stdin, there's really not much overhead at all and it finishes in 7.2 secs, with gzip using ~95% CPU and node ~30%:
``` js
var dataLength = 0
var time = process.hrtime()
process.stdin.on('data', function(data) {
if (!dataLength) dataLength = data.length
})
process.stdin.on('end', function() {
var diff = process.hrtime(time)
console.log('stream took %d seconds', diff[0] + diff[1] / 1e9)
if (dataLength) console.log('buffer length was %d', dataLength)
})
```
``` console
gzip -cd file.gz | node spawn-bench.js
stream took 7.2209929479999995 seconds
buffer length was 65536
```
Similarly, if I process the entire gzip file in node using zlib, I get relatively ok performance too, taking 9.8 secs:
``` js
var dataLength = 0
var time = process.hrtime()
fs.createReadStream(file, {highWaterMark: 32 * 1024 * 1024})
.pipe(zlib.createGunzip({chunkSize: 32 * 1024 * 1024}))
.on('data', function(data) {
if (!dataLength) dataLength = data.length
})
.on('end', function() {
var diff = process.hrtime(time)
console.log('stream took %d seconds', diff[0] + diff[1] / 1e9)
if (dataLength) console.log('buffer length was %d', dataLength)
})
```
``` console
node spawn-bench.js
stream took 9.836914587 seconds
buffer length was 33554432
```
So one thing I suspected was the stream buffer size – you can see when spawning it's only 8k, whereas it's 64k when piping from stdin, and I had better performance upping it even further (to 32MB) when reading from the file and gzipping in node.
However, I couldn't see any way to change this with spawn, so I can't really determine if this is the cause.
Is there something else I'm missing? A 4x slowdown seems to be far greater than expected.
|
1.0
|
Very slow child_process.spawn stdout stream performance on macOS - The `child_process.spawn` stdout stream seems to be taking ~4x longer than other "similar" methods of I/O streaming with similarly sized data streams.
Running on Node.js v4.2.1, Mac OS X 10.11, Macbook Air 1.7GHz
As a baseline, the following program, where `file` is a 472MB gzip file, and `stdout` is set to `ignore` takes 6.5 secs with the gzip process using ~100% CPU:
``` js
var dataLength = 0
var time = process.hrtime()
var gzip = spawn('gzip', ['-cd', file], {stdio: ['ignore', 'ignore', 'ignore']})
if (gzip.stdout != null) {
gzip.stdout.on('data', function(data) {
if (!dataLength) dataLength = data.length
})
}
gzip.on('close', function(code) {
var diff = process.hrtime(time)
console.log('stream took %d seconds', diff[0] + diff[1] / 1e9)
if (dataLength) console.log('buffer length was %d', dataLength)
})
```
``` console
$ node spawn-bench.js
stream took 6.497701762 seconds
```
If I set the `stdout` option from `ignore` to `pipe`, it suddenly takes 27 secs, with gzip only using ~35% CPU and node using ~75%:
``` js
var dataLength = 0
var time = process.hrtime()
var gzip = spawn('gzip', ['-cd', file], {stdio: ['ignore', 'pipe', 'ignore']})
if (gzip.stdout != null) {
gzip.stdout.on('data', function(data) {
if (!dataLength) dataLength = data.length
})
}
gzip.on('close', function(code) {
var diff = process.hrtime(time)
console.log('stream took %d seconds', diff[0] + diff[1] / 1e9)
if (dataLength) console.log('buffer length was %d', dataLength)
})
```
``` console
$ node spawn-bench.js
stream took 27.406851714 seconds
buffer length was 8192
```
Fine, so that 4x overhead could just be standard pipe/stream overhead, but if I pipe in from stdin, there's really not much overhead at all and it finishes in 7.2 secs, with gzip using ~95% CPU and node ~30%:
``` js
var dataLength = 0
var time = process.hrtime()
process.stdin.on('data', function(data) {
if (!dataLength) dataLength = data.length
})
process.stdin.on('end', function() {
var diff = process.hrtime(time)
console.log('stream took %d seconds', diff[0] + diff[1] / 1e9)
if (dataLength) console.log('buffer length was %d', dataLength)
})
```
``` console
gzip -cd file.gz | node spawn-bench.js
stream took 7.2209929479999995 seconds
buffer length was 65536
```
Similarly, if I process the entire gzip file in node using zlib, I get relatively ok performance too, taking 9.8 secs:
``` js
var dataLength = 0
var time = process.hrtime()
fs.createReadStream(file, {highWaterMark: 32 * 1024 * 1024})
.pipe(zlib.createGunzip({chunkSize: 32 * 1024 * 1024}))
.on('data', function(data) {
if (!dataLength) dataLength = data.length
})
.on('end', function() {
var diff = process.hrtime(time)
console.log('stream took %d seconds', diff[0] + diff[1] / 1e9)
if (dataLength) console.log('buffer length was %d', dataLength)
})
```
``` console
node spawn-bench.js
stream took 9.836914587 seconds
buffer length was 33554432
```
So one thing I suspected was the stream buffer size – you can see when spawning it's only 8k, whereas it's 64k when piping from stdin, and I had better performance upping it even further (to 32MB) when reading from the file and gzipping in node.
However, I couldn't see any way to change this with spawn, so I can't really determine if this is the cause.
Is there something else I'm missing? A 4x slowdown seems to be far greater than expected.
|
process
|
very slow child process spawn stdout stream performance on macos the child process spawn stdout stream seems to be taking longer than other similar methods of i o streaming with similarly sized data streams running on node js mac os x macbook air as a baseline the following program where file is a gzip file and stdout is set to ignore takes secs with the gzip process using cpu js var datalength var time process hrtime var gzip spawn gzip stdio if gzip stdout null gzip stdout on data function data if datalength datalength data length gzip on close function code var diff process hrtime time console log stream took d seconds diff diff if datalength console log buffer length was d datalength console node spawn bench js stream took seconds if i set the stdout option from ignore to pipe it suddenly takes secs with gzip only using cpu and node using js var datalength var time process hrtime var gzip spawn gzip stdio if gzip stdout null gzip stdout on data function data if datalength datalength data length gzip on close function code var diff process hrtime time console log stream took d seconds diff diff if datalength console log buffer length was d datalength console node spawn bench js stream took seconds buffer length was fine so that overhead could just be standard pipe stream overhead but if i pipe in from stdin there s really not much overhead at all and it finishes in secs with gzip using cpu and node js var datalength var time process hrtime process stdin on data function data if datalength datalength data length process stdin on end function var diff process hrtime time console log stream took d seconds diff diff if datalength console log buffer length was d datalength console gzip cd file gz node spawn bench js stream took seconds buffer length was similarly if i process the entire gzip file in node using zlib i get relatively ok performance too taking secs js var datalength var time process hrtime fs createreadstream file highwatermark pipe zlib creategunzip chunksize on data function data if datalength datalength data length on end function var diff process hrtime time console log stream took d seconds diff diff if datalength console log buffer length was d datalength console node spawn bench js stream took seconds buffer length was so one thing i suspected was the stream buffer size – you can see when spawning it s only whereas it s when piping from stdin and i had better performance upping it even further to when reading from the file and gzipping in node however i couldn t see any way to change this with spawn so i can t really determine if this is the cause is there something else i m missing a slowdown seems to be far greater than expected
| 1
|
230,421
| 17,616,036,155
|
IssuesEvent
|
2021-08-18 09:49:10
|
rism-digital/verovio
|
https://api.github.com/repos/rism-digital/verovio
|
closed
|
Possible improvements to the test suite output page
|
enhancement documentation / test suite low priority
|
In the new test suite it would be handy to be able to copy and paste a link to view and reference the exact test.
|
1.0
|
Possible improvements to the test suite output page - In the new test suite it would be handy to be able to copy and paste a link to view and reference the exact test.
|
non_process
|
possible improvements to the test suite output page in the new test suite it would be handy to be able to copy and paste a link to view and reference the exact test
| 0
|
2,335
| 5,142,772,284
|
IssuesEvent
|
2017-01-12 14:19:05
|
jimbrown75/Permit-Vision-Enhancements
|
https://api.github.com/repos/jimbrown75/Permit-Vision-Enhancements
|
opened
|
Provide facility for guest signature for Permit Requester
|
Could Fix Low Priority Process Related
|
If a new and short term (often one time only) contractor comes to Site they are not a system user. They can use the Guest Signature as a Permit Holder, but they will often have to define work scope too as Permit Requester, however in current build the Permit Requester must be a system user.
T&S have raised this as a challenge on assets where there are very few staff on Site. The issuer cannot request as well due to 2 signatures required.
|
1.0
|
Provide facility for guest signature for Permit Requester - If a new and short term (often one time only) contractor comes to Site they are not a system user. They can use the Guest Signature as a Permit Holder, but they will often have to define work scope too as Permit Requester, however in current build the Permit Requester must be a system user.
T&S have raised this as a challenge on assets where there are very few staff on Site. The issuer cannot request as well due to 2 signatures required.
|
process
|
provide facility for guest signature for permit requester if a new and short term often one time only contractor comes to site they are not a system user they can use the guest signature as a permit holder but they will often have to define work scope too as permit requester however in current build the permit requester must be a system user t s have raised this as a challenge on assets where there are very few staff on site the issuer cannot request as well due to signatures required
| 1
|
5,620
| 8,476,917,445
|
IssuesEvent
|
2018-10-25 00:08:15
|
easy-software-ufal/annotations_repos
|
https://api.github.com/repos/easy-software-ufal/annotations_repos
|
opened
|
elastic/elasticsearch-net "DataMember" attribute is not applied when deserializing searching query
|
ADA C# test wrong processing
|
Issue: `https://github.com/elastic/elasticsearch-net/issues/3107`
PR: `https://github.com/elastic/elasticsearch-net/commit/57c44dba240e48f2110a61e59fc91bdd662e810d`
|
1.0
|
elastic/elasticsearch-net "DataMember" attribute is not applied when deserializing searching query - Issue: `https://github.com/elastic/elasticsearch-net/issues/3107`
PR: `https://github.com/elastic/elasticsearch-net/commit/57c44dba240e48f2110a61e59fc91bdd662e810d`
|
process
|
elastic elasticsearch net datamember attribute is not applied when deserializing searching query issue pr
| 1
|
79,722
| 10,138,129,389
|
IssuesEvent
|
2019-08-02 17:04:52
|
YugaByte/yugabyte-db
|
https://api.github.com/repos/YugaByte/yugabyte-db
|
opened
|
Doc support for index attributes in JSONB columns
|
area/documentation
|
Answer by @rkarthik007 here, but GitHub issue was never filed by the user:
https://stackoverflow.com/questions/56610522/does-yugabyte-s-sql-support-index-attributes-inside-a-jsonb-column?rq=1
|
1.0
|
Doc support for index attributes in JSONB columns - Answer by @rkarthik007 here, but GitHub issue was never filed by the user:
https://stackoverflow.com/questions/56610522/does-yugabyte-s-sql-support-index-attributes-inside-a-jsonb-column?rq=1
|
non_process
|
doc support for index attributes in jsonb columns answer by here but github issue was never filed by the user
| 0
|
339,453
| 30,448,651,602
|
IssuesEvent
|
2023-07-16 02:03:01
|
ContinualAI/avalanche
|
https://api.github.com/repos/ContinualAI/avalanche
|
opened
|
Creation of a new envirnment failed for python 3.8
|
test Continuous integration
|
Here are the differences between the last working environment and the new one that I tried to run:
```
9c9,10
< - aiohttp=3.7.4.post0=py38h497a2fe_1
---
> - aiohttp=3.8.4=py38h01eb140_1
> - aiosignal=1.3.1=pyhd8ed1ab_0
12c13
< - async-timeout=3.0.1=py_1000
---
> - async-timeout=4.0.2=pyhd8ed1ab_0
18,20c19,21
< - brotli=1.0.9=h166bdaf_8
< - brotli-bin=1.0.9=h166bdaf_8
< - brotlipy=0.7.0=py38h0a891b7_1005
---
> - brotli=1.0.9=h166bdaf_9
> - brotli-bin=1.0.9=h166bdaf_9
> - brotli-python=1.0.9=py38hfa26641_9
22c23
< - c-ares=1.19.0=hd590300_0
---
> - c-ares=1.19.1=hd590300_0
24c25
< - cachetools=5.3.0=pyhd8ed1ab_0
---
> - cachetools=5.3.1=pyhd8ed1ab_0
28,30c29,30
< - chardet=4.0.0=py38h578d9bd_3
< - charset-normalizer=3.1.0=pyhd8ed1ab_0
< - click=8.1.3=unix_pyhd8ed1ab_2
---
> - charset-normalizer=3.2.0=pyhd8ed1ab_0
> - click=8.1.5=unix_pyh707e725_0
32c32
< - contourpy=1.0.7=py38hfbd4bf9_0
---
> - contourpy=1.1.0=py38h7f3f72f_0
34c34
< - cryptography=40.0.2=py38h3d167d9_0
---
> - cryptography=41.0.2=py38hcdda232_0
36c36
< - cython=0.29.34=py38h8dc9893_0
---
> - cython=0.29.36=py38h17151c0_0
41c41
< - fftw=3.3.10=nompi_hc118613_107
---
> - fftw=3.3.10=nompi_hc118613_108
49c49
< - fonttools=4.39.4=py38h01eb140_0
---
> - fonttools=4.41.0=py38h01eb140_0
50a51
> - frozenlist=1.4.0=py38h01eb140_0
53,55c54,56
< - gitpython=3.1.31=pyhd8ed1ab_0
< - glib=2.76.2=hfc55251_0
< - glib-tools=2.76.2=hfc55251_0
---
> - gitpython=3.1.32=pyhd8ed1ab_0
> - glib=2.76.4=hfc55251_0
> - glib-tools=2.76.4=hfc55251_0
58,59c59,60
< - google-auth=2.18.1=pyh1a96a4e_0
< - google-auth-oauthlib=1.0.0=pyhd8ed1ab_0
---
> - google-auth=2.17.3=pyh1a96a4e_0
> - google-auth-oauthlib=1.0.0=pyhd8ed1ab_1
62c63
< - grpcio=1.54.2=py38h8dc9893_2
---
> - grpcio=1.56.2=py38h94a1851_0
65c66
< - gstreamer-orc=0.4.33=h166bdaf_0
---
> - gstreamer-orc=0.4.34=hd590300_0
69,71c70,72
< - importlib-metadata=6.6.0=pyha770c72_0
< - importlib-resources=5.12.0=pyhd8ed1ab_0
< - importlib_resources=5.12.0=pyhd8ed1ab_0
---
> - importlib-metadata=6.8.0=pyha770c72_0
> - importlib-resources=6.0.0=pyhd8ed1ab_1
> - importlib_resources=6.0.0=pyhd8ed1ab_1
73c74
< - joblib=1.2.0=pyhd8ed1ab_0
---
> - joblib=1.3.0=pyhd8ed1ab_1
82c83
< - libabseil=20230125.2=cxx17_h59595ed_2
---
> - libabseil=20230125.3=cxx17_h59595ed_0
84,86c85,87
< - libbrotlicommon=1.0.9=h166bdaf_8
< - libbrotlidec=1.0.9=h166bdaf_8
< - libbrotlienc=1.0.9=h166bdaf_8
---
> - libbrotlicommon=1.0.9=h166bdaf_9
> - libbrotlidec=1.0.9=h166bdaf_9
> - libbrotlienc=1.0.9=h166bdaf_9
98,99c99,100
< - libflac=1.4.2=h27087fc_0
< - libgcc-ng=12.2.0=h65d4601_19
---
> - libflac=1.4.3=h59595ed_0
> - libgcc-ng=13.1.0=he5830b7_0
101,106c102,107
< - libgfortran-ng=12.2.0=h69a702a_19
< - libgfortran5=12.2.0=h337968e_19
< - libglib=2.76.2=hebfc3b9_0
< - libgomp=12.2.0=h65d4601_19
< - libgpg-error=1.46=h620e276_0
< - libgrpc=1.54.2=hb20ce57_2
---
> - libgfortran-ng=13.1.0=h69a702a_0
> - libgfortran5=13.1.0=h15d22d2_0
> - libglib=2.76.4=hebfc3b9_0
> - libgomp=13.1.0=he5830b7_0
> - libgpg-error=1.47=h71f35ed_0
> - libgrpc=1.56.2=h3905398_0
117c118
< - libprotobuf=3.21.12=h3eb15da_0
---
> - libprotobuf=4.23.3=hd1fb520_0
120c121
< - libstdcxx-ng=12.2.0=h46fd767_19
---
> - libstdcxx-ng=13.1.0=hfd8a6a1_0
127c128
< - libwebp-base=1.3.0=h0b41bf4_0
---
> - libwebp-base=1.3.1=hd590300_0
131,132c132,133
< - libzlib=1.2.13=h166bdaf_4
< - llvm-openmp=16.0.4=h4dfa4b3_0
---
> - libzlib=1.2.13=hd590300_5
> - llvm-openmp=16.0.6=h4dfa4b3_0
135,137c136,138
< - markupsafe=2.1.2=py38h1de0b5d_0
< - matplotlib=3.7.1=py38h578d9bd_0
< - matplotlib-base=3.7.1=py38hd6c3c57_0
---
> - markupsafe=2.1.3=py38h01eb140_0
> - matplotlib=3.7.2=py38h578d9bd_0
> - matplotlib-base=3.7.2=py38hf5b0b65_0
144,146c145,147
< - mysql-common=8.0.32=hf1915f5_2
< - mysql-libs=8.0.32=hca2cd23_2
< - ncurses=6.3=h27087fc_1
---
> - mysql-common=8.0.33=hf1915f5_2
> - mysql-libs=8.0.33=hca2cd23_2
> - ncurses=6.4=hcb278e6_0
150c151
< - numpy=1.24.3=py38h59b608b_0
---
> - numpy=1.24.4=py38h59b608b_0
154c155
< - openssl=3.1.0=hd590300_3
---
> - openssl=3.1.1=hd590300_1
159c160
< - pip=23.1.2=pyhd8ed1ab_0
---
> - pip=23.2=pyhd8ed1ab_0
161c162
< - platformdirs=3.5.1=pyhd8ed1ab_0
---
> - platformdirs=3.9.0=pyhd8ed1ab_0
164c165
< - protobuf=4.21.12=py38h8dc9893_0
---
> - protobuf=4.23.3=py38h830738e_0
175c176
< - pyopenssl=23.1.1=pyhd8ed1ab_0
---
> - pyopenssl=23.2.0=pyhd8ed1ab_1
180c181
< - python=3.8.16=he550d4f_1_cpython
---
> - python=3.8.17=he550d4f_0_cpython
191c192
< - requests=2.29.0=pyhd8ed1ab_0
---
> - requests=2.31.0=pyhd8ed1ab_0
194c195
< - scikit-learn=1.2.2=py38hd4b6e60_1
---
> - scikit-learn=1.3.0=py38hc099248_0
196c197
< - sentry-sdk=1.21.1=pyhd8ed1ab_0
---
> - sentry-sdk=1.28.1=pyhd8ed1ab_0
198c199
< - setuptools=67.7.2=pyhd8ed1ab_0
---
> - setuptools=68.0.0=pyhd8ed1ab_0
205c206
< - threadpoolctl=3.1.0=pyh8a188c0_0
---
> - threadpoolctl=3.2.0=pyha21a80b_0
213,214c214,215
< - typing-extensions=4.5.0=hd8ed1ab_0
< - typing_extensions=4.5.0=pyha770c72_0
---
> - typing-extensions=4.7.1=hd8ed1ab_0
> - typing_extensions=4.7.1=pyha770c72_0
216,219c217,220
< - urllib3=1.26.15=pyhd8ed1ab_0
< - wandb=0.15.3=pyhd8ed1ab_0
< - werkzeug=2.3.4=pyhd8ed1ab_0
< - wheel=0.40.0=pyhd8ed1ab_0
---
> - urllib3=2.0.3=pyhd8ed1ab_1
> - wandb=0.15.5=pyhd8ed1ab_0
> - werkzeug=2.3.6=pyhd8ed1ab_0
> - wheel=0.40.0=pyhd8ed1ab_1
227,228c228,229
< - xorg-libice=1.0.10=h7f98852_0
< - xorg-libsm=1.2.3=hd9c2040_1000
---
> - xorg-libice=1.1.1=hd590300_0
> - xorg-libsm=1.2.4=h7391055_0
230c231
< - xorg-libxau=1.0.9=h7f98852_0
---
> - xorg-libxau=1.0.11=hd590300_0
239,242c240,243
< - yarl=1.9.1=py38h01eb140_0
< - zipp=3.15.0=pyhd8ed1ab_0
< - zlib=1.2.13=h166bdaf_4
< - zstd=1.5.2=h3eb15da_6
---
> - yarl=1.9.2=py38h01eb140_0
> - zipp=3.16.2=pyhd8ed1ab_0
> - zlib=1.2.13=hd590300_5
> - zstd=1.5.2=hfc55251_7
249c250
< - filelock==3.12.0
---
> - filelock==3.12.2
256,257c257,258
< - opencv-python==4.7.0.72
< - plotly==5.14.1
---
> - opencv-python==4.8.0.74
> - plotly==5.15.0
```
|
1.0
|
Creation of a new envirnment failed for python 3.8 - Here are the differences between the last working environment and the new one that I tried to run:
```
9c9,10
< - aiohttp=3.7.4.post0=py38h497a2fe_1
---
> - aiohttp=3.8.4=py38h01eb140_1
> - aiosignal=1.3.1=pyhd8ed1ab_0
12c13
< - async-timeout=3.0.1=py_1000
---
> - async-timeout=4.0.2=pyhd8ed1ab_0
18,20c19,21
< - brotli=1.0.9=h166bdaf_8
< - brotli-bin=1.0.9=h166bdaf_8
< - brotlipy=0.7.0=py38h0a891b7_1005
---
> - brotli=1.0.9=h166bdaf_9
> - brotli-bin=1.0.9=h166bdaf_9
> - brotli-python=1.0.9=py38hfa26641_9
22c23
< - c-ares=1.19.0=hd590300_0
---
> - c-ares=1.19.1=hd590300_0
24c25
< - cachetools=5.3.0=pyhd8ed1ab_0
---
> - cachetools=5.3.1=pyhd8ed1ab_0
28,30c29,30
< - chardet=4.0.0=py38h578d9bd_3
< - charset-normalizer=3.1.0=pyhd8ed1ab_0
< - click=8.1.3=unix_pyhd8ed1ab_2
---
> - charset-normalizer=3.2.0=pyhd8ed1ab_0
> - click=8.1.5=unix_pyh707e725_0
32c32
< - contourpy=1.0.7=py38hfbd4bf9_0
---
> - contourpy=1.1.0=py38h7f3f72f_0
34c34
< - cryptography=40.0.2=py38h3d167d9_0
---
> - cryptography=41.0.2=py38hcdda232_0
36c36
< - cython=0.29.34=py38h8dc9893_0
---
> - cython=0.29.36=py38h17151c0_0
41c41
< - fftw=3.3.10=nompi_hc118613_107
---
> - fftw=3.3.10=nompi_hc118613_108
49c49
< - fonttools=4.39.4=py38h01eb140_0
---
> - fonttools=4.41.0=py38h01eb140_0
50a51
> - frozenlist=1.4.0=py38h01eb140_0
53,55c54,56
< - gitpython=3.1.31=pyhd8ed1ab_0
< - glib=2.76.2=hfc55251_0
< - glib-tools=2.76.2=hfc55251_0
---
> - gitpython=3.1.32=pyhd8ed1ab_0
> - glib=2.76.4=hfc55251_0
> - glib-tools=2.76.4=hfc55251_0
58,59c59,60
< - google-auth=2.18.1=pyh1a96a4e_0
< - google-auth-oauthlib=1.0.0=pyhd8ed1ab_0
---
> - google-auth=2.17.3=pyh1a96a4e_0
> - google-auth-oauthlib=1.0.0=pyhd8ed1ab_1
62c63
< - grpcio=1.54.2=py38h8dc9893_2
---
> - grpcio=1.56.2=py38h94a1851_0
65c66
< - gstreamer-orc=0.4.33=h166bdaf_0
---
> - gstreamer-orc=0.4.34=hd590300_0
69,71c70,72
< - importlib-metadata=6.6.0=pyha770c72_0
< - importlib-resources=5.12.0=pyhd8ed1ab_0
< - importlib_resources=5.12.0=pyhd8ed1ab_0
---
> - importlib-metadata=6.8.0=pyha770c72_0
> - importlib-resources=6.0.0=pyhd8ed1ab_1
> - importlib_resources=6.0.0=pyhd8ed1ab_1
73c74
< - joblib=1.2.0=pyhd8ed1ab_0
---
> - joblib=1.3.0=pyhd8ed1ab_1
82c83
< - libabseil=20230125.2=cxx17_h59595ed_2
---
> - libabseil=20230125.3=cxx17_h59595ed_0
84,86c85,87
< - libbrotlicommon=1.0.9=h166bdaf_8
< - libbrotlidec=1.0.9=h166bdaf_8
< - libbrotlienc=1.0.9=h166bdaf_8
---
> - libbrotlicommon=1.0.9=h166bdaf_9
> - libbrotlidec=1.0.9=h166bdaf_9
> - libbrotlienc=1.0.9=h166bdaf_9
98,99c99,100
< - libflac=1.4.2=h27087fc_0
< - libgcc-ng=12.2.0=h65d4601_19
---
> - libflac=1.4.3=h59595ed_0
> - libgcc-ng=13.1.0=he5830b7_0
101,106c102,107
< - libgfortran-ng=12.2.0=h69a702a_19
< - libgfortran5=12.2.0=h337968e_19
< - libglib=2.76.2=hebfc3b9_0
< - libgomp=12.2.0=h65d4601_19
< - libgpg-error=1.46=h620e276_0
< - libgrpc=1.54.2=hb20ce57_2
---
> - libgfortran-ng=13.1.0=h69a702a_0
> - libgfortran5=13.1.0=h15d22d2_0
> - libglib=2.76.4=hebfc3b9_0
> - libgomp=13.1.0=he5830b7_0
> - libgpg-error=1.47=h71f35ed_0
> - libgrpc=1.56.2=h3905398_0
117c118
< - libprotobuf=3.21.12=h3eb15da_0
---
> - libprotobuf=4.23.3=hd1fb520_0
120c121
< - libstdcxx-ng=12.2.0=h46fd767_19
---
> - libstdcxx-ng=13.1.0=hfd8a6a1_0
127c128
< - libwebp-base=1.3.0=h0b41bf4_0
---
> - libwebp-base=1.3.1=hd590300_0
131,132c132,133
< - libzlib=1.2.13=h166bdaf_4
< - llvm-openmp=16.0.4=h4dfa4b3_0
---
> - libzlib=1.2.13=hd590300_5
> - llvm-openmp=16.0.6=h4dfa4b3_0
135,137c136,138
< - markupsafe=2.1.2=py38h1de0b5d_0
< - matplotlib=3.7.1=py38h578d9bd_0
< - matplotlib-base=3.7.1=py38hd6c3c57_0
---
> - markupsafe=2.1.3=py38h01eb140_0
> - matplotlib=3.7.2=py38h578d9bd_0
> - matplotlib-base=3.7.2=py38hf5b0b65_0
144,146c145,147
< - mysql-common=8.0.32=hf1915f5_2
< - mysql-libs=8.0.32=hca2cd23_2
< - ncurses=6.3=h27087fc_1
---
> - mysql-common=8.0.33=hf1915f5_2
> - mysql-libs=8.0.33=hca2cd23_2
> - ncurses=6.4=hcb278e6_0
150c151
< - numpy=1.24.3=py38h59b608b_0
---
> - numpy=1.24.4=py38h59b608b_0
154c155
< - openssl=3.1.0=hd590300_3
---
> - openssl=3.1.1=hd590300_1
159c160
< - pip=23.1.2=pyhd8ed1ab_0
---
> - pip=23.2=pyhd8ed1ab_0
161c162
< - platformdirs=3.5.1=pyhd8ed1ab_0
---
> - platformdirs=3.9.0=pyhd8ed1ab_0
164c165
< - protobuf=4.21.12=py38h8dc9893_0
---
> - protobuf=4.23.3=py38h830738e_0
175c176
< - pyopenssl=23.1.1=pyhd8ed1ab_0
---
> - pyopenssl=23.2.0=pyhd8ed1ab_1
180c181
< - python=3.8.16=he550d4f_1_cpython
---
> - python=3.8.17=he550d4f_0_cpython
191c192
< - requests=2.29.0=pyhd8ed1ab_0
---
> - requests=2.31.0=pyhd8ed1ab_0
194c195
< - scikit-learn=1.2.2=py38hd4b6e60_1
---
> - scikit-learn=1.3.0=py38hc099248_0
196c197
< - sentry-sdk=1.21.1=pyhd8ed1ab_0
---
> - sentry-sdk=1.28.1=pyhd8ed1ab_0
198c199
< - setuptools=67.7.2=pyhd8ed1ab_0
---
> - setuptools=68.0.0=pyhd8ed1ab_0
205c206
< - threadpoolctl=3.1.0=pyh8a188c0_0
---
> - threadpoolctl=3.2.0=pyha21a80b_0
213,214c214,215
< - typing-extensions=4.5.0=hd8ed1ab_0
< - typing_extensions=4.5.0=pyha770c72_0
---
> - typing-extensions=4.7.1=hd8ed1ab_0
> - typing_extensions=4.7.1=pyha770c72_0
216,219c217,220
< - urllib3=1.26.15=pyhd8ed1ab_0
< - wandb=0.15.3=pyhd8ed1ab_0
< - werkzeug=2.3.4=pyhd8ed1ab_0
< - wheel=0.40.0=pyhd8ed1ab_0
---
> - urllib3=2.0.3=pyhd8ed1ab_1
> - wandb=0.15.5=pyhd8ed1ab_0
> - werkzeug=2.3.6=pyhd8ed1ab_0
> - wheel=0.40.0=pyhd8ed1ab_1
227,228c228,229
< - xorg-libice=1.0.10=h7f98852_0
< - xorg-libsm=1.2.3=hd9c2040_1000
---
> - xorg-libice=1.1.1=hd590300_0
> - xorg-libsm=1.2.4=h7391055_0
230c231
< - xorg-libxau=1.0.9=h7f98852_0
---
> - xorg-libxau=1.0.11=hd590300_0
239,242c240,243
< - yarl=1.9.1=py38h01eb140_0
< - zipp=3.15.0=pyhd8ed1ab_0
< - zlib=1.2.13=h166bdaf_4
< - zstd=1.5.2=h3eb15da_6
---
> - yarl=1.9.2=py38h01eb140_0
> - zipp=3.16.2=pyhd8ed1ab_0
> - zlib=1.2.13=hd590300_5
> - zstd=1.5.2=hfc55251_7
249c250
< - filelock==3.12.0
---
> - filelock==3.12.2
256,257c257,258
< - opencv-python==4.7.0.72
< - plotly==5.14.1
---
> - opencv-python==4.8.0.74
> - plotly==5.15.0
```
|
non_process
|
creation of a new envirnment failed for python here are the differences between the last working environment and the new one that i tried to run aiohttp aiohttp aiosignal async timeout py async timeout brotli brotli bin brotlipy brotli brotli bin brotli python c ares c ares cachetools cachetools chardet charset normalizer click unix charset normalizer click unix contourpy contourpy cryptography cryptography cython cython fftw nompi fftw nompi fonttools fonttools frozenlist gitpython glib glib tools gitpython glib glib tools google auth google auth oauthlib google auth google auth oauthlib grpcio grpcio gstreamer orc gstreamer orc importlib metadata importlib resources importlib resources importlib metadata importlib resources importlib resources joblib joblib libabseil libabseil libbrotlicommon libbrotlidec libbrotlienc libbrotlicommon libbrotlidec libbrotlienc libflac libgcc ng libflac libgcc ng libgfortran ng libglib libgomp libgpg error libgrpc libgfortran ng libglib libgomp libgpg error libgrpc libprotobuf libprotobuf libstdcxx ng libstdcxx ng libwebp base libwebp base libzlib llvm openmp libzlib llvm openmp markupsafe matplotlib matplotlib base markupsafe matplotlib matplotlib base mysql common mysql libs ncurses mysql common mysql libs ncurses numpy numpy openssl openssl pip pip platformdirs platformdirs protobuf protobuf pyopenssl pyopenssl python cpython python cpython requests requests scikit learn scikit learn sentry sdk sentry sdk setuptools setuptools threadpoolctl threadpoolctl typing extensions typing extensions typing extensions typing extensions wandb werkzeug wheel wandb werkzeug wheel xorg libice xorg libsm xorg libice xorg libsm xorg libxau xorg libxau yarl zipp zlib zstd yarl zipp zlib zstd filelock filelock opencv python plotly opencv python plotly
| 0
|
262,161
| 22,819,948,882
|
IssuesEvent
|
2022-07-12 00:29:56
|
zeek/zeek
|
https://api.github.com/repos/zeek/zeek
|
closed
|
The `scripts.base.utils.dir` btest is unstable
|
Type: Bug :bug: Area: CI/Testing
|
... at least on my system:
```
$ while true; do btest ./scripts/base/utils/dir.test || break; done
all 1 tests successful
all 1 tests successful
all 1 tests successful
all 1 tests successful
all 1 tests successful
all 1 tests successful
all 1 tests successful
[ 0%] scripts.base.utils.dir ... failed
1 of 1 test failed
```
I'll take a look at this; just creating a ticket for now in case anyone else sees this.
|
1.0
|
The `scripts.base.utils.dir` btest is unstable - ... at least on my system:
```
$ while true; do btest ./scripts/base/utils/dir.test || break; done
all 1 tests successful
all 1 tests successful
all 1 tests successful
all 1 tests successful
all 1 tests successful
all 1 tests successful
all 1 tests successful
[ 0%] scripts.base.utils.dir ... failed
1 of 1 test failed
```
I'll take a look at this; just creating a ticket for now in case anyone else sees this.
|
non_process
|
the scripts base utils dir btest is unstable at least on my system while true do btest scripts base utils dir test break done all tests successful all tests successful all tests successful all tests successful all tests successful all tests successful all tests successful scripts base utils dir failed of test failed i ll take a look at this just creating a ticket for now in case anyone else sees this
| 0
|
785,244
| 27,605,838,498
|
IssuesEvent
|
2023-03-09 12:57:23
|
jenkins-x/jx
|
https://api.github.com/repos/jenkins-x/jx
|
closed
|
Link to user in Jira in changelog
|
kind/enhancement priority/important-longterm area/issue-tracker
|
### Summary
### Steps to reproduce the behavior
Create a release for a project with Jira as the issue tracker
### Expected behavior
Links to assignees for related issues in the changelog leads to the user in Jira.
### Actual behavior
When a changelog is generated for a project with Jira as the issue tracker the links to issues are correct, but links to the assignees of the issues go to github (where the jira user is likely to not exist or actually is another user).
|
1.0
|
Link to user in Jira in changelog - ### Summary
### Steps to reproduce the behavior
Create a release for a project with Jira as the issue tracker
### Expected behavior
Links to assignees for related issues in the changelog leads to the user in Jira.
### Actual behavior
When a changelog is generated for a project with Jira as the issue tracker the links to issues are correct, but links to the assignees of the issues go to github (where the jira user is likely to not exist or actually is another user).
|
non_process
|
link to user in jira in changelog summary steps to reproduce the behavior create a release for a project with jira as the issue tracker expected behavior links to assignees for related issues in the changelog leads to the user in jira actual behavior when a changelog is generated for a project with jira as the issue tracker the links to issues are correct but links to the assignees of the issues go to github where the jira user is likely to not exist or actually is another user
| 0
|
266,446
| 8,367,761,904
|
IssuesEvent
|
2018-10-04 13:10:50
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Need to scroll a lot horizontally to see the full xml message when debugging
|
Component/Composer Component/Debugger Imported Priority/High Type/Improvement
|
<a href="https://github.com/dilinisg"><img src="https://avatars2.githubusercontent.com/u/1845370?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [dilinisg](https://github.com/dilinisg)**
_Tuesday Aug 29, 2017 at 09:32 GMT_
_Originally opened as https://github.com/ballerina-lang/composer/issues/2954_
----
Release 0.93
Need to scroll a lot horizontally to see the full xml message when there are multiple namespaces

|
1.0
|
Need to scroll a lot horizontally to see the full xml message when debugging - <a href="https://github.com/dilinisg"><img src="https://avatars2.githubusercontent.com/u/1845370?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [dilinisg](https://github.com/dilinisg)**
_Tuesday Aug 29, 2017 at 09:32 GMT_
_Originally opened as https://github.com/ballerina-lang/composer/issues/2954_
----
Release 0.93
Need to scroll a lot horizontally to see the full xml message when there are multiple namespaces

|
non_process
|
need to scroll a lot horizontally to see the full xml message when debugging issue by tuesday aug at gmt originally opened as release need to scroll a lot horizontally to see the full xml message when there are multiple namespaces
| 0
|
1,715
| 4,364,870,660
|
IssuesEvent
|
2016-08-03 08:43:06
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
closed
|
Nginx already configured before setup
|
process_wontfix type_bug
|
When you install the packages of openvstorage on a node, the nginx configuration is already in place before the setup is ran.
We noticed this when trying to access the GUI before the installation is ran or when you try to access the GUI on a EXTRA node.
|
1.0
|
Nginx already configured before setup - When you install the packages of openvstorage on a node, the nginx configuration is already in place before the setup is ran.
We noticed this when trying to access the GUI before the installation is ran or when you try to access the GUI on a EXTRA node.
|
process
|
nginx already configured before setup when you install the packages of openvstorage on a node the nginx configuration is already in place before the setup is ran we noticed this when trying to access the gui before the installation is ran or when you try to access the gui on a extra node
| 1
|
5,300
| 8,121,816,604
|
IssuesEvent
|
2018-08-16 09:26:09
|
openvstorage/arakoon
|
https://api.github.com/repos/openvstorage/arakoon
|
closed
|
Arakoon_exc.Exception(4, "Not_found")
|
process_wontfix
|
If a key doesn't exist, `arakoon --get ` will throw an exception. For example,
[root@node-03 ~]# arakoon -config /opt/OpenvStorage/config/arakoon_config.ini --get ovs/framework/scheduling/celery
Uncaught exception:
Arakoon_exc.Exception(4, "Not_found")
Raised at file "src/core/lwt.ml", line 805, characters 16-23
Called from file "src/unix/lwt_main.ml", line 34, characters 8-18
Called from file "src/client/client_main.ml" (inlined), line 36, characters 12-27
Called from file "src/client/client_main.ml", line 87, characters 2-7
Called from file "src/main/arakoon.ml" (inlined), line 493, characters 13-50
Called from file "src/main/arakoon.ml", line 612, characters 7-23
Called from file "src/main/arakoon.ml", line 626, characters 9-16
It would be friendly if it reports sth like 'no key xxx defined', 'empty value' or exits silently instead of throwing an exception as above.
|
1.0
|
Arakoon_exc.Exception(4, "Not_found") - If a key doesn't exist, `arakoon --get ` will throw an exception. For example,
[root@node-03 ~]# arakoon -config /opt/OpenvStorage/config/arakoon_config.ini --get ovs/framework/scheduling/celery
Uncaught exception:
Arakoon_exc.Exception(4, "Not_found")
Raised at file "src/core/lwt.ml", line 805, characters 16-23
Called from file "src/unix/lwt_main.ml", line 34, characters 8-18
Called from file "src/client/client_main.ml" (inlined), line 36, characters 12-27
Called from file "src/client/client_main.ml", line 87, characters 2-7
Called from file "src/main/arakoon.ml" (inlined), line 493, characters 13-50
Called from file "src/main/arakoon.ml", line 612, characters 7-23
Called from file "src/main/arakoon.ml", line 626, characters 9-16
It would be friendly if it reports sth like 'no key xxx defined', 'empty value' or exits silently instead of throwing an exception as above.
|
process
|
arakoon exc exception not found if a key doesn t exist arakoon get will throw an exception for example arakoon config opt openvstorage config arakoon config ini get ovs framework scheduling celery uncaught exception arakoon exc exception not found raised at file src core lwt ml line characters called from file src unix lwt main ml line characters called from file src client client main ml inlined line characters called from file src client client main ml line characters called from file src main arakoon ml inlined line characters called from file src main arakoon ml line characters called from file src main arakoon ml line characters it would be friendly if it reports sth like no key xxx defined empty value or exits silently instead of throwing an exception as above
| 1
|
605,782
| 18,740,514,884
|
IssuesEvent
|
2021-11-04 13:06:55
|
AY2122S1-CS2103T-W13-3/tp
|
https://api.github.com/repos/AY2122S1-CS2103T-W13-3/tp
|
closed
|
Sort feature does not work when list is edited
|
bug priority.Medium
|
Before (game 1 has date being `2021-10-30`):

After (edited previous game 1 to have date `2020-10-30`):

Game 1 should have been moved down to maintain sortedness of the game list.
|
1.0
|
Sort feature does not work when list is edited - Before (game 1 has date being `2021-10-30`):

After (edited previous game 1 to have date `2020-10-30`):

Game 1 should have been moved down to maintain sortedness of the game list.
|
non_process
|
sort feature does not work when list is edited before game has date being after edited previous game to have date game should have been moved down to maintain sortedness of the game list
| 0
|
13,233
| 15,704,377,220
|
IssuesEvent
|
2021-03-26 14:56:06
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Prisma generate should fail for Typescript users below 4.1
|
good first issue kind/improvement process/candidate team/client
|
## Problem
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
The Prisma Client's generated types requires Typescript 4.1 and above. Right now we show a warning, but should fail because it's too easy to miss this hard requirement.
```
npm run build
> website-v8@1.0.0 build /Users/m/dev/src/github.com/prisma/homepage-v8
> npm run generate:explore && next build
> website-v8@1.0.0 generate:explore /Users/m/dev/src/github.com/prisma/homepage-v8
> prisma generate --schema=./ui/explore/schema.prisma && node ./scripts/generate-explore.js
Prisma schema loaded from ui/explore/schema.prisma
**prisma:warn Your typescript version is 4.0.3, which is outdated. Please update it to 4.1.0 or newer in order to use Prisma Client.**
✔ Generated Prisma Client (2.19.0) to ./node_modules/@prisma/client in 152ms
You can now start using Prisma Client in your code. Reference: https://pris.ly/d/client
import { PrismaClient } from '@prisma/client'
const prisma = new PrismaClient()
Browserslist: caniuse-lite is outdated. Please run:
npx browserslist@latest --update-db
Failed to compile.
./node_modules/.prisma/client/index.d.ts:11:33
Type error: Type expected.
9 | type UnwrapPromise<P extends any> = P extends Promise<infer R> ? R : P
10 | type UnwrapTuple<Tuple extends readonly unknown[]> = {
> 11 | [K in keyof Tuple]: K extends `${number}` ? Tuple[K] extends PrismaPromise<infer X> ? X : UnwrapPromise<Tuple[K]> : UnwrapPromise<Tuple[K]>
| ^
12 | };
13 |
14 |
info - Creating an optimized production build .npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! website-v8@1.0.0 build: `npm run generate:explore && next build`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the website-v8@1.0.0 build script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /Users/m/.npm/_logs/2021-03-19T12_46_24_629Z-debug.log
```
## Suggested solution
<!-- A clear and concise description of what you want to happen. -->
Show an error during `prisma generate` if the user is using a Typescript version that's less than 4.1.
**Warning** that this is only a problem for Typescript users, not Javascript users. We shouldn't make this a hard requirement for Javascript users.
|
1.0
|
Prisma generate should fail for Typescript users below 4.1 - ## Problem
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
The Prisma Client's generated types requires Typescript 4.1 and above. Right now we show a warning, but should fail because it's too easy to miss this hard requirement.
```
npm run build
> website-v8@1.0.0 build /Users/m/dev/src/github.com/prisma/homepage-v8
> npm run generate:explore && next build
> website-v8@1.0.0 generate:explore /Users/m/dev/src/github.com/prisma/homepage-v8
> prisma generate --schema=./ui/explore/schema.prisma && node ./scripts/generate-explore.js
Prisma schema loaded from ui/explore/schema.prisma
**prisma:warn Your typescript version is 4.0.3, which is outdated. Please update it to 4.1.0 or newer in order to use Prisma Client.**
✔ Generated Prisma Client (2.19.0) to ./node_modules/@prisma/client in 152ms
You can now start using Prisma Client in your code. Reference: https://pris.ly/d/client
import { PrismaClient } from '@prisma/client'
const prisma = new PrismaClient()
Browserslist: caniuse-lite is outdated. Please run:
npx browserslist@latest --update-db
Failed to compile.
./node_modules/.prisma/client/index.d.ts:11:33
Type error: Type expected.
9 | type UnwrapPromise<P extends any> = P extends Promise<infer R> ? R : P
10 | type UnwrapTuple<Tuple extends readonly unknown[]> = {
> 11 | [K in keyof Tuple]: K extends `${number}` ? Tuple[K] extends PrismaPromise<infer X> ? X : UnwrapPromise<Tuple[K]> : UnwrapPromise<Tuple[K]>
| ^
12 | };
13 |
14 |
info - Creating an optimized production build .npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! website-v8@1.0.0 build: `npm run generate:explore && next build`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the website-v8@1.0.0 build script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /Users/m/.npm/_logs/2021-03-19T12_46_24_629Z-debug.log
```
## Suggested solution
<!-- A clear and concise description of what you want to happen. -->
Show an error during `prisma generate` if the user is using a Typescript version that's less than 4.1.
**Warning** that this is only a problem for Typescript users, not Javascript users. We shouldn't make this a hard requirement for Javascript users.
|
process
|
prisma generate should fail for typescript users below problem the prisma client s generated types requires typescript and above right now we show a warning but should fail because it s too easy to miss this hard requirement npm run build website build users m dev src github com prisma homepage npm run generate explore next build website generate explore users m dev src github com prisma homepage prisma generate schema ui explore schema prisma node scripts generate explore js prisma schema loaded from ui explore schema prisma prisma warn your typescript version is which is outdated please update it to or newer in order to use prisma client ✔ generated prisma client to node modules prisma client in you can now start using prisma client in your code reference import prismaclient from prisma client const prisma new prismaclient browserslist caniuse lite is outdated please run npx browserslist latest update db failed to compile node modules prisma client index d ts type error type expected type unwrappromise p extends promise r p type unwraptuple k extends number tuple extends prismapromise x unwrappromise unwrappromise info creating an optimized production build npm err code elifecycle npm err errno npm err website build npm run generate explore next build npm err exit status npm err npm err failed at the website build script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err users m npm logs debug log suggested solution show an error during prisma generate if the user is using a typescript version that s less than warning that this is only a problem for typescript users not javascript users we shouldn t make this a hard requirement for javascript users
| 1
|
123,674
| 4,867,003,158
|
IssuesEvent
|
2016-11-15 02:10:20
|
FredHutch/Oncoscape
|
https://api.github.com/repos/FredHutch/Oncoscape
|
closed
|
TCGAbrca copy number and mut patient ids do not intersect
|
bug datasets priority
|
> head(rownames(mtx.mu))
> [1] "TCGA.BH.A0DZ.01" "TCGA.AN.A046.01" "TCGA.BH.A0DG.01" "TCGA.BH.A1F8.01" "TCGA.A7.A13D.01" "TCGA.A8.A0A6.01"
> head(rownames(mtx.cn))
> [1] "TCGA.05.4244.01" "TCGA.05.4249.01" "TCGA.05.4250.01" "TCGA.05.4382.01" "TCGA.05.4384.01" "TCGA.05.4389.01"
> length(intersect(rownames(mtx.cn), rownames(mtx.mu))) # [1] 0
dim(mtx.cn) # [1] 515 22184
dim(mtx.mut) # [1] 980 15417
|
1.0
|
TCGAbrca copy number and mut patient ids do not intersect - > head(rownames(mtx.mu))
> [1] "TCGA.BH.A0DZ.01" "TCGA.AN.A046.01" "TCGA.BH.A0DG.01" "TCGA.BH.A1F8.01" "TCGA.A7.A13D.01" "TCGA.A8.A0A6.01"
> head(rownames(mtx.cn))
> [1] "TCGA.05.4244.01" "TCGA.05.4249.01" "TCGA.05.4250.01" "TCGA.05.4382.01" "TCGA.05.4384.01" "TCGA.05.4389.01"
> length(intersect(rownames(mtx.cn), rownames(mtx.mu))) # [1] 0
dim(mtx.cn) # [1] 515 22184
dim(mtx.mut) # [1] 980 15417
|
non_process
|
tcgabrca copy number and mut patient ids do not intersect head rownames mtx mu tcga bh tcga an tcga bh tcga bh tcga tcga head rownames mtx cn tcga tcga tcga tcga tcga tcga length intersect rownames mtx cn rownames mtx mu dim mtx cn dim mtx mut
| 0
|
3,955
| 6,892,352,935
|
IssuesEvent
|
2017-11-22 20:40:28
|
PWRFLcreative/Lightwork-Mapper
|
https://api.github.com/repos/PWRFLcreative/Lightwork-Mapper
|
opened
|
Try using image sequence for capture
|
Processing
|
Store sequence of images, instead of raw stream (sync problems) or recorded video. (performance problems)
|
1.0
|
Try using image sequence for capture - Store sequence of images, instead of raw stream (sync problems) or recorded video. (performance problems)
|
process
|
try using image sequence for capture store sequence of images instead of raw stream sync problems or recorded video performance problems
| 1
|
15,635
| 19,805,527,312
|
IssuesEvent
|
2022-01-19 06:04:21
|
redwoodjs/redwood
|
https://api.github.com/repos/redwoodjs/redwood
|
closed
|
`rw-test-app` improvements
|
triage/processing
|
cc @jtoar / @dthyresson / @thedavidprice
I noticed while working on #3515 that the generated `jest` tests and `storybook` stories are broken.
For example, the generated `BlogPost.stories.tsx`, and `BlogPost.test.tsx` do not have the proper mock passed into them (see screenshot).

There is more, such as the wrong query from `'types/graphql'` being used in `BlogPostCell.tsx`. We try to `import type { FindBlogPostQuery } from 'types/graphql'` , but `FindBlogPostQuery` does not exist (there are types such as `BlogPostQuery` , and `FindPostById` but it is sort of unclear which to use [probably `BlogPostQuery`]).
---
Looking at some prior art in the codebase, I think the gist of what is needed to fix is updating some of the `codemods` in https://github.com/redwoodjs/redwood/tree/main/tasks/test-project, maybe leveraging https://github.com/redwoodjs/redwood/tree/main/tasks/e2e/cypress/integration/03-storybook/codemods ?
---
Linked #3287
|
1.0
|
`rw-test-app` improvements - cc @jtoar / @dthyresson / @thedavidprice
I noticed while working on #3515 that the generated `jest` tests and `storybook` stories are broken.
For example, the generated `BlogPost.stories.tsx`, and `BlogPost.test.tsx` do not have the proper mock passed into them (see screenshot).

There is more, such as the wrong query from `'types/graphql'` being used in `BlogPostCell.tsx`. We try to `import type { FindBlogPostQuery } from 'types/graphql'` , but `FindBlogPostQuery` does not exist (there are types such as `BlogPostQuery` , and `FindPostById` but it is sort of unclear which to use [probably `BlogPostQuery`]).
---
Looking at some prior art in the codebase, I think the gist of what is needed to fix is updating some of the `codemods` in https://github.com/redwoodjs/redwood/tree/main/tasks/test-project, maybe leveraging https://github.com/redwoodjs/redwood/tree/main/tasks/e2e/cypress/integration/03-storybook/codemods ?
---
Linked #3287
|
process
|
rw test app improvements cc jtoar dthyresson thedavidprice i noticed while working on that the generated jest tests and storybook stories are broken for example the generated blogpost stories tsx and blogpost test tsx do not have the proper mock passed into them see screenshot there is more such as the wrong query from types graphql being used in blogpostcell tsx we try to import type findblogpostquery from types graphql but findblogpostquery does not exist there are types such as blogpostquery and findpostbyid but it is sort of unclear which to use looking at some prior art in the codebase i think the gist of what is needed to fix is updating some of the codemods in maybe leveraging linked
| 1
|
3,122
| 6,153,851,397
|
IssuesEvent
|
2017-06-28 11:04:20
|
openvstorage/volumedriver
|
https://api.github.com/repos/openvstorage/volumedriver
|
closed
|
Observation: halted volumes after volume move to node with full root partition
|
process_wontfix
|
The move from volumes failed due to a full root partition on the destination node. All of the volumes got in halted state (on the destination node). The MDS failed.
```
root@stor-04:~# df -h
Filesystem Size Used Avail Use% Mounted on
udev 126G 0 126G 0% /dev
tmpfs 26G 2.6G 23G 11% /run
/dev/md1 20G 19G 0 100% /
tmpfs 126G 236K 126G 1% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 126G 0 126G 0% /sys/fs/cgroup
/dev/md1 20G 19G 0 100% /home
```
```
2017-06-22 11:13:24 251318 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e38eb - info - execute: Arguments for migrateVolume are {[force:false,volume_id:c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:24 251384 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e38ec - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:24 251404 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e38ed - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:24 251449 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VFSObjectRouter - 00000000003e38ee - info - migrate: Migrating c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:24 251784 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VFSObjectRouter - 00000000003e38ef - info - migrate_: Trying to migrate c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d from data01PAwWiMYt0g9OwJZl, only steal if offline OnlyStealFromOfflineNode::T
2017-06-22 11:13:25 280926 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VolManager - 00000000003e38f0 - notice - Backend Restart, Namespace: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d, owner tag 106, START
2017-06-22 11:13:25 280997 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f1 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:25 281118 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f2 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:25 281136 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f3 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:25 281348 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38f4 - info - Logger: Entering invalidate_cache c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 281365 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e38f5 - info - invalidate_cache_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: requesting cache invalidation
2017-06-22 11:13:25 281440 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38f6 - info - ~Logger: Exiting invalidate_cache for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 281474 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f7 - info - logMessage: TCP_transport(10.109.3.44, 26229)
2017-06-22 11:13:25 281545 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f8 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:25 281560 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f9 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:25 281753 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38fa - info - Logger: Entering invalidate_cache c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 281767 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e38fb - info - invalidate_cache_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: requesting cache invalidation
2017-06-22 11:13:25 285147 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38fc - info - ~Logger: Exiting invalidate_cache for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 286064 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38fd - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:25 323160 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e38fe - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:25 327447 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38ff - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:25 327483 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3900 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 327514 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3901 - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:25 327615 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3902 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:25 327746 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3903 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:25 327767 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3904 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:25 327902 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3905 - warning - do_wrap_: Retrying with new connection (retry: 1, sleep before retry: 1000 milliseconds)
2017-06-22 11:13:26 328053 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3906 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:26 372696 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3907 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:26 385614 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3908 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:26 385684 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3909 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:26 385721 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e390a - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:26 385972 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e390b - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:26 386143 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e390c - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:26 386172 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e390d - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:26 386349 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e390e - warning - do_wrap_: Retrying with new connection (retry: 2, sleep before retry: 2000 milliseconds)
2017-06-22 11:13:27 334223 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e390f - info - work: You have 2 connections running
2017-06-22 11:13:27 334294 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e3910 - info - cleanup_threads_locked: Trying to join 2 threads
2017-06-22 11:13:27 334382 +0200 - stor-04.be-g8-4 - 29131/0x00007fab857f9700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e3911 - info - execute: Arguments for volumeInfo are {[redirect_fenced:true,volume_id:c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a]}
2017-06-22 11:13:27 520215 +0200 - stor-04.be-g8-4 - 29131/0x00007fab7a7fb700 - volumedriverfs/VFSRemoteNode - 00000000003e3912 - info - operator(): got keepalive rsp from data01OpKZbWWujImRDMBt
2017-06-22 11:13:27 520232 +0200 - stor-04.be-g8-4 - 29131/0x00007fab79ffa700 - volumedriverfs/VFSRemoteNode - 00000000003e3913 - info - operator(): got keepalive rsp from data01PAwWiMYt0g9OwJZl
2017-06-22 11:13:27 520242 +0200 - stor-04.be-g8-4 - 29131/0x00007fab7affc700 - volumedriverfs/VFSRemoteNode - 00000000003e3914 - info - operator(): got keepalive rsp from data012BSgz3tz2dJvdW1e
2017-06-22 11:13:28 386486 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3915 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:28 412787 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3916 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:28 415098 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3917 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:28 415138 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3918 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:28 415170 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3919 - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:28 415275 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e391a - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:28 415401 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e391b - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:28 415428 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e391c - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:28 415673 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e391d - warning - do_wrap_: Retrying with new connection (retry: 3, sleep before retry: 4000 milliseconds)
2017-06-22 11:13:29 340281 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e391e - info - work: You have 3 connections running
2017-06-22 11:13:29 340349 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e391f - info - cleanup_threads_locked: Trying to join 3 threads
2017-06-22 11:13:29 340427 +0200 - stor-04.be-g8-4 - 29131/0x00007fabb97fa700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e3920 - info - execute: Arguments for volumePerformanceCountersV3 are {[reset:false,volume_id:c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a]}
2017-06-22 11:13:31 605394 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3921 - info - Logger: Entering read fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a snapshots.xml
2017-06-22 11:13:31 625614 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/AlbaConnection - 00000000003e3922 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:31 625926 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3923 - error - ~Logger: Exiting read for fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a snapshots.xml with exception
2017-06-22 11:13:31 625953 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3924 - error - ~Logger: Exiting read for fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a
2017-06-22 11:13:31 625974 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendInterface - 00000000003e3925 - error - do_wrap_: Problem with connection 0x7fab14024a40: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:31 625989 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendInterface - 00000000003e3926 - error - do_wrap_: Giving up connection 0x7fab14024a40
2017-06-22 11:13:31 626121 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendInterface - 00000000003e3927 - error - fillObject: Problem getting snapshots.xml from fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:31 626513 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/MDSMetaDataBackend - 00000000003e3928 - info - ~MDSMetaDataBackend: fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a: used clusters: 0
2017-06-22 11:13:31 626599 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/PeriodicActionPoolTaskImpl - 00000000003e3929 - error - run: mds-poll-namespace-fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a: caught exception: (Unix.Unix_error "No space left on device" write "") - ignored
2017-06-22 11:13:32 371894 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e392a - info - work: You have 4 connections running
2017-06-22 11:13:32 371943 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e392b - info - cleanup_threads_locked: Trying to join 4 threads
2017-06-22 11:13:32 372096 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e392c - info - execute: Arguments for ListClientConnections are {[vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:32 372170 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e392d - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:32 372201 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e392e - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:32 372218 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e392f - info - execute: Call ListClientConnections took 5.7e-05 seconds
2017-06-22 11:13:32 372276 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCConnection - 00000000003e3930 - info - operator(): Closing socket 102
2017-06-22 11:13:32 415845 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3931 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:32 455585 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3932 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:32 455871 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3933 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:32 455901 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3934 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:32 455925 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3935 - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:32 456071 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3936 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:32 456206 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3937 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:32 456234 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3938 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:32 456368 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3939 - warning - do_wrap_: Retrying with new connection (retry: 4, sleep before retry: 8000 milliseconds)
2017-06-22 11:13:36 769956 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e393a - info - work: You have 5 connections running
2017-06-22 11:13:36 770018 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e393b - info - cleanup_threads_locked: Trying to join 5 threads
2017-06-22 11:13:36 770078 +0200 - stor-04.be-g8-4 - 29131/0x00007fab667f8700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e393c - info - execute: Arguments for volumeDriverPerformanceCountersV3 are {[reset:false,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:36 770164 +0200 - stor-04.be-g8-4 - 29131/0x00007fab667f8700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e393d - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:36 770189 +0200 - stor-04.be-g8-4 - 29131/0x00007fab667f8700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e393e - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:40 456516 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e393f - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:40 488395 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3940 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:40 488615 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3941 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:40 488636 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3942 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:40 488653 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3943 - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:40 488758 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3944 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:40 488863 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3945 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:40 488882 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3946 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:40 489001 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3947 - warning - do_wrap_: Retrying with new connection (retry: 5, sleep before retry: 16000 milliseconds)
2017-06-22 11:13:42 520385 +0200 - stor-04.be-g8-4 - 29131/0x00007fab7affc700 - volumedriverfs/VFSRemoteNode - 00000000003e3948 - info - operator(): got keepalive rsp from data012BSgz3tz2dJvdW1e
2017-06-22 11:13:42 520419 +0200 - stor-04.be-g8-4 - 29131/0x00007fab7a7fb700 - volumedriverfs/VFSRemoteNode - 00000000003e3949 - info - operator(): got keepalive rsp from data01OpKZbWWujImRDMBt
2017-06-22 11:13:42 520415 +0200 - stor-04.be-g8-4 - 29131/0x00007fab79ffa700 - volumedriverfs/VFSRemoteNode - 00000000003e3949 - info - operator(): got keepalive rsp from data01PAwWiMYt0g9OwJZl
2017-06-22 11:13:46 290833 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e394a - info - work: You have 5 connections running
2017-06-22 11:13:46 290901 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e394b - info - cleanup_threads_locked: Trying to join 5 threads
2017-06-22 11:13:46 291026 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e394c - info - execute: Arguments for volumeDriverPerformanceCountersV3 are {[reset:false,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:46 291092 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e394d - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:46 291111 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e394e - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:50 811098 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MetaDataServerTable - 00000000003e394f - info - work_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: running periodic action
2017-06-22 11:13:50 811174 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MDSMetaDataBackend - 00000000003e3950 - info - MDSMetaDataBackend: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:50 811332 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MDSMetaDataBackend - 00000000003e3951 - info - lastCorkUUID: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: 3e8a6914-e4f4-4d2e-9343-c1aef081a0d8
2017-06-22 11:13:50 811395 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MDSMetaDataBackend - 00000000003e3952 - info - scrub_id: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: scrub ID 28ee90a9-374b-49e8-85aa-07bf303aad5f
2017-06-22 11:13:50 811430 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/CachedMetaDataStore - 00000000003e3953 - info - init_pages_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: page capacity (entries): 64, max cached pages: 256
2017-06-22 11:13:50 811537 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MetaDataStoreBuilder - 00000000003e3954 - info - update_metadata_store_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: bringing MetaDataStore in sync with backend, requested interval ( 3e8a6914-e4f4-4d2e-9343-c1aef081a0d8, --], check scrub ID: CheckScrubId::F, dry run:DryRun::F, full rebuild: false
2017-06-22 11:13:50 815978 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3955 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:50 816185 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3956 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:50 816216 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3957 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:50 816448 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3958 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml
2017-06-22 11:13:50 842623 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaConnection - 00000000003e3959 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:50 848631 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e395a - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml with exception
2017-06-22 11:13:50 848676 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e395b - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:50 848706 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e395c - error - do_wrap_: Problem with connection 0x7fab18007ae0: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:50 848855 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e395d - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:50 849003 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e395e - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:50 849031 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e395f - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:50 849166 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e3960 - warning - do_wrap_: Retrying with new connection (retry: 1, sleep before retry: 1000 milliseconds)
2017-06-22 11:13:51 849290 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3961 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml
2017-06-22 11:13:51 874257 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaConnection - 00000000003e3962 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:51 874479 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3963 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml with exception
2017-06-22 11:13:51 874503 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3964 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:51 874525 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e3965 - error - do_wrap_: Problem with connection 0x7fab18007ae0: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:51 874685 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3966 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:51 874808 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3967 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:51 874833 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3968 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:51 874974 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e3969 - warning - do_wrap_: Retrying with new connection (retry: 2, sleep before retry: 2000 milliseconds)
2017-06-22 11:13:53 875101 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e396a - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml
2017-06-22 11:13:53 898792 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaConnection - 00000000003e396b - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:53 899042 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e396c - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml with exception
2017-06-22 11:13:53 899066 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e396d - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:53 899088 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e396e - error - do_wrap_: Problem with connection 0x7fab18007ae0: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:53 899212 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e396f - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:53 899330 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3970 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:53 899371 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3971 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:53 899499 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e3972 - warning - do_wrap_: Retrying with new connection (retry: 3, sleep before retry: 4000 milliseconds)
2017-06-22 11:13:56 270787 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e3973 - info - work: You have 6 connections running
2017-06-22 11:13:56 270850 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e3974 - info - cleanup_threads_locked: Trying to join 6 threads
2017-06-22 11:13:56 270976 +0200 - stor-04.be-g8-4 - 29131/0x00007fab65ff7700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e3975 - info - execute: Arguments for volumeDriverPerformanceCountersV3 are {[reset:false,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:56 271041 +0200 - stor-04.be-g8-4 - 29131/0x00007fab65ff7700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e3976 - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:56 271063 +0200 - stor-04.be-g8-4 - 29131/0x00007fab65ff7700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e3977 - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:56 489152 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3978 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:56 527885 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3979 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528152 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e397a - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:56 528179 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e397b - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:56 528219 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e397c - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528234 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e397d - error - do_wrap_: Giving up connection 0x7faa0d9a5040
2017-06-22 11:13:56 528405 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e397e - error - fillObject: Problem getting volume_configuration from c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528466 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VolManager - 00000000003e397f - notice - Backend Restart, Namespace: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d, owner tag 106, EXCEPTION
2017-06-22 11:13:56 528495 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VFSLocalNode - 00000000003e3980 - error - restart_volume_from_backend_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: failed to restart volume: : (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528604 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e3981 - error - execute: migrateVolume Caught std::exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528607 +0200 - stor-04.be-g8-4 - 29131/0x00007fab857f9700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e3982 - info - execute: Object c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d is not present on this node - figuring out if it lives elsewhere.
2017-06-22 11:13:56 528715 +0200 - stor-04.be-g8-4 - 29131/0x00007fabb97fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e3983 - info - execute: Object c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d is not present on this node - figuring out if it lives elsewhere.
2017-06-22 11:13:56 528727 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCConnection - 00000000003e3984 - info - operator(): Closing socket 139
```
Volumedriver config:
```
root@stor-04:~# ovs config get ovs/vpools/215304e1-f23b-4a2d-aba9-67937960a17a/hosts/data01OdDMLuDUBCNEneoS/config
{
"content_addressed_cache": {
"read_cache_serialization_path": [],
"serialize_read_cache": false
},
"backend_connection_manager": {
"backend_interface_retries_on_error": 5,
"backend_type": "MULTI",
"1": {
"alba_connection_preset": "gigPreset",
"backend_type": "ALBA",
"alba_connection_transport": "TCP",
"alba_connection_timeout": 15,
"alba_connection_use_rora": true,
"alba_connection_host": "10.109.3.44",
"alba_connection_rora_manifest_cache_capacity": 17179869184,
"alba_connection_port": 26229
},
"0": {
"alba_connection_preset": "gigPreset",
"backend_type": "ALBA",
"alba_connection_transport": "TCP",
"alba_connection_timeout": 15,
"alba_connection_use_rora": true,
"alba_connection_host": "10.109.3.44",
"alba_connection_rora_manifest_cache_capacity": 17179869184,
"alba_connection_port": 26228
},
"backend_interface_retry_backoff_multiplier": 2.0,
"backend_interface_retry_interval_secs": 1
},
"distributed_lock_store": {
"dls_arakoon_cluster_nodes": [
{
"host": "10.109.3.42",
"node_id": "2BSgz3tz2dJvdW1e",
"port": 26416
},
{
"host": "10.109.3.41",
"node_id": "OpKZbWWujImRDMBt",
"port": 26416
},
{
"host": "10.109.3.43",
"node_id": "PAwWiMYt0g9OwJZl",
"port": 26416
}
],
"dls_arakoon_cluster_id": "voldrv",
"dls_type": "Arakoon"
},
"volume_registry": {
"vregistry_arakoon_cluster_id": "voldrv",
"vregistry_arakoon_cluster_nodes": [
{
"host": "10.109.3.42",
"node_id": "2BSgz3tz2dJvdW1e",
"port": 26416
},
{
"host": "10.109.3.41",
"node_id": "OpKZbWWujImRDMBt",
"port": 26416
},
{
"host": "10.109.3.43",
"node_id": "PAwWiMYt0g9OwJZl",
"port": 26416
}
]
},
"volume_router_cluster": {
"vrouter_cluster_id": "215304e1-f23b-4a2d-aba9-67937960a17a"
},
"volume_router": {
"vrouter_migrate_timeout_ms": 60000,
"vrouter_sco_multiplier": 16384,
"vrouter_keepalive_interval_secs": "5",
"vrouter_volume_read_threshold": 0,
"vrouter_id": "data01OdDMLuDUBCNEneoS",
"vrouter_use_fencing": true,
"vrouter_keepalive_time_secs": "15",
"vrouter_file_write_threshold": 0,
"vrouter_volume_write_threshold": 0,
"vrouter_backend_sync_timeout_ms": 60000,
"vrouter_min_workers": 4,
"vrouter_keepalive_retries": "2",
"vrouter_routing_retries": 10,
"vrouter_redirect_timeout_ms": "120000",
"vrouter_file_read_threshold": 0,
"vrouter_max_workers": 16
},
"volume_manager": {
"clean_interval": 1,
"tlog_path": "/mnt/ssd3/data01_db_tlog_1",
"non_disposable_scos_factor": 2.0,
"metadata_path": "/mnt/ssd3/data01_db_md_1",
"number_of_scos_in_tlog": 1,
"default_cluster_size": 4096,
"dtl_throttle_usecs": 4000
},
"scocache": {
"trigger_gap": "8GiB",
"backoff_gap": "10GiB",
"scocache_mount_points": [
{
"path": "/mnt/ssd1/data01_write_sco_1",
"size": "100188160KiB"
},
{
"path": "/mnt/ssd4/data01_write_sco_1",
"size": "100188160KiB"
}
]
},
"filesystem": {
"fs_raw_disk_suffix": ".raw",
"fs_enable_shm_interface": 0,
"fs_metadata_backend_arakoon_cluster_nodes": [],
"fs_dtl_mode": "Synchronous",
"fs_metadata_backend_mds_nodes": [
{
"host": "10.109.3.44",
"port": 26304
}
],
"fs_metadata_backend_type": "MDS",
"fs_virtual_disk_format": "raw",
"fs_enable_network_interface": 1,
"fs_dtl_config_mode": "Automatic",
"fs_dtl_host": "",
"fs_file_event_rules": [
{
"fs_file_event_rule_calls": [
"Rename"
],
"fs_file_event_rule_path_regex": ".*"
}
]
},
"file_driver": {
"fd_extent_cache_capacity": "1024",
"fd_namespace": "fd-data01-215304e1-f23b-4a2d-aba9-67937960a17a",
"fd_cache_path": "/mnt/ssd1/data01_write_fd_1"
},
"threadpool_component": {
"num_threads": 16
},
"event_publisher": {
"events_amqp_routing_key": "volumerouter",
"events_amqp_uris": [
{
"amqp_uri": "amqp://ovs:0penv5tor4ge@10.109.3.41:5672"
},
{
"amqp_uri": "amqp://ovs:0penv5tor4ge@10.109.3.42:5672"
},
{
"amqp_uri": "amqp://ovs:0penv5tor4ge@10.109.3.43:5672"
}
]
},
"metadata_server": {
"mds_nodes": [
{
"host": "10.109.3.44",
"scratch_directory": "/mnt/ssd3/data01_db_mds_1/scratch",
"port": 26304,
"db_directory": "/mnt/ssd3/data01_db_mds_1/db"
}
]
},
"distributed_transaction_log": {
"dtl_path": "/mnt/ssd5/data01_dtl_1",
"dtl_transport": "TCP"
},
"network_interface": {
"network_max_neighbour_distance": 9999
}
}
```
|
1.0
|
Observation: halted volumes after volume move to node with full root partition - The move from volumes failed due to a full root partition on the destination node. All of the volumes got in halted state (on the destination node). The MDS failed.
```
root@stor-04:~# df -h
Filesystem Size Used Avail Use% Mounted on
udev 126G 0 126G 0% /dev
tmpfs 26G 2.6G 23G 11% /run
/dev/md1 20G 19G 0 100% /
tmpfs 126G 236K 126G 1% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 126G 0 126G 0% /sys/fs/cgroup
/dev/md1 20G 19G 0 100% /home
```
```
2017-06-22 11:13:24 251318 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e38eb - info - execute: Arguments for migrateVolume are {[force:false,volume_id:c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:24 251384 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e38ec - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:24 251404 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e38ed - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:24 251449 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VFSObjectRouter - 00000000003e38ee - info - migrate: Migrating c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:24 251784 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VFSObjectRouter - 00000000003e38ef - info - migrate_: Trying to migrate c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d from data01PAwWiMYt0g9OwJZl, only steal if offline OnlyStealFromOfflineNode::T
2017-06-22 11:13:25 280926 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VolManager - 00000000003e38f0 - notice - Backend Restart, Namespace: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d, owner tag 106, START
2017-06-22 11:13:25 280997 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f1 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:25 281118 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f2 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:25 281136 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f3 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:25 281348 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38f4 - info - Logger: Entering invalidate_cache c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 281365 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e38f5 - info - invalidate_cache_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: requesting cache invalidation
2017-06-22 11:13:25 281440 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38f6 - info - ~Logger: Exiting invalidate_cache for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 281474 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f7 - info - logMessage: TCP_transport(10.109.3.44, 26229)
2017-06-22 11:13:25 281545 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f8 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:25 281560 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e38f9 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:25 281753 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38fa - info - Logger: Entering invalidate_cache c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 281767 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e38fb - info - invalidate_cache_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: requesting cache invalidation
2017-06-22 11:13:25 285147 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38fc - info - ~Logger: Exiting invalidate_cache for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 286064 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38fd - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:25 323160 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e38fe - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:25 327447 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e38ff - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:25 327483 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3900 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:25 327514 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3901 - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:25 327615 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3902 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:25 327746 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3903 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:25 327767 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3904 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:25 327902 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3905 - warning - do_wrap_: Retrying with new connection (retry: 1, sleep before retry: 1000 milliseconds)
2017-06-22 11:13:26 328053 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3906 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:26 372696 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3907 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:26 385614 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3908 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:26 385684 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3909 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:26 385721 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e390a - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:26 385972 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e390b - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:26 386143 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e390c - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:26 386172 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e390d - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:26 386349 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e390e - warning - do_wrap_: Retrying with new connection (retry: 2, sleep before retry: 2000 milliseconds)
2017-06-22 11:13:27 334223 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e390f - info - work: You have 2 connections running
2017-06-22 11:13:27 334294 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e3910 - info - cleanup_threads_locked: Trying to join 2 threads
2017-06-22 11:13:27 334382 +0200 - stor-04.be-g8-4 - 29131/0x00007fab857f9700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e3911 - info - execute: Arguments for volumeInfo are {[redirect_fenced:true,volume_id:c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a]}
2017-06-22 11:13:27 520215 +0200 - stor-04.be-g8-4 - 29131/0x00007fab7a7fb700 - volumedriverfs/VFSRemoteNode - 00000000003e3912 - info - operator(): got keepalive rsp from data01OpKZbWWujImRDMBt
2017-06-22 11:13:27 520232 +0200 - stor-04.be-g8-4 - 29131/0x00007fab79ffa700 - volumedriverfs/VFSRemoteNode - 00000000003e3913 - info - operator(): got keepalive rsp from data01PAwWiMYt0g9OwJZl
2017-06-22 11:13:27 520242 +0200 - stor-04.be-g8-4 - 29131/0x00007fab7affc700 - volumedriverfs/VFSRemoteNode - 00000000003e3914 - info - operator(): got keepalive rsp from data012BSgz3tz2dJvdW1e
2017-06-22 11:13:28 386486 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3915 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:28 412787 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3916 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:28 415098 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3917 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:28 415138 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3918 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:28 415170 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3919 - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:28 415275 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e391a - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:28 415401 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e391b - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:28 415428 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e391c - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:28 415673 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e391d - warning - do_wrap_: Retrying with new connection (retry: 3, sleep before retry: 4000 milliseconds)
2017-06-22 11:13:29 340281 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e391e - info - work: You have 3 connections running
2017-06-22 11:13:29 340349 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e391f - info - cleanup_threads_locked: Trying to join 3 threads
2017-06-22 11:13:29 340427 +0200 - stor-04.be-g8-4 - 29131/0x00007fabb97fa700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e3920 - info - execute: Arguments for volumePerformanceCountersV3 are {[reset:false,volume_id:c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a]}
2017-06-22 11:13:31 605394 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3921 - info - Logger: Entering read fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a snapshots.xml
2017-06-22 11:13:31 625614 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/AlbaConnection - 00000000003e3922 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:31 625926 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3923 - error - ~Logger: Exiting read for fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a snapshots.xml with exception
2017-06-22 11:13:31 625953 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3924 - error - ~Logger: Exiting read for fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a
2017-06-22 11:13:31 625974 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendInterface - 00000000003e3925 - error - do_wrap_: Problem with connection 0x7fab14024a40: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:31 625989 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendInterface - 00000000003e3926 - error - do_wrap_: Giving up connection 0x7fab14024a40
2017-06-22 11:13:31 626121 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/BackendInterface - 00000000003e3927 - error - fillObject: Problem getting snapshots.xml from fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:31 626513 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/MDSMetaDataBackend - 00000000003e3928 - info - ~MDSMetaDataBackend: fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a: used clusters: 0
2017-06-22 11:13:31 626599 +0200 - stor-04.be-g8-4 - 29131/0x00007fac53fff700 - volumedriverfs/PeriodicActionPoolTaskImpl - 00000000003e3929 - error - run: mds-poll-namespace-fa1b9e20-32e6-4d62-a4cf-93f9a32f7a6a: caught exception: (Unix.Unix_error "No space left on device" write "") - ignored
2017-06-22 11:13:32 371894 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e392a - info - work: You have 4 connections running
2017-06-22 11:13:32 371943 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e392b - info - cleanup_threads_locked: Trying to join 4 threads
2017-06-22 11:13:32 372096 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e392c - info - execute: Arguments for ListClientConnections are {[vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:32 372170 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e392d - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:32 372201 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e392e - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:32 372218 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e392f - info - execute: Call ListClientConnections took 5.7e-05 seconds
2017-06-22 11:13:32 372276 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCConnection - 00000000003e3930 - info - operator(): Closing socket 102
2017-06-22 11:13:32 415845 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3931 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:32 455585 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3932 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:32 455871 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3933 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:32 455901 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3934 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:32 455925 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3935 - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:32 456071 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3936 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:32 456206 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3937 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:32 456234 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3938 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:32 456368 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3939 - warning - do_wrap_: Retrying with new connection (retry: 4, sleep before retry: 8000 milliseconds)
2017-06-22 11:13:36 769956 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e393a - info - work: You have 5 connections running
2017-06-22 11:13:36 770018 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e393b - info - cleanup_threads_locked: Trying to join 5 threads
2017-06-22 11:13:36 770078 +0200 - stor-04.be-g8-4 - 29131/0x00007fab667f8700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e393c - info - execute: Arguments for volumeDriverPerformanceCountersV3 are {[reset:false,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:36 770164 +0200 - stor-04.be-g8-4 - 29131/0x00007fab667f8700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e393d - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:36 770189 +0200 - stor-04.be-g8-4 - 29131/0x00007fab667f8700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e393e - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:40 456516 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e393f - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:40 488395 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3940 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:40 488615 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3941 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:40 488636 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3942 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:40 488653 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3943 - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:40 488758 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3944 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:40 488863 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3945 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:40 488882 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaProxyClient - 00000000003e3946 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:40 489001 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e3947 - warning - do_wrap_: Retrying with new connection (retry: 5, sleep before retry: 16000 milliseconds)
2017-06-22 11:13:42 520385 +0200 - stor-04.be-g8-4 - 29131/0x00007fab7affc700 - volumedriverfs/VFSRemoteNode - 00000000003e3948 - info - operator(): got keepalive rsp from data012BSgz3tz2dJvdW1e
2017-06-22 11:13:42 520419 +0200 - stor-04.be-g8-4 - 29131/0x00007fab7a7fb700 - volumedriverfs/VFSRemoteNode - 00000000003e3949 - info - operator(): got keepalive rsp from data01OpKZbWWujImRDMBt
2017-06-22 11:13:42 520415 +0200 - stor-04.be-g8-4 - 29131/0x00007fab79ffa700 - volumedriverfs/VFSRemoteNode - 00000000003e3949 - info - operator(): got keepalive rsp from data01PAwWiMYt0g9OwJZl
2017-06-22 11:13:46 290833 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e394a - info - work: You have 5 connections running
2017-06-22 11:13:46 290901 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e394b - info - cleanup_threads_locked: Trying to join 5 threads
2017-06-22 11:13:46 291026 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e394c - info - execute: Arguments for volumeDriverPerformanceCountersV3 are {[reset:false,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:46 291092 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e394d - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:46 291111 +0200 - stor-04.be-g8-4 - 29131/0x00007fab677fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e394e - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:50 811098 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MetaDataServerTable - 00000000003e394f - info - work_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: running periodic action
2017-06-22 11:13:50 811174 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MDSMetaDataBackend - 00000000003e3950 - info - MDSMetaDataBackend: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:50 811332 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MDSMetaDataBackend - 00000000003e3951 - info - lastCorkUUID: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: 3e8a6914-e4f4-4d2e-9343-c1aef081a0d8
2017-06-22 11:13:50 811395 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MDSMetaDataBackend - 00000000003e3952 - info - scrub_id: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: scrub ID 28ee90a9-374b-49e8-85aa-07bf303aad5f
2017-06-22 11:13:50 811430 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/CachedMetaDataStore - 00000000003e3953 - info - init_pages_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: page capacity (entries): 64, max cached pages: 256
2017-06-22 11:13:50 811537 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/MetaDataStoreBuilder - 00000000003e3954 - info - update_metadata_store_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: bringing MetaDataStore in sync with backend, requested interval ( 3e8a6914-e4f4-4d2e-9343-c1aef081a0d8, --], check scrub ID: CheckScrubId::F, dry run:DryRun::F, full rebuild: false
2017-06-22 11:13:50 815978 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3955 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:50 816185 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3956 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:50 816216 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3957 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:50 816448 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3958 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml
2017-06-22 11:13:50 842623 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaConnection - 00000000003e3959 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:50 848631 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e395a - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml with exception
2017-06-22 11:13:50 848676 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e395b - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:50 848706 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e395c - error - do_wrap_: Problem with connection 0x7fab18007ae0: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:50 848855 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e395d - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:50 849003 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e395e - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:50 849031 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e395f - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:50 849166 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e3960 - warning - do_wrap_: Retrying with new connection (retry: 1, sleep before retry: 1000 milliseconds)
2017-06-22 11:13:51 849290 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3961 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml
2017-06-22 11:13:51 874257 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaConnection - 00000000003e3962 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:51 874479 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3963 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml with exception
2017-06-22 11:13:51 874503 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3964 - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:51 874525 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e3965 - error - do_wrap_: Problem with connection 0x7fab18007ae0: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:51 874685 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3966 - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:51 874808 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3967 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:51 874833 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3968 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:51 874974 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e3969 - warning - do_wrap_: Retrying with new connection (retry: 2, sleep before retry: 2000 milliseconds)
2017-06-22 11:13:53 875101 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e396a - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml
2017-06-22 11:13:53 898792 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaConnection - 00000000003e396b - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:53 899042 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e396c - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d snapshots.xml with exception
2017-06-22 11:13:53 899066 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e396d - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:53 899088 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e396e - error - do_wrap_: Problem with connection 0x7fab18007ae0: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:53 899212 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e396f - info - logMessage: TCP_transport(10.109.3.44, 26228)
2017-06-22 11:13:53 899330 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3970 - info - logMessage: make_proxy_client( rora_config=RoraConfig{ manifest_cache_size= 17179869184, asd_connection_pool_size= 5 } )
2017-06-22 11:13:53 899371 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/AlbaProxyClient - 00000000003e3971 - info - logMessage: RoraProxy_client( _asd_connection_pool_size = 5 ...)
2017-06-22 11:13:53 899499 +0200 - stor-04.be-g8-4 - 29131/0x00007fac60ff9700 - volumedriverfs/BackendInterface - 00000000003e3972 - warning - do_wrap_: Retrying with new connection (retry: 3, sleep before retry: 4000 milliseconds)
2017-06-22 11:13:56 270787 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e3973 - info - work: You have 6 connections running
2017-06-22 11:13:56 270850 +0200 - stor-04.be-g8-4 - 29131/0x00007faba27fc700 - volumedriverfs/MTServer - 00000000003e3974 - info - cleanup_threads_locked: Trying to join 6 threads
2017-06-22 11:13:56 270976 +0200 - stor-04.be-g8-4 - 29131/0x00007fab65ff7700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e3975 - info - execute: Arguments for volumeDriverPerformanceCountersV3 are {[reset:false,vrouter_cluster_id:215304e1-f23b-4a2d-aba9-67937960a17a,vrouter_id:data01OdDMLuDUBCNEneoS]}
2017-06-22 11:13:56 271041 +0200 - stor-04.be-g8-4 - 29131/0x00007fab65ff7700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e3976 - info - execute: Execution on data01OdDMLuDUBCNEneoS requested
2017-06-22 11:13:56 271063 +0200 - stor-04.be-g8-4 - 29131/0x00007fab65ff7700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e3977 - info - execute: Node ID data01OdDMLuDUBCNEneoS is ours - good
2017-06-22 11:13:56 489152 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e3978 - info - Logger: Entering read c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration
2017-06-22 11:13:56 527885 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/AlbaConnection - 00000000003e3979 - error - convert_exceptions_: read object: caught Alba proxy exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528152 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e397a - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d volume_configuration with exception
2017-06-22 11:13:56 528179 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendConnectionInterfaceLogger - 00000000003e397b - error - ~Logger: Exiting read for c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d
2017-06-22 11:13:56 528219 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e397c - error - do_wrap_: Problem with connection 0x7faa0d9a5040: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528234 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e397d - error - do_wrap_: Giving up connection 0x7faa0d9a5040
2017-06-22 11:13:56 528405 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/BackendInterface - 00000000003e397e - error - fillObject: Problem getting volume_configuration from c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528466 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VolManager - 00000000003e397f - notice - Backend Restart, Namespace: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d, owner tag 106, EXCEPTION
2017-06-22 11:13:56 528495 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/VFSLocalNode - 00000000003e3980 - error - restart_volume_from_backend_: c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d: failed to restart volume: : (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528604 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCTimingWrapper - 00000000003e3981 - error - execute: migrateVolume Caught std::exception: (Unix.Unix_error "No space left on device" write "")
2017-06-22 11:13:56 528607 +0200 - stor-04.be-g8-4 - 29131/0x00007fab857f9700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e3982 - info - execute: Object c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d is not present on this node - figuring out if it lives elsewhere.
2017-06-22 11:13:56 528715 +0200 - stor-04.be-g8-4 - 29131/0x00007fabb97fa700 - volumedriverfs/XMLRPCRedirectWrapper - 00000000003e3983 - info - execute: Object c9c0b837-a21e-4ff3-b2d6-7a5bd79e596d is not present on this node - figuring out if it lives elsewhere.
2017-06-22 11:13:56 528727 +0200 - stor-04.be-g8-4 - 29131/0x00007fab797f9700 - volumedriverfs/XMLRPCConnection - 00000000003e3984 - info - operator(): Closing socket 139
```
Volumedriver config:
```
root@stor-04:~# ovs config get ovs/vpools/215304e1-f23b-4a2d-aba9-67937960a17a/hosts/data01OdDMLuDUBCNEneoS/config
{
"content_addressed_cache": {
"read_cache_serialization_path": [],
"serialize_read_cache": false
},
"backend_connection_manager": {
"backend_interface_retries_on_error": 5,
"backend_type": "MULTI",
"1": {
"alba_connection_preset": "gigPreset",
"backend_type": "ALBA",
"alba_connection_transport": "TCP",
"alba_connection_timeout": 15,
"alba_connection_use_rora": true,
"alba_connection_host": "10.109.3.44",
"alba_connection_rora_manifest_cache_capacity": 17179869184,
"alba_connection_port": 26229
},
"0": {
"alba_connection_preset": "gigPreset",
"backend_type": "ALBA",
"alba_connection_transport": "TCP",
"alba_connection_timeout": 15,
"alba_connection_use_rora": true,
"alba_connection_host": "10.109.3.44",
"alba_connection_rora_manifest_cache_capacity": 17179869184,
"alba_connection_port": 26228
},
"backend_interface_retry_backoff_multiplier": 2.0,
"backend_interface_retry_interval_secs": 1
},
"distributed_lock_store": {
"dls_arakoon_cluster_nodes": [
{
"host": "10.109.3.42",
"node_id": "2BSgz3tz2dJvdW1e",
"port": 26416
},
{
"host": "10.109.3.41",
"node_id": "OpKZbWWujImRDMBt",
"port": 26416
},
{
"host": "10.109.3.43",
"node_id": "PAwWiMYt0g9OwJZl",
"port": 26416
}
],
"dls_arakoon_cluster_id": "voldrv",
"dls_type": "Arakoon"
},
"volume_registry": {
"vregistry_arakoon_cluster_id": "voldrv",
"vregistry_arakoon_cluster_nodes": [
{
"host": "10.109.3.42",
"node_id": "2BSgz3tz2dJvdW1e",
"port": 26416
},
{
"host": "10.109.3.41",
"node_id": "OpKZbWWujImRDMBt",
"port": 26416
},
{
"host": "10.109.3.43",
"node_id": "PAwWiMYt0g9OwJZl",
"port": 26416
}
]
},
"volume_router_cluster": {
"vrouter_cluster_id": "215304e1-f23b-4a2d-aba9-67937960a17a"
},
"volume_router": {
"vrouter_migrate_timeout_ms": 60000,
"vrouter_sco_multiplier": 16384,
"vrouter_keepalive_interval_secs": "5",
"vrouter_volume_read_threshold": 0,
"vrouter_id": "data01OdDMLuDUBCNEneoS",
"vrouter_use_fencing": true,
"vrouter_keepalive_time_secs": "15",
"vrouter_file_write_threshold": 0,
"vrouter_volume_write_threshold": 0,
"vrouter_backend_sync_timeout_ms": 60000,
"vrouter_min_workers": 4,
"vrouter_keepalive_retries": "2",
"vrouter_routing_retries": 10,
"vrouter_redirect_timeout_ms": "120000",
"vrouter_file_read_threshold": 0,
"vrouter_max_workers": 16
},
"volume_manager": {
"clean_interval": 1,
"tlog_path": "/mnt/ssd3/data01_db_tlog_1",
"non_disposable_scos_factor": 2.0,
"metadata_path": "/mnt/ssd3/data01_db_md_1",
"number_of_scos_in_tlog": 1,
"default_cluster_size": 4096,
"dtl_throttle_usecs": 4000
},
"scocache": {
"trigger_gap": "8GiB",
"backoff_gap": "10GiB",
"scocache_mount_points": [
{
"path": "/mnt/ssd1/data01_write_sco_1",
"size": "100188160KiB"
},
{
"path": "/mnt/ssd4/data01_write_sco_1",
"size": "100188160KiB"
}
]
},
"filesystem": {
"fs_raw_disk_suffix": ".raw",
"fs_enable_shm_interface": 0,
"fs_metadata_backend_arakoon_cluster_nodes": [],
"fs_dtl_mode": "Synchronous",
"fs_metadata_backend_mds_nodes": [
{
"host": "10.109.3.44",
"port": 26304
}
],
"fs_metadata_backend_type": "MDS",
"fs_virtual_disk_format": "raw",
"fs_enable_network_interface": 1,
"fs_dtl_config_mode": "Automatic",
"fs_dtl_host": "",
"fs_file_event_rules": [
{
"fs_file_event_rule_calls": [
"Rename"
],
"fs_file_event_rule_path_regex": ".*"
}
]
},
"file_driver": {
"fd_extent_cache_capacity": "1024",
"fd_namespace": "fd-data01-215304e1-f23b-4a2d-aba9-67937960a17a",
"fd_cache_path": "/mnt/ssd1/data01_write_fd_1"
},
"threadpool_component": {
"num_threads": 16
},
"event_publisher": {
"events_amqp_routing_key": "volumerouter",
"events_amqp_uris": [
{
"amqp_uri": "amqp://ovs:0penv5tor4ge@10.109.3.41:5672"
},
{
"amqp_uri": "amqp://ovs:0penv5tor4ge@10.109.3.42:5672"
},
{
"amqp_uri": "amqp://ovs:0penv5tor4ge@10.109.3.43:5672"
}
]
},
"metadata_server": {
"mds_nodes": [
{
"host": "10.109.3.44",
"scratch_directory": "/mnt/ssd3/data01_db_mds_1/scratch",
"port": 26304,
"db_directory": "/mnt/ssd3/data01_db_mds_1/db"
}
]
},
"distributed_transaction_log": {
"dtl_path": "/mnt/ssd5/data01_dtl_1",
"dtl_transport": "TCP"
},
"network_interface": {
"network_max_neighbour_distance": 9999
}
}
```
|
process
|
observation halted volumes after volume move to node with full root partition the move from volumes failed due to a full root partition on the destination node all of the volumes got in halted state on the destination node the mds failed root stor df h filesystem size used avail use mounted on udev dev tmpfs run dev tmpfs dev shm tmpfs run lock tmpfs sys fs cgroup dev home stor be volumedriverfs xmlrpctimingwrapper info execute arguments for migratevolume are stor be volumedriverfs xmlrpcredirectwrapper info execute execution on requested stor be volumedriverfs xmlrpcredirectwrapper info execute node id is ours good stor be volumedriverfs vfsobjectrouter info migrate migrating stor be volumedriverfs vfsobjectrouter info migrate trying to migrate from only steal if offline onlystealfromofflinenode t stor be volumedriverfs volmanager notice backend restart namespace owner tag start stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendconnectioninterfacelogger info logger entering invalidate cache stor be volumedriverfs albaconnection info invalidate cache requesting cache invalidation stor be volumedriverfs backendconnectioninterfacelogger info logger exiting invalidate cache for stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendconnectioninterfacelogger info logger entering invalidate cache stor be volumedriverfs albaconnection info invalidate cache requesting cache invalidation stor be volumedriverfs backendconnectioninterfacelogger info logger exiting invalidate cache for stor be volumedriverfs backendconnectioninterfacelogger info logger entering read volume configuration stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for volume configuration with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendinterface warning do wrap retrying with new connection retry sleep before retry milliseconds stor be volumedriverfs backendconnectioninterfacelogger info logger entering read volume configuration stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for volume configuration with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendinterface warning do wrap retrying with new connection retry sleep before retry milliseconds stor be volumedriverfs mtserver info work you have connections running stor be volumedriverfs mtserver info cleanup threads locked trying to join threads stor be volumedriverfs xmlrpctimingwrapper info execute arguments for volumeinfo are stor be volumedriverfs vfsremotenode info operator got keepalive rsp from stor be volumedriverfs vfsremotenode info operator got keepalive rsp from stor be volumedriverfs vfsremotenode info operator got keepalive rsp from stor be volumedriverfs backendconnectioninterfacelogger info logger entering read volume configuration stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for volume configuration with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendinterface warning do wrap retrying with new connection retry sleep before retry milliseconds stor be volumedriverfs mtserver info work you have connections running stor be volumedriverfs mtserver info cleanup threads locked trying to join threads stor be volumedriverfs xmlrpctimingwrapper info execute arguments for are stor be volumedriverfs backendconnectioninterfacelogger info logger entering read snapshots xml stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for snapshots xml with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs backendinterface error do wrap giving up connection stor be volumedriverfs backendinterface error fillobject problem getting snapshots xml from unix unix error no space left on device write stor be volumedriverfs mdsmetadatabackend info mdsmetadatabackend used clusters stor be volumedriverfs periodicactionpooltaskimpl error run mds poll namespace caught exception unix unix error no space left on device write ignored stor be volumedriverfs mtserver info work you have connections running stor be volumedriverfs mtserver info cleanup threads locked trying to join threads stor be volumedriverfs xmlrpctimingwrapper info execute arguments for listclientconnections are stor be volumedriverfs xmlrpcredirectwrapper info execute execution on requested stor be volumedriverfs xmlrpcredirectwrapper info execute node id is ours good stor be volumedriverfs xmlrpctimingwrapper info execute call listclientconnections took seconds stor be volumedriverfs xmlrpcconnection info operator closing socket stor be volumedriverfs backendconnectioninterfacelogger info logger entering read volume configuration stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for volume configuration with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendinterface warning do wrap retrying with new connection retry sleep before retry milliseconds stor be volumedriverfs mtserver info work you have connections running stor be volumedriverfs mtserver info cleanup threads locked trying to join threads stor be volumedriverfs xmlrpctimingwrapper info execute arguments for are stor be volumedriverfs xmlrpcredirectwrapper info execute execution on requested stor be volumedriverfs xmlrpcredirectwrapper info execute node id is ours good stor be volumedriverfs backendconnectioninterfacelogger info logger entering read volume configuration stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for volume configuration with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendinterface warning do wrap retrying with new connection retry sleep before retry milliseconds stor be volumedriverfs vfsremotenode info operator got keepalive rsp from stor be volumedriverfs vfsremotenode info operator got keepalive rsp from stor be volumedriverfs vfsremotenode info operator got keepalive rsp from stor be volumedriverfs mtserver info work you have connections running stor be volumedriverfs mtserver info cleanup threads locked trying to join threads stor be volumedriverfs xmlrpctimingwrapper info execute arguments for are stor be volumedriverfs xmlrpcredirectwrapper info execute execution on requested stor be volumedriverfs xmlrpcredirectwrapper info execute node id is ours good stor be volumedriverfs metadataservertable info work running periodic action stor be volumedriverfs mdsmetadatabackend info mdsmetadatabackend stor be volumedriverfs mdsmetadatabackend info lastcorkuuid stor be volumedriverfs mdsmetadatabackend info scrub id scrub id stor be volumedriverfs cachedmetadatastore info init pages page capacity entries max cached pages stor be volumedriverfs metadatastorebuilder info update metadata store bringing metadatastore in sync with backend requested interval check scrub id checkscrubid f dry run dryrun f full rebuild false stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendconnectioninterfacelogger info logger entering read snapshots xml stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for snapshots xml with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendinterface warning do wrap retrying with new connection retry sleep before retry milliseconds stor be volumedriverfs backendconnectioninterfacelogger info logger entering read snapshots xml stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for snapshots xml with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendinterface warning do wrap retrying with new connection retry sleep before retry milliseconds stor be volumedriverfs backendconnectioninterfacelogger info logger entering read snapshots xml stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for snapshots xml with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs albaproxyclient info logmessage tcp transport stor be volumedriverfs albaproxyclient info logmessage make proxy client rora config roraconfig manifest cache size asd connection pool size stor be volumedriverfs albaproxyclient info logmessage roraproxy client asd connection pool size stor be volumedriverfs backendinterface warning do wrap retrying with new connection retry sleep before retry milliseconds stor be volumedriverfs mtserver info work you have connections running stor be volumedriverfs mtserver info cleanup threads locked trying to join threads stor be volumedriverfs xmlrpctimingwrapper info execute arguments for are stor be volumedriverfs xmlrpcredirectwrapper info execute execution on requested stor be volumedriverfs xmlrpcredirectwrapper info execute node id is ours good stor be volumedriverfs backendconnectioninterfacelogger info logger entering read volume configuration stor be volumedriverfs albaconnection error convert exceptions read object caught alba proxy exception unix unix error no space left on device write stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for volume configuration with exception stor be volumedriverfs backendconnectioninterfacelogger error logger exiting read for stor be volumedriverfs backendinterface error do wrap problem with connection unix unix error no space left on device write stor be volumedriverfs backendinterface error do wrap giving up connection stor be volumedriverfs backendinterface error fillobject problem getting volume configuration from unix unix error no space left on device write stor be volumedriverfs volmanager notice backend restart namespace owner tag exception stor be volumedriverfs vfslocalnode error restart volume from backend failed to restart volume unix unix error no space left on device write stor be volumedriverfs xmlrpctimingwrapper error execute migratevolume caught std exception unix unix error no space left on device write stor be volumedriverfs xmlrpcredirectwrapper info execute object is not present on this node figuring out if it lives elsewhere stor be volumedriverfs xmlrpcredirectwrapper info execute object is not present on this node figuring out if it lives elsewhere stor be volumedriverfs xmlrpcconnection info operator closing socket volumedriver config root stor ovs config get ovs vpools hosts config content addressed cache read cache serialization path serialize read cache false backend connection manager backend interface retries on error backend type multi alba connection preset gigpreset backend type alba alba connection transport tcp alba connection timeout alba connection use rora true alba connection host alba connection rora manifest cache capacity alba connection port alba connection preset gigpreset backend type alba alba connection transport tcp alba connection timeout alba connection use rora true alba connection host alba connection rora manifest cache capacity alba connection port backend interface retry backoff multiplier backend interface retry interval secs distributed lock store dls arakoon cluster nodes host node id port host node id opkzbwwujimrdmbt port host node id port dls arakoon cluster id voldrv dls type arakoon volume registry vregistry arakoon cluster id voldrv vregistry arakoon cluster nodes host node id port host node id opkzbwwujimrdmbt port host node id port volume router cluster vrouter cluster id volume router vrouter migrate timeout ms vrouter sco multiplier vrouter keepalive interval secs vrouter volume read threshold vrouter id vrouter use fencing true vrouter keepalive time secs vrouter file write threshold vrouter volume write threshold vrouter backend sync timeout ms vrouter min workers vrouter keepalive retries vrouter routing retries vrouter redirect timeout ms vrouter file read threshold vrouter max workers volume manager clean interval tlog path mnt db tlog non disposable scos factor metadata path mnt db md number of scos in tlog default cluster size dtl throttle usecs scocache trigger gap backoff gap scocache mount points path mnt write sco size path mnt write sco size filesystem fs raw disk suffix raw fs enable shm interface fs metadata backend arakoon cluster nodes fs dtl mode synchronous fs metadata backend mds nodes host port fs metadata backend type mds fs virtual disk format raw fs enable network interface fs dtl config mode automatic fs dtl host fs file event rules fs file event rule calls rename fs file event rule path regex file driver fd extent cache capacity fd namespace fd fd cache path mnt write fd threadpool component num threads event publisher events amqp routing key volumerouter events amqp uris amqp uri amqp ovs amqp uri amqp ovs amqp uri amqp ovs metadata server mds nodes host scratch directory mnt db mds scratch port db directory mnt db mds db distributed transaction log dtl path mnt dtl dtl transport tcp network interface network max neighbour distance
| 1
|
21,963
| 6,227,689,392
|
IssuesEvent
|
2017-07-10 21:19:07
|
XceedBoucherS/TestImport5
|
https://api.github.com/repos/XceedBoucherS/TestImport5
|
closed
|
TimePicker is not fully culture aware
|
CodePlex
|
<b>tobias[CodePlex]</b> <br />Binding to a DateTime handles the DateTime in a incorrect way.
nbsp
I have attached a sample project where the problem is visible (I don't know if this problem can be replicated on a English Windows system):
Sample DateTime: 3. June 2011 (the problem occurs only, if the day is lower inclusive 12)
I have two UserControls: On the left the quotcorrectquot DateTime is displayed, on the right the Date is quotmessed upquot and is now the 6. March 2011.
The difference between the two UserControls is that on the right I use the TimePicker and bind its Value property to my ViewModels DateTime.
|
1.0
|
TimePicker is not fully culture aware - <b>tobias[CodePlex]</b> <br />Binding to a DateTime handles the DateTime in a incorrect way.
nbsp
I have attached a sample project where the problem is visible (I don't know if this problem can be replicated on a English Windows system):
Sample DateTime: 3. June 2011 (the problem occurs only, if the day is lower inclusive 12)
I have two UserControls: On the left the quotcorrectquot DateTime is displayed, on the right the Date is quotmessed upquot and is now the 6. March 2011.
The difference between the two UserControls is that on the right I use the TimePicker and bind its Value property to my ViewModels DateTime.
|
non_process
|
timepicker is not fully culture aware tobias binding to a datetime handles the datetime in a incorrect way nbsp i have attached a sample project where the problem is visible i don t know if this problem can be replicated on a english windows system sample datetime june the problem occurs only if the day is lower inclusive i have two usercontrols on the left the quotcorrectquot datetime is displayed on the right the date is quotmessed upquot and is now the march the difference between the two usercontrols is that on the right i use the timepicker and bind its value property to my viewmodels datetime
| 0
|
40,038
| 5,267,113,556
|
IssuesEvent
|
2017-02-04 19:20:59
|
paperjs/paper.js
|
https://api.github.com/repos/paperjs/paper.js
|
closed
|
Optimize boolean operations when there are no crossings.
|
cat: boolean-operations status: needs-tests type: improvemnet
|
When there are no crossings, the result can already be known ahead of `tracePaths()`, probably leading to a massive speed-up:
- intersect: return `null`
- unite: return a compound path with both operands
- subtract: return the first operand
- exclude: same as unite.
- divide: no change needed, since it redirects to the other methods.
|
1.0
|
Optimize boolean operations when there are no crossings. - When there are no crossings, the result can already be known ahead of `tracePaths()`, probably leading to a massive speed-up:
- intersect: return `null`
- unite: return a compound path with both operands
- subtract: return the first operand
- exclude: same as unite.
- divide: no change needed, since it redirects to the other methods.
|
non_process
|
optimize boolean operations when there are no crossings when there are no crossings the result can already be known ahead of tracepaths probably leading to a massive speed up intersect return null unite return a compound path with both operands subtract return the first operand exclude same as unite divide no change needed since it redirects to the other methods
| 0
|
22,718
| 32,040,067,371
|
IssuesEvent
|
2023-09-22 18:31:22
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
procpath 1.8.1 has 2 GuardDog issues
|
guarddog exec-base64 silent-process-execution
|
https://pypi.org/project/procpath
https://inspector.pypi.io/project/procpath
```{
"dependency": "procpath",
"version": "1.8.1",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "Procpath-1.8.1/procpath/utility.py:27",
"code": " env = subprocess.check_output('\\n'.join(script), shell=True, encoding='utf-8')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "Procpath-1.8.1/procpath/test.py:2386",
"code": " p = subprocess.Popen(\n ['timeout', '0.25', 'tail', '---disable-inotify', '-f', f'{f.name}', f.name],\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subp... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpfi5m3q_f/procpath"
}
}```
|
1.0
|
procpath 1.8.1 has 2 GuardDog issues - https://pypi.org/project/procpath
https://inspector.pypi.io/project/procpath
```{
"dependency": "procpath",
"version": "1.8.1",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "Procpath-1.8.1/procpath/utility.py:27",
"code": " env = subprocess.check_output('\\n'.join(script), shell=True, encoding='utf-8')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "Procpath-1.8.1/procpath/test.py:2386",
"code": " p = subprocess.Popen(\n ['timeout', '0.25', 'tail', '---disable-inotify', '-f', f'{f.name}', f.name],\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subp... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpfi5m3q_f/procpath"
}
}```
|
process
|
procpath has guarddog issues dependency procpath version result issues errors results exec location procpath procpath utility py code env subprocess check output n join script shell true encoding utf message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n silent process execution location procpath procpath test py code p subprocess popen n n stdin subprocess devnull n stdout subprocess devnull n stderr subp message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp f procpath
| 1
|
4,055
| 6,988,621,710
|
IssuesEvent
|
2017-12-14 13:38:11
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
process._tickDomainCallback is not a function using 9.3.0
|
process
|
* **Version**: 9.3.0
* **Platform**: Windows 10 64 bit
* **Subsystem**: process
<!-- Enter your issue details below this comment. -->
Using 9.3.0 only I'm seeing `process._tickDomainCallback is not a function`, (we're using [deasync](https://github.com/abbr/deasync/blob/60fbfa97f26a79ccf918d1d3188ac3a7e0d685ff/index.js#L65) which uses that function)
I don't see this issue with 9.2.1, or 8.9.3. Easily reproducible with `node -e "process._tickDomainCallback()"`, this function doesn't seem to be documented and the underscore I presume means it's more internal? So maybe this is intentional?
|
1.0
|
process._tickDomainCallback is not a function using 9.3.0 - * **Version**: 9.3.0
* **Platform**: Windows 10 64 bit
* **Subsystem**: process
<!-- Enter your issue details below this comment. -->
Using 9.3.0 only I'm seeing `process._tickDomainCallback is not a function`, (we're using [deasync](https://github.com/abbr/deasync/blob/60fbfa97f26a79ccf918d1d3188ac3a7e0d685ff/index.js#L65) which uses that function)
I don't see this issue with 9.2.1, or 8.9.3. Easily reproducible with `node -e "process._tickDomainCallback()"`, this function doesn't seem to be documented and the underscore I presume means it's more internal? So maybe this is intentional?
|
process
|
process tickdomaincallback is not a function using version platform windows bit subsystem process using only i m seeing process tickdomaincallback is not a function we re using which uses that function i don t see this issue with or easily reproducible with node e process tickdomaincallback this function doesn t seem to be documented and the underscore i presume means it s more internal so maybe this is intentional
| 1
|
7,594
| 10,704,766,978
|
IssuesEvent
|
2019-10-24 12:27:52
|
googleapis/google-cloud-cpp
|
https://api.github.com/repos/googleapis/google-cloud-cpp
|
closed
|
opensuse build fails due to gRPC.
|
type: process
|
The OpenSuse build started failing due to a duplicate definition of gettid() in gRPC: https://github.com/grpc/grpc/issues/19964
However, we weren't able to upgrade to a version of gRPC that included the fix due to a "make install" problem: https://github.com/grpc/grpc/issues/20392
That issue is now resolved so we should be able to upgrade gRPC and re-enable the OpenSuse build. Carlos is trying this in common in https://github.com/googleapis/google-cloud-cpp-common/pull/75 - if that is ok we can apply it here
|
1.0
|
opensuse build fails due to gRPC. - The OpenSuse build started failing due to a duplicate definition of gettid() in gRPC: https://github.com/grpc/grpc/issues/19964
However, we weren't able to upgrade to a version of gRPC that included the fix due to a "make install" problem: https://github.com/grpc/grpc/issues/20392
That issue is now resolved so we should be able to upgrade gRPC and re-enable the OpenSuse build. Carlos is trying this in common in https://github.com/googleapis/google-cloud-cpp-common/pull/75 - if that is ok we can apply it here
|
process
|
opensuse build fails due to grpc the opensuse build started failing due to a duplicate definition of gettid in grpc however we weren t able to upgrade to a version of grpc that included the fix due to a make install problem that issue is now resolved so we should be able to upgrade grpc and re enable the opensuse build carlos is trying this in common in if that is ok we can apply it here
| 1
|
7,265
| 10,421,123,639
|
IssuesEvent
|
2019-09-16 04:36:09
|
eqcorrscan/EQcorrscan
|
https://api.github.com/repos/eqcorrscan/EQcorrscan
|
closed
|
Slow processing during template-generation
|
utils.pre_processing
|
**Describe the bug**
This is a long-standing issue that I haven't fixed, @emilyws1 ran into this recently. This is mostly a note to better document this until I re-write the processing.
When processing moderate length data (particularly day-long data) the processing functions look like they hang. What is happening is that the fft used for resampling and filtering is not efficient when the length of data cannot be easily divided by prime-factors. This means that when data has 8640001 data points (one day at 100Hz + one sample) the processing is slow. Working with data that is 8640000 length is much faster.
|
1.0
|
Slow processing during template-generation - **Describe the bug**
This is a long-standing issue that I haven't fixed, @emilyws1 ran into this recently. This is mostly a note to better document this until I re-write the processing.
When processing moderate length data (particularly day-long data) the processing functions look like they hang. What is happening is that the fft used for resampling and filtering is not efficient when the length of data cannot be easily divided by prime-factors. This means that when data has 8640001 data points (one day at 100Hz + one sample) the processing is slow. Working with data that is 8640000 length is much faster.
|
process
|
slow processing during template generation describe the bug this is a long standing issue that i haven t fixed ran into this recently this is mostly a note to better document this until i re write the processing when processing moderate length data particularly day long data the processing functions look like they hang what is happening is that the fft used for resampling and filtering is not efficient when the length of data cannot be easily divided by prime factors this means that when data has data points one day at one sample the processing is slow working with data that is length is much faster
| 1
|
8,097
| 11,273,767,241
|
IssuesEvent
|
2020-01-14 17:09:31
|
googleapis/google-cloud-cpp
|
https://api.github.com/repos/googleapis/google-cloud-cpp
|
closed
|
[Flake]: Troubleshoot Kokoro + Windows failures.
|
status: blocked type: process
|
We have see a couple of builds fail with this:
```
Start 77: embedded_server_test
kokoro_log_reader: Exhausted retries waiting for bash to return an integer exit code, returning exit code 137 instead. This is not a Kokoro error! This usually means: 1) Your build was killed by something in the VM 2) Exited normally, but that the parent shell of the nohup'd build process is IO blocked and could not write out the exit_code_file in 60 seconds. In this case be sure that all processes launched by your build are properly cleaned up before exiting to free up resources and reduce IO contention, or 3) There was a BSOD, the VM rebooted, the builder has reconnected but the process is of course gone.
Found windows minidump directory c:/windows/minidump, looks like we had a BSOD. Copying minidumps to t:/src/win_minidumps. Be sure to add an artifacts regex to collect data from **/win_minidumps/*.dmp
Minidump files found:
```
Sounds like a operating system crash, whether induced by our tests or something else is hard to tell right now.
|
1.0
|
[Flake]: Troubleshoot Kokoro + Windows failures. - We have see a couple of builds fail with this:
```
Start 77: embedded_server_test
kokoro_log_reader: Exhausted retries waiting for bash to return an integer exit code, returning exit code 137 instead. This is not a Kokoro error! This usually means: 1) Your build was killed by something in the VM 2) Exited normally, but that the parent shell of the nohup'd build process is IO blocked and could not write out the exit_code_file in 60 seconds. In this case be sure that all processes launched by your build are properly cleaned up before exiting to free up resources and reduce IO contention, or 3) There was a BSOD, the VM rebooted, the builder has reconnected but the process is of course gone.
Found windows minidump directory c:/windows/minidump, looks like we had a BSOD. Copying minidumps to t:/src/win_minidumps. Be sure to add an artifacts regex to collect data from **/win_minidumps/*.dmp
Minidump files found:
```
Sounds like a operating system crash, whether induced by our tests or something else is hard to tell right now.
|
process
|
troubleshoot kokoro windows failures we have see a couple of builds fail with this start embedded server test kokoro log reader exhausted retries waiting for bash to return an integer exit code returning exit code instead this is not a kokoro error this usually means your build was killed by something in the vm exited normally but that the parent shell of the nohup d build process is io blocked and could not write out the exit code file in seconds in this case be sure that all processes launched by your build are properly cleaned up before exiting to free up resources and reduce io contention or there was a bsod the vm rebooted the builder has reconnected but the process is of course gone found windows minidump directory c windows minidump looks like we had a bsod copying minidumps to t src win minidumps be sure to add an artifacts regex to collect data from win minidumps dmp minidump files found sounds like a operating system crash whether induced by our tests or something else is hard to tell right now
| 1
|
283,150
| 24,527,489,900
|
IssuesEvent
|
2022-10-11 14:06:39
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
Gutenberg List View hides inner block plus
|
[Status] Needs More Info Needs Testing [Feature] List View
|
### Description
If you're using inner blocks and the inner blocks plus is far left, when you have the nav open it hides it. You should be able to add blocks with the nav open all the while, it's shouldn't hide any content? Attached is open and closed, you can see how the plus is missing then open, so a user wouldn't know how to add an inner block, it's just empty.


### Step-by-step reproduction instructions
This might be an ACF blocks related issue, but I would expect the same on a custom inner block.
1. Create inner block that is far left without a container.
2. Check for inner block + with both nav open and close.
### Screenshots, screen recording, code snippet
_No response_
### Environment info
_No response_
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
No
|
1.0
|
Gutenberg List View hides inner block plus - ### Description
If you're using inner blocks and the inner blocks plus is far left, when you have the nav open it hides it. You should be able to add blocks with the nav open all the while, it's shouldn't hide any content? Attached is open and closed, you can see how the plus is missing then open, so a user wouldn't know how to add an inner block, it's just empty.


### Step-by-step reproduction instructions
This might be an ACF blocks related issue, but I would expect the same on a custom inner block.
1. Create inner block that is far left without a container.
2. Check for inner block + with both nav open and close.
### Screenshots, screen recording, code snippet
_No response_
### Environment info
_No response_
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
No
|
non_process
|
gutenberg list view hides inner block plus description if you re using inner blocks and the inner blocks plus is far left when you have the nav open it hides it you should be able to add blocks with the nav open all the while it s shouldn t hide any content attached is open and closed you can see how the plus is missing then open so a user wouldn t know how to add an inner block it s just empty step by step reproduction instructions this might be an acf blocks related issue but i would expect the same on a custom inner block create inner block that is far left without a container check for inner block with both nav open and close screenshots screen recording code snippet no response environment info no response please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg no
| 0
|
37,870
| 5,146,616,884
|
IssuesEvent
|
2017-01-13 02:08:33
|
Chengifei/Phyco
|
https://api.github.com/repos/Chengifei/Phyco
|
closed
|
expression_tree: destructor call in move assignment freeze the execution
|
tested
|
Easy to fix yet reason is unclear.
|
1.0
|
expression_tree: destructor call in move assignment freeze the execution - Easy to fix yet reason is unclear.
|
non_process
|
expression tree destructor call in move assignment freeze the execution easy to fix yet reason is unclear
| 0
|
220,934
| 17,270,282,195
|
IssuesEvent
|
2021-07-22 18:50:32
|
libp2p/go-libp2p
|
https://api.github.com/repos/libp2p/go-libp2p
|
closed
|
TestMdnsDiscovery fails
|
test_failure
|
I'm not sure where to start debugging this.
Arch Linux / No {nf,ip}tables / go1.9.3
```
--- FAIL: TestMdnsDiscovery (2.07s)
mdns_test.go:51: dial attempt failed: failed to dial <peer.ID VJcuj6> (default failure)
FAIL
FAIL github.com/libp2p/go-libp2p/p2p/discovery 2.104s
```
|
1.0
|
TestMdnsDiscovery fails - I'm not sure where to start debugging this.
Arch Linux / No {nf,ip}tables / go1.9.3
```
--- FAIL: TestMdnsDiscovery (2.07s)
mdns_test.go:51: dial attempt failed: failed to dial <peer.ID VJcuj6> (default failure)
FAIL
FAIL github.com/libp2p/go-libp2p/p2p/discovery 2.104s
```
|
non_process
|
testmdnsdiscovery fails i m not sure where to start debugging this arch linux no nf ip tables fail testmdnsdiscovery mdns test go dial attempt failed failed to dial default failure fail fail github com go discovery
| 0
|
9,943
| 12,976,090,565
|
IssuesEvent
|
2020-07-21 18:08:54
|
obinnaokechukwu/internship-2020
|
https://api.github.com/repos/obinnaokechukwu/internship-2020
|
opened
|
Translate newly added text into multiple languages
|
process
|
Translate newly added text into multiple languages
1. Find all newly added strings
2. Get translation
3. Update language files with translation
|
1.0
|
Translate newly added text into multiple languages - Translate newly added text into multiple languages
1. Find all newly added strings
2. Get translation
3. Update language files with translation
|
process
|
translate newly added text into multiple languages translate newly added text into multiple languages find all newly added strings get translation update language files with translation
| 1
|
3,296
| 9,414,716,776
|
IssuesEvent
|
2019-04-10 10:49:45
|
xXxH4n5Wur5t360NoPullsxXx/vodafonica
|
https://api.github.com/repos/xXxH4n5Wur5t360NoPullsxXx/vodafonica
|
closed
|
Webanwendungskomponente zu Subkomponenten aufschlüsseln
|
architecture
|
Siehe [Diagramm](https://github.com/xXxH4n5Wur5t360NoPullsxXx/vodafonica/blob/master/doc/architecture/t-architecture.pdf)
Die Komponente der Webanwendung muss noch in die einzelnen GUI Komponenten aufgeschlüsselt werden (Hatten wir an der Tafel stehen)
|
1.0
|
Webanwendungskomponente zu Subkomponenten aufschlüsseln - Siehe [Diagramm](https://github.com/xXxH4n5Wur5t360NoPullsxXx/vodafonica/blob/master/doc/architecture/t-architecture.pdf)
Die Komponente der Webanwendung muss noch in die einzelnen GUI Komponenten aufgeschlüsselt werden (Hatten wir an der Tafel stehen)
|
non_process
|
webanwendungskomponente zu subkomponenten aufschlüsseln siehe die komponente der webanwendung muss noch in die einzelnen gui komponenten aufgeschlüsselt werden hatten wir an der tafel stehen
| 0
|
102,951
| 4,163,323,072
|
IssuesEvent
|
2016-06-18 01:48:34
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
cAdvisor /stats/summary endpoint in kubelet returns incorrect cpu usage numbers
|
kind/bug priority/P0 team/node
|
### Environment
Kubernetes version: 1.2.3
Docker version: 1.10.3
3 node(c4.xLarge) cluster on AWS running CoreOS 1010.4.0.
### Issue
After facing an issue with incorrect metrics being reported by heapster kubernetes/heapster#1177 I tried querying the cadvisor /stats/summary endpoint directly to see if that would give me consistent values for node cpu usage.
I have one pod with cpu request=1000m and limit=1000m. In that pod I run a busy loop to consume a 100% of the cpu. This is what top shows on the node.
<img width="648" alt="screen shot 2016-06-10 at 10 07 52 am" src="https://cloud.githubusercontent.com/assets/6660644/15975336/e922fe28-2f01-11e6-9dd9-7482fa52b4e4.png">
I query the /stats/summary endpoint every 5 seconds, however it seems that the latest timestamps are only updated every 15 seconds or so. Checking the `summary.Node.CPU.UsageNanoCores` value from the summary returned gives me the following output(formatted):
```
TS:2016-06-09T20:05:20-07:00, Percentage:105.539566 Val:1055395658
TS:2016-06-09T20:05:31-07:00, Percentage:107.416097 Val:1074160974
TS:2016-06-09T20:05:44-07:00, Percentage:108.195877 Val:1081958770
TS:2016-06-09T20:05:59-07:00, Percentage:106.670910 Val:1066709101
TS:2016-06-09T20:06:19-07:00, Percentage:14.360576 Val:143605762
TS:2016-06-09T20:06:31-07:00, Percentage:108.000277 Val:1080002769
TS:2016-06-09T20:06:41-07:00, Percentage:108.373232 Val:1083732315
TS:2016-06-09T20:06:56-07:00, Percentage:107.025070 Val:1070250700
TS:2016-06-09T20:07:16-07:00, Percentage:13.004869 Val:130048687
TS:2016-06-09T20:07:31-07:00, Percentage:106.839146 Val:1068391461
TS:2016-06-09T20:07:48-07:00, Percentage:107.614464 Val:1076144640
TS:2016-06-09T20:08:06-07:00, Percentage:4.232330 Val:42323305
TS:2016-06-09T20:08:20-07:00, Percentage:106.009173 Val:1060091732
TS:2016-06-09T20:08:35-07:00, Percentage:108.121440 Val:1081214401
TS:2016-06-09T20:08:50-07:00, Percentage:106.659561 Val:1066595609
TS:2016-06-09T20:09:07-07:00, Percentage:1.724644 Val:17246439
TS:2016-06-09T20:09:19-07:00, Percentage:106.633227 Val:1066332268
TS:2016-06-09T20:09:38-07:00, Percentage:9.938621 Val:99386209
TS:2016-06-09T20:09:53-07:00, Percentage:107.046112 Val:1070461118
TS:2016-06-09T20:10:10-07:00, Percentage:3.373636 Val:33736361
TS:2016-06-09T20:10:25-07:00, Percentage:107.338541 Val:1073385413
TS:2016-06-09T20:10:39-07:00, Percentage:108.575783 Val:1085757834
TS:2016-06-09T20:10:54-07:00, Percentage:107.055382 Val:1070553817
TS:2016-06-09T20:11:13-07:00, Percentage:7.869509 Val:78695088
TS:2016-06-09T20:11:32-07:00, Percentage:11.476262 Val:114762620
TS:2016-06-09T20:11:45-07:00, Percentage:106.928681 Val:1069286811
TS:2016-06-09T20:12:03-07:00, Percentage:3.309632 Val:33096320
TS:2016-06-09T20:12:15-07:00, Percentage:105.832345 Val:1058323450
TS:2016-06-09T20:12:34-07:00, Percentage:5.079409 Val:50794090
TS:2016-06-09T20:12:47-07:00, Percentage:106.305439 Val:1063054389
TS:2016-06-09T20:13:05-07:00, Percentage:3.613690 Val:36136900
TS:2016-06-09T20:13:24-07:00, Percentage:9.785441 Val:97854409
TS:2016-06-09T20:13:43-07:00, Percentage:12.661783 Val:126617830
```
As you can see I'm not getting a steady report of near 100% cpu usage values for UsageNanoCores. Any idea why this might be the case or how I can debug this issue. Also is there any way I can change the resolution of the summary stats to get more fine grained reporting.
|
1.0
|
cAdvisor /stats/summary endpoint in kubelet returns incorrect cpu usage numbers - ### Environment
Kubernetes version: 1.2.3
Docker version: 1.10.3
3 node(c4.xLarge) cluster on AWS running CoreOS 1010.4.0.
### Issue
After facing an issue with incorrect metrics being reported by heapster kubernetes/heapster#1177 I tried querying the cadvisor /stats/summary endpoint directly to see if that would give me consistent values for node cpu usage.
I have one pod with cpu request=1000m and limit=1000m. In that pod I run a busy loop to consume a 100% of the cpu. This is what top shows on the node.
<img width="648" alt="screen shot 2016-06-10 at 10 07 52 am" src="https://cloud.githubusercontent.com/assets/6660644/15975336/e922fe28-2f01-11e6-9dd9-7482fa52b4e4.png">
I query the /stats/summary endpoint every 5 seconds, however it seems that the latest timestamps are only updated every 15 seconds or so. Checking the `summary.Node.CPU.UsageNanoCores` value from the summary returned gives me the following output(formatted):
```
TS:2016-06-09T20:05:20-07:00, Percentage:105.539566 Val:1055395658
TS:2016-06-09T20:05:31-07:00, Percentage:107.416097 Val:1074160974
TS:2016-06-09T20:05:44-07:00, Percentage:108.195877 Val:1081958770
TS:2016-06-09T20:05:59-07:00, Percentage:106.670910 Val:1066709101
TS:2016-06-09T20:06:19-07:00, Percentage:14.360576 Val:143605762
TS:2016-06-09T20:06:31-07:00, Percentage:108.000277 Val:1080002769
TS:2016-06-09T20:06:41-07:00, Percentage:108.373232 Val:1083732315
TS:2016-06-09T20:06:56-07:00, Percentage:107.025070 Val:1070250700
TS:2016-06-09T20:07:16-07:00, Percentage:13.004869 Val:130048687
TS:2016-06-09T20:07:31-07:00, Percentage:106.839146 Val:1068391461
TS:2016-06-09T20:07:48-07:00, Percentage:107.614464 Val:1076144640
TS:2016-06-09T20:08:06-07:00, Percentage:4.232330 Val:42323305
TS:2016-06-09T20:08:20-07:00, Percentage:106.009173 Val:1060091732
TS:2016-06-09T20:08:35-07:00, Percentage:108.121440 Val:1081214401
TS:2016-06-09T20:08:50-07:00, Percentage:106.659561 Val:1066595609
TS:2016-06-09T20:09:07-07:00, Percentage:1.724644 Val:17246439
TS:2016-06-09T20:09:19-07:00, Percentage:106.633227 Val:1066332268
TS:2016-06-09T20:09:38-07:00, Percentage:9.938621 Val:99386209
TS:2016-06-09T20:09:53-07:00, Percentage:107.046112 Val:1070461118
TS:2016-06-09T20:10:10-07:00, Percentage:3.373636 Val:33736361
TS:2016-06-09T20:10:25-07:00, Percentage:107.338541 Val:1073385413
TS:2016-06-09T20:10:39-07:00, Percentage:108.575783 Val:1085757834
TS:2016-06-09T20:10:54-07:00, Percentage:107.055382 Val:1070553817
TS:2016-06-09T20:11:13-07:00, Percentage:7.869509 Val:78695088
TS:2016-06-09T20:11:32-07:00, Percentage:11.476262 Val:114762620
TS:2016-06-09T20:11:45-07:00, Percentage:106.928681 Val:1069286811
TS:2016-06-09T20:12:03-07:00, Percentage:3.309632 Val:33096320
TS:2016-06-09T20:12:15-07:00, Percentage:105.832345 Val:1058323450
TS:2016-06-09T20:12:34-07:00, Percentage:5.079409 Val:50794090
TS:2016-06-09T20:12:47-07:00, Percentage:106.305439 Val:1063054389
TS:2016-06-09T20:13:05-07:00, Percentage:3.613690 Val:36136900
TS:2016-06-09T20:13:24-07:00, Percentage:9.785441 Val:97854409
TS:2016-06-09T20:13:43-07:00, Percentage:12.661783 Val:126617830
```
As you can see I'm not getting a steady report of near 100% cpu usage values for UsageNanoCores. Any idea why this might be the case or how I can debug this issue. Also is there any way I can change the resolution of the summary stats to get more fine grained reporting.
|
non_process
|
cadvisor stats summary endpoint in kubelet returns incorrect cpu usage numbers environment kubernetes version docker version node xlarge cluster on aws running coreos issue after facing an issue with incorrect metrics being reported by heapster kubernetes heapster i tried querying the cadvisor stats summary endpoint directly to see if that would give me consistent values for node cpu usage i have one pod with cpu request and limit in that pod i run a busy loop to consume a of the cpu this is what top shows on the node img width alt screen shot at am src i query the stats summary endpoint every seconds however it seems that the latest timestamps are only updated every seconds or so checking the summary node cpu usagenanocores value from the summary returned gives me the following output formatted ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val ts percentage val as you can see i m not getting a steady report of near cpu usage values for usagenanocores any idea why this might be the case or how i can debug this issue also is there any way i can change the resolution of the summary stats to get more fine grained reporting
| 0
|
2,363
| 5,166,357,806
|
IssuesEvent
|
2017-01-17 16:02:41
|
jmbredal/NAV
|
https://api.github.com/repos/jmbredal/NAV
|
closed
|
NAV daemons don't properly re-open log files on SIGHUP
|
Fix Released Medium process-control
|
The alertEngine, pping and servicemon daemons do not properly close and
re-open their logfiles when sent a SIGHUP signal.
When logrotate is used to rotate the log files of these processes, they
will keep their file descriptors and continue writing log messages to
non-existant files - which will actually occupy more and more space on the
partition.
Usually one would configure logrotate to issue a HUP signal to these kinds
of processes, but this won't work for the three daemons in question, and it
should.
[http://sourceforge.net/tracker/index.php?func=detail&aid=1697527&group_id=107608&atid=648170]
|
1.0
|
NAV daemons don't properly re-open log files on SIGHUP - The alertEngine, pping and servicemon daemons do not properly close and
re-open their logfiles when sent a SIGHUP signal.
When logrotate is used to rotate the log files of these processes, they
will keep their file descriptors and continue writing log messages to
non-existant files - which will actually occupy more and more space on the
partition.
Usually one would configure logrotate to issue a HUP signal to these kinds
of processes, but this won't work for the three daemons in question, and it
should.
[http://sourceforge.net/tracker/index.php?func=detail&aid=1697527&group_id=107608&atid=648170]
|
process
|
nav daemons don t properly re open log files on sighup the alertengine pping and servicemon daemons do not properly close and re open their logfiles when sent a sighup signal when logrotate is used to rotate the log files of these processes they will keep their file descriptors and continue writing log messages to non existant files which will actually occupy more and more space on the partition usually one would configure logrotate to issue a hup signal to these kinds of processes but this won t work for the three daemons in question and it should
| 1
|
11,232
| 14,007,905,591
|
IssuesEvent
|
2020-10-28 22:28:35
|
kubeflow/internal-acls
|
https://api.github.com/repos/kubeflow/internal-acls
|
closed
|
Automate approval of requests
|
area/front-end kind/feature kind/process lifecycle/stale
|
/kind process
For policy reasons, I don't think we can have non-Googlers approving requests to change the GitHub policy because granting a non-googler permission to approve PRs would be overly permissive; i.e. its probably permissive for non googlers to approve
joining the github org or a possible a github team; but changing github admins should not be allowed.
If we become bottlenecked on Googlers approving routine requests then we should automate the approval of routine requests.
Building a GitHub app that can verify only safe changes are included and adding an LGTM should be relatively straightforward.
/cc @animeshsingh
|
1.0
|
Automate approval of requests - /kind process
For policy reasons, I don't think we can have non-Googlers approving requests to change the GitHub policy because granting a non-googler permission to approve PRs would be overly permissive; i.e. its probably permissive for non googlers to approve
joining the github org or a possible a github team; but changing github admins should not be allowed.
If we become bottlenecked on Googlers approving routine requests then we should automate the approval of routine requests.
Building a GitHub app that can verify only safe changes are included and adding an LGTM should be relatively straightforward.
/cc @animeshsingh
|
process
|
automate approval of requests kind process for policy reasons i don t think we can have non googlers approving requests to change the github policy because granting a non googler permission to approve prs would be overly permissive i e its probably permissive for non googlers to approve joining the github org or a possible a github team but changing github admins should not be allowed if we become bottlenecked on googlers approving routine requests then we should automate the approval of routine requests building a github app that can verify only safe changes are included and adding an lgtm should be relatively straightforward cc animeshsingh
| 1
|
18,200
| 24,254,878,424
|
IssuesEvent
|
2022-09-27 16:53:24
|
eosnetworkfoundation/devrel
|
https://api.github.com/repos/eosnetworkfoundation/devrel
|
closed
|
Github training 1 - working with Github Editor
|
Process
|
Train project manager on using the web-based Github Editor for:
- creating/editing files
- creating commits
- push changes
- check changes on remote
|
1.0
|
Github training 1 - working with Github Editor - Train project manager on using the web-based Github Editor for:
- creating/editing files
- creating commits
- push changes
- check changes on remote
|
process
|
github training working with github editor train project manager on using the web based github editor for creating editing files creating commits push changes check changes on remote
| 1
|
76,130
| 9,382,428,006
|
IssuesEvent
|
2019-04-04 22:24:19
|
oasis-tcs/sarif-spec
|
https://api.github.com/repos/oasis-tcs/sarif-spec
|
closed
|
Refine URL normalization guidance for file URLS
|
CSD.2 bug design-approved impact-non-breaking-change merged resolved-fixed
|
Producers SHOULD not include .. or . path segments in a URI after forming the full URI. Before converting to a URI, paths that contain a .. or . should be normalized to a canonical absolute path using an appropriate algorithm for the operating system on which the tool ran. This is necessary as the path /d1/../f naively converted to a URI is file:///d1/../f which resolves to file:///f using RFC 3986. If /d1 is a symbolic link to the directory d2/d3 then the correct URI is file:///d2/f.
NOTE: If a SARIF producer cannot determine the correct canonical representation of the path for some reason such as the original file system not being available, then the producer MAY generate URIs with .. segments.
For file scheme URIs, consumer must not normalize .. segments out of the path. Any paths that contain a .. segment should treat the directory formed by the segments prior to and including the .. segment as if it were a unique directory in the file system, even if RFC 3986 normalization produces identical.
|
1.0
|
Refine URL normalization guidance for file URLS - Producers SHOULD not include .. or . path segments in a URI after forming the full URI. Before converting to a URI, paths that contain a .. or . should be normalized to a canonical absolute path using an appropriate algorithm for the operating system on which the tool ran. This is necessary as the path /d1/../f naively converted to a URI is file:///d1/../f which resolves to file:///f using RFC 3986. If /d1 is a symbolic link to the directory d2/d3 then the correct URI is file:///d2/f.
NOTE: If a SARIF producer cannot determine the correct canonical representation of the path for some reason such as the original file system not being available, then the producer MAY generate URIs with .. segments.
For file scheme URIs, consumer must not normalize .. segments out of the path. Any paths that contain a .. segment should treat the directory formed by the segments prior to and including the .. segment as if it were a unique directory in the file system, even if RFC 3986 normalization produces identical.
|
non_process
|
refine url normalization guidance for file urls producers should not include or path segments in a uri after forming the full uri before converting to a uri paths that contain a or should be normalized to a canonical absolute path using an appropriate algorithm for the operating system on which the tool ran this is necessary as the path f naively converted to a uri is file f which resolves to file f using rfc if is a symbolic link to the directory then the correct uri is file f note if a sarif producer cannot determine the correct canonical representation of the path for some reason such as the original file system not being available then the producer may generate uris with segments for file scheme uris consumer must not normalize segments out of the path any paths that contain a segment should treat the directory formed by the segments prior to and including the segment as if it were a unique directory in the file system even if rfc normalization produces identical
| 0
|
9,144
| 12,203,193,879
|
IssuesEvent
|
2020-04-30 10:10:59
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
Observability | Structured logs for doc index updater
|
EPIC - Auto Batch Process :oncoming_automobile:
|
## User want
Part of #408
As a tech user
I want set structured logs for doc index updater
So I can have meaningful data to monitor/alert/ debug
### Technical acceptance criteria
Should have structured logs with this suggested schema
```json
{
"timestamp": "2020-03-18T13:13:05.442Z",
"service_name": "medicines/doc-index-updater",
"correlation_id": "92bb6eb5-b8c8-4ad3-8e5a-65bb71e34729",
"event": {
"label": "DOCUMENT_MANAGER_CREATE",
"payload": {
//
}
},
"level": "INFO"
}
```
for the following events/interactions:
- [ ] Sentinel <--> Adding a job to the Document Manager
- [ ] Document Manager <--> Service bus
- [ ] Service bus <--> Check in
- [ ] Check in <--> Sentinel file system
- [ ] Check in <--> blob storage
- [ ] Sentinel <--> Checking job status
- [ ] State manager <--> Redis
- [ ] Serve logs from warp
- [ ] Delete from Blob Storage
- [ ] Delete from Index
- [ ] Create from Index
Note the following:
- [ ] In the event of everything working as expected, log to **DEBUG**.
- [ ] In the event of something which presents a possible but recoverable problem, log to **INFO**.
- [ ] In the event of something causing an individual job to be impossible to process, log to **WARN**.
- [ ] In the event of something fatal for all jobs occurring, log to **ERROR**.
**Size**
M
**Value**
**Effort**
### Exit Criteria met
- [ x] Backlog
- [ x] Discovery
- [x ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
1.0
|
Observability | Structured logs for doc index updater - ## User want
Part of #408
As a tech user
I want set structured logs for doc index updater
So I can have meaningful data to monitor/alert/ debug
### Technical acceptance criteria
Should have structured logs with this suggested schema
```json
{
"timestamp": "2020-03-18T13:13:05.442Z",
"service_name": "medicines/doc-index-updater",
"correlation_id": "92bb6eb5-b8c8-4ad3-8e5a-65bb71e34729",
"event": {
"label": "DOCUMENT_MANAGER_CREATE",
"payload": {
//
}
},
"level": "INFO"
}
```
for the following events/interactions:
- [ ] Sentinel <--> Adding a job to the Document Manager
- [ ] Document Manager <--> Service bus
- [ ] Service bus <--> Check in
- [ ] Check in <--> Sentinel file system
- [ ] Check in <--> blob storage
- [ ] Sentinel <--> Checking job status
- [ ] State manager <--> Redis
- [ ] Serve logs from warp
- [ ] Delete from Blob Storage
- [ ] Delete from Index
- [ ] Create from Index
Note the following:
- [ ] In the event of everything working as expected, log to **DEBUG**.
- [ ] In the event of something which presents a possible but recoverable problem, log to **INFO**.
- [ ] In the event of something causing an individual job to be impossible to process, log to **WARN**.
- [ ] In the event of something fatal for all jobs occurring, log to **ERROR**.
**Size**
M
**Value**
**Effort**
### Exit Criteria met
- [ x] Backlog
- [ x] Discovery
- [x ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
process
|
observability structured logs for doc index updater user want part of as a tech user i want set structured logs for doc index updater so i can have meaningful data to monitor alert debug technical acceptance criteria should have structured logs with this suggested schema json timestamp service name medicines doc index updater correlation id event label document manager create payload level info for the following events interactions sentinel adding a job to the document manager document manager service bus service bus check in check in sentinel file system check in blob storage sentinel checking job status state manager redis serve logs from warp delete from blob storage delete from index create from index note the following in the event of everything working as expected log to debug in the event of something which presents a possible but recoverable problem log to info in the event of something causing an individual job to be impossible to process log to warn in the event of something fatal for all jobs occurring log to error size m value effort exit criteria met backlog discovery duxd development quality assurance release and validate
| 1
|
69,796
| 3,314,943,895
|
IssuesEvent
|
2015-11-06 09:14:45
|
chebfun/chebfun
|
https://api.github.com/repos/chebfun/chebfun
|
closed
|
What should the single output behaviour from RATINTERP be?
|
low-priority question
|
When a user invokes `r = ratinterp(f, np, nq)` what is returned is a chebfun representation of the numerator `p` in the rational interpolant. This confused me for a while, as I was expecting a single output
call to return the function handle `r = @(x) p(x)/q(x)` (since by itself `p` is not so useful?).
The same is true of `chebpade()`. Note that `padeapprox()` returns the function handle as the first output.
Edit: `pade()` --> `padeapprox()`.
|
1.0
|
What should the single output behaviour from RATINTERP be? - When a user invokes `r = ratinterp(f, np, nq)` what is returned is a chebfun representation of the numerator `p` in the rational interpolant. This confused me for a while, as I was expecting a single output
call to return the function handle `r = @(x) p(x)/q(x)` (since by itself `p` is not so useful?).
The same is true of `chebpade()`. Note that `padeapprox()` returns the function handle as the first output.
Edit: `pade()` --> `padeapprox()`.
|
non_process
|
what should the single output behaviour from ratinterp be when a user invokes r ratinterp f np nq what is returned is a chebfun representation of the numerator p in the rational interpolant this confused me for a while as i was expecting a single output call to return the function handle r x p x q x since by itself p is not so useful the same is true of chebpade note that padeapprox returns the function handle as the first output edit pade padeapprox
| 0
|
18,727
| 24,615,016,054
|
IssuesEvent
|
2022-10-15 07:30:38
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
6502 JMP indirect incorrectly disassembled
|
Type: Bug Feature: Processor/6502 Status: Internal
|
**Describe the bug**
The way Ghidra represents an indirect JMP is misleading. An indirect JMP (0x6c) is where the CPU reads the contents of a memory location and then sets PC to that address. This is opposed to an absolute JMP (0x4c) which just jumps to the passed address.
This is often used to make look up tables which are often dynamic.
Ghidra tries to be clever and reads the memory location and puts in a link to the result; which causes it to be inconsistent.
For example (in the below, CLIV is set to 0x0208),
```
OSCLI
fff7 6c 08 02 JMP CLIV
```
Is an indirect JMP, which should be disassembled as JMP (CLIV); but is disassembled as a JMP to 0x0208 when it should be the address **at** 0x0208.
But this is inconsistent as it can sometimes return a variable to the actual destination:
```
OSBYTE XREF[5]: FUN_0fc2:0fcd(c),
fff4 6c 0a 02 JMP offset FUN_e772
```
I can't work out a pattern of when it sets the destination to the indirect value or the absolute value.
This method makes it looks exactly the same as an absolute JMP:
```
2292 4c 33 23 JMP LAB_2333
```
This could cause misinterpretations in program flows and; if the program is using this for dynamic tables then will cause it to be wrong in many circumstances.
**To Reproduce**
Create binary file with at least:
4c 0a 02
6c 0a 02
Where 0x020a has a value.
**Expected behavior**
I'd expect an indirect JMP to disassemble using the indirect mode and an absolute JMP to use the absolute mode:
```
4c 0a 02 JMP 0x020a Absolute
6c 0a 02 JMP (0x020a) Indirect
```
|
1.0
|
6502 JMP indirect incorrectly disassembled - **Describe the bug**
The way Ghidra represents an indirect JMP is misleading. An indirect JMP (0x6c) is where the CPU reads the contents of a memory location and then sets PC to that address. This is opposed to an absolute JMP (0x4c) which just jumps to the passed address.
This is often used to make look up tables which are often dynamic.
Ghidra tries to be clever and reads the memory location and puts in a link to the result; which causes it to be inconsistent.
For example (in the below, CLIV is set to 0x0208),
```
OSCLI
fff7 6c 08 02 JMP CLIV
```
Is an indirect JMP, which should be disassembled as JMP (CLIV); but is disassembled as a JMP to 0x0208 when it should be the address **at** 0x0208.
But this is inconsistent as it can sometimes return a variable to the actual destination:
```
OSBYTE XREF[5]: FUN_0fc2:0fcd(c),
fff4 6c 0a 02 JMP offset FUN_e772
```
I can't work out a pattern of when it sets the destination to the indirect value or the absolute value.
This method makes it looks exactly the same as an absolute JMP:
```
2292 4c 33 23 JMP LAB_2333
```
This could cause misinterpretations in program flows and; if the program is using this for dynamic tables then will cause it to be wrong in many circumstances.
**To Reproduce**
Create binary file with at least:
4c 0a 02
6c 0a 02
Where 0x020a has a value.
**Expected behavior**
I'd expect an indirect JMP to disassemble using the indirect mode and an absolute JMP to use the absolute mode:
```
4c 0a 02 JMP 0x020a Absolute
6c 0a 02 JMP (0x020a) Indirect
```
|
process
|
jmp indirect incorrectly disassembled describe the bug the way ghidra represents an indirect jmp is misleading an indirect jmp is where the cpu reads the contents of a memory location and then sets pc to that address this is opposed to an absolute jmp which just jumps to the passed address this is often used to make look up tables which are often dynamic ghidra tries to be clever and reads the memory location and puts in a link to the result which causes it to be inconsistent for example in the below cliv is set to oscli jmp cliv is an indirect jmp which should be disassembled as jmp cliv but is disassembled as a jmp to when it should be the address at but this is inconsistent as it can sometimes return a variable to the actual destination osbyte xref fun c jmp offset fun i can t work out a pattern of when it sets the destination to the indirect value or the absolute value this method makes it looks exactly the same as an absolute jmp jmp lab this could cause misinterpretations in program flows and if the program is using this for dynamic tables then will cause it to be wrong in many circumstances to reproduce create binary file with at least where has a value expected behavior i d expect an indirect jmp to disassemble using the indirect mode and an absolute jmp to use the absolute mode jmp absolute jmp indirect
| 1
|
30,646
| 5,830,442,152
|
IssuesEvent
|
2017-05-08 16:50:32
|
18F/web-design-standards
|
https://api.github.com/repos/18F/web-design-standards
|
opened
|
Write v1.2 JS API compatibility layer
|
[Priority] Blocker [Release] 1.2 [Skill] DevOps [Skill] Front end [Type] Communication [Type] Documentation
|
This is the primary task related to the decision made in https://github.com/18F/web-design-standards/issues/1885#issuecomment-299921655. This work is underway in #1908, but for posterity, here's what we need to do:
- [ ] Identify private submodules (APIs) that users are `require()`-ing
- [ ] Back-port the relevant v1.1 unit tests to the [1.2 release branch](../tree/release-1.2.0) in a separate spec directory
- [ ] Write compatibility shims for the relevant submodules (components), and mark them so that they're easy to remove in v2.0
|
1.0
|
Write v1.2 JS API compatibility layer - This is the primary task related to the decision made in https://github.com/18F/web-design-standards/issues/1885#issuecomment-299921655. This work is underway in #1908, but for posterity, here's what we need to do:
- [ ] Identify private submodules (APIs) that users are `require()`-ing
- [ ] Back-port the relevant v1.1 unit tests to the [1.2 release branch](../tree/release-1.2.0) in a separate spec directory
- [ ] Write compatibility shims for the relevant submodules (components), and mark them so that they're easy to remove in v2.0
|
non_process
|
write js api compatibility layer this is the primary task related to the decision made in this work is underway in but for posterity here s what we need to do identify private submodules apis that users are require ing back port the relevant unit tests to the tree release in a separate spec directory write compatibility shims for the relevant submodules components and mark them so that they re easy to remove in
| 0
|
85,154
| 24,525,038,281
|
IssuesEvent
|
2022-10-11 12:32:37
|
microsoft/fluentui
|
https://api.github.com/repos/microsoft/fluentui
|
closed
|
[Feature]: Remove unnecessary checks from screener code
|
Area: Build System Type: Epic
|
### Library
React / v8 (@fluentui/react)
### Describe the feature that you would like added
With the changes from #24292, the ``skipScreener`` environment variable can now be used to determine whether a screener check should be run or not, so other redundant checks and usage of other environment variables can be discarded.
### Have you discussed this feature with our team
ling1726
### Additional context
_No response_
### Validations
- [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
|
1.0
|
[Feature]: Remove unnecessary checks from screener code - ### Library
React / v8 (@fluentui/react)
### Describe the feature that you would like added
With the changes from #24292, the ``skipScreener`` environment variable can now be used to determine whether a screener check should be run or not, so other redundant checks and usage of other environment variables can be discarded.
### Have you discussed this feature with our team
ling1726
### Additional context
_No response_
### Validations
- [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
|
non_process
|
remove unnecessary checks from screener code library react fluentui react describe the feature that you would like added with the changes from the skipscreener environment variable can now be used to determine whether a screener check should be run or not so other redundant checks and usage of other environment variables can be discarded have you discussed this feature with our team additional context no response validations check that there isn t already an issue that request the same feature to avoid creating a duplicate
| 0
|
15,121
| 3,310,368,759
|
IssuesEvent
|
2015-11-05 08:36:56
|
geetsisbac/WCVVENIXYFVIRBXH3BYTI6TE
|
https://api.github.com/repos/geetsisbac/WCVVENIXYFVIRBXH3BYTI6TE
|
closed
|
Glt7RyHmocTWQXh4CQT0By3KWR1Kj/WR/spy4+F7Y2SQcC13+DtDS/Dw1D8DeDC7u50A6KslorZMBPnemFl3J/UXMw9QpgFLG5SEq3w+E/mswE4EcfeLXPF75b+ysE8fcWHe6lXR3RrXl1inIAASfw/V2deikXheIFI9IDuxEME=
|
design
|
JuZiQzs98LCkwQ8Uju7zCbp2KefIdJBxAH/GX7CRvmvel0XGLHHK2hprZXqG7/Qp23idmcBGRuhKxIOuDA0hLw7AlhTh+iLllRwOFWBNzjkjVaOrVvgmz9BWOrb2qxrc7ZJ5dKxOJ8RypmEyGCZqaHY+9ijgRpjOTNUrKOee1wR2NbIykPONWmtGE/I+Jmu7zmDrikp/Io4Y+OJcurV3b5vFYQ5iveYP1+QUXo8ODLSqRzfr7oUSVP0JUgKvqPp7d2T2WRMvHQbyNffLTcxP+j+9Tth8Ti4qlWWyiTiGapEGRFZI6qT7X9DfJkXwGSc1OH5JPKvGoNS9MugWFooofxQ7DJCc9v9Vu1xUjkvinMIh5Q5WCrrsU7krUSC0kHnSeffkOiaTM6uWVpcMqqCXk+B+zDjLH9rWZERKj4vSYiobDpE/MVh+06EcQAxU1VagNTFTS/m7f2tydyLChhjwc/pKhlm0R6juUnBAOTo872AYM7p1XzipucwvF9OkTVU47CJTqngSAMpr5qxTyPkazBtRYDPVUIv3pfUbt1C1x+Uc6wbqlzIFNstQY+Iz+U/QU07FNDfk+gxBkHqc0TYkWQT12srC0WLcCGTUlWqLZ+2mvVstgsP4OppIAq0af5H/MuC8UFlhKtxNiX/x+bYF5vBx+6moOgekg5BXB6CEFMOpp/TFl61glswcqHBosNF3r8dirVa3Xy6jMeVVY+yneNOzOBibwy8LfbDCcJ6bfebBs7H8l2qx4v+iaFk/u6RrTsY/5UcHXHczMeybssQlMy5Wnr1DLZSlcbsuypSB8SBujJotWnUJPtbmx0wcjcp+pvbPzICk0aBwQB5BCeT8ITkaJ5mYFmKPurCNQ3UdbmV57oqxUdh+U+4S+XmNSsEOLWEnLmfy5RiBTYZiaIFOfW0tFiPH/NgED26PPPX1ItnLSYFYtwjYMHnTJB0Bggy3vrRaheRIJrRhbMLmlelntZyWzQOm45hAX8ttwEX9frHC4JMLM33Uj9RFBZAuzHMw9EJEvNnNI981+g7Vxkd4SV85F1VqcY0fYUnpddOVjc0PVo+yMsGhCsw4lIGKKLJdnA19qRbngXlnQ+URrOtqWHrDU0wmZsBwUh+97LZJ63QZOOzEH3dj/ksoxxicZgbkp38WCdeaU4SdBUBtyeGN0gn6cCU1IuH7Jnoy5m2IcpnIaZRKApBPFzqL3JFgafzrnzzdmuWJ838oMsq6iTobMZh8YXPnV06sYjPMFEKcG+TkG32r0dqhB6cUI05Jbh0rzpvyNJnVMmRe4/Vxand39+o8RV+M6Zt0IdDHcI/DrdJ57oqxUdh+U+4S+XmNSsEOO/I9wBvSeDOfv+B/GmxUKkPfs/0cpmkadOvBdfTNw3BEM0nBaeiy8R7sQ+lmSv6rYefTevoZryqSV63YFekEhbMNmwLRnoyNjIKD6sCJaYhfI4dDlZBp4+CSMPJ8ZFhORTUFtD7xvIccROAF622A4s5Vlyq5uBMhkclKE6exTeFYNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S7hgEFsM8Q+rhrMzV5k8dZBJhhpu7vEkYpNGAuqciOYVwOj1BO2PSJesb2Nw/aBWoF3Wl+HbBZkWHnO1ZvTeUHqQxNEyz7Bdnedfx9JzDSjcPq2vDn9gWT21to5D7Ces6Y4mewISwDxm2M3P+B/IgJDNLVn+bPr1MXL3ElC4PuoePsZv192ZAoVKtphpYO/Vbc3r9Eswt8JgohlQNrc/K/Ymh/wWaH/zPCvotBBl3Qu1edhMo4M2X1Bgf97uYkXfXhtgBfUfRR+9k7HwAOjS+/T+QbZtunnKec8p/oRcvYmrBRRZK2TPEigsCTs/byfc1hDLAwY+2sf8zR2tbe6I3d4+rE07PbqaCw5XBmi7SzYEKugXiy3k8/Rg+aWumZvZVi86pHL+YmsMMS7cqDzJ+fru6RK2/iergcwaQo0O+DQ2aoeaQoWfXi/7+/rvrdD35/Bx+6moOgekg5BXB6CEFMOpp/TFl61glswcqHBosNF3d+tDg60bI7Nfe+0F6Wp4XtOzOBibwy8LfbDCcJ6bfeZBbHmwaVUSrhqGP2MLH4Q9wC5gUncCuk5ceFC4/ABt1kmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagMPnYwzy5FvSaGUsybv62JSCbfRhjBqvcR8UlrQIysbQANOFBbhYmc0AxKBfytjphuv4y3jeYqcet3gLS6pNrx9ReqLhjevvLzkxSHtXj9EXl00jTwnn3RAjpbze0Q9C3Zy8rJu+ZHL0iTuOk73ab3WKi7zlnYqHqt9rnnh+Gg4pYSFYk9xeQNmwMJ8Ng3ZeaSqlxuXaZmLtKxPndGrXokGhXN72Z9ZgcCXrE3Og9+QRYNssv63MLq96TzAFC1x5iv4QSofPEc3wl50vE6XI1Q/o2HyqrO4w47Og9uKk7uFKYjlgOpXApd52n9TTPJFBZtoO7z4CXWYyWbaDIJ9YwBp3mGU9jN5YPaHytgVjzxPxpcnlq1718+7t7NeJnVteFjxCE5fcK0/IavHLG7D7ol9sz4u9UnArnMa5e9zBGJCcJIvN9vStY3A2UbqnirRP/06g/HAnJK8RxAgcaMqIegPuZb+ymtbR8Cstvj3OJnrwK02lF2TzRHy0BpA3hggLg/zQefcma4VEr5KgoJOXATxrzkwyNhvraMaOe/t+txyQXncd248YVsBnAAAi32q3kirzqscXrG+uvBOaM7ptPYr2EA0r996EDFl0/nXBdNuVTLiN5cCcP8tLD7DzEWpkisX3F3B/LAusojTszMcf34pJmJrJmrRZXfCkmTJVV3ul5mypqMCD5aH/a5DUoOjM3CtNpRdk80R8tAaQN4YIC4P80Hn3JmuFRK+SoKCTlwE8HBkfgkbuRQAhEG4mqy2Iwtq4VJ8ubDq2PouiQqmIJrdYZujLjfyKazntDCTobswbVTCroEJwbnGFvtyo1UNLFcv7vpwfvku9j6kCTtkuBbWKGUpLwXUjJg31qO1xrnrIjQVDNHcPtOlWB5kiGve7+MZZZ/douxw8HHFFcFtSW24RFuXEllM0av2zZFj4WtAjKTG5NVJOmFyNv4QXQd6YS31LAkG8RpDP3UaHOyv2jyOt4KvcNtUm89vi+D+c8I4EKaM+qr2ydNN7CpmP6MyNEl+775FVJ3kHHNA9lkxW/RiOwsViYb1Lu8NjJ27XepPkRy7gKsseeyv7Z+AgSQTbIMZZZ/douxw8HHFFcFtSW24RFuXEllM0av2zZFj4WtAjGD8LS9TGLN3Zln/aybHgMq+RN3X0uW0glMGl5GShYznNyWidSsIc5fSI6LV481LYKaM+qr2ydNN7CpmP6MyNERxp+V21lGFj7hUe0XOstcazp4JRRrzlKfv15RUflLHMB78kE3IYfad928d/LzgvK6uFhAGlMBMgDKchhlHiFK9sz4u9UnArnMa5e9zBGJCfiV/NCFWayAAoaGZdQadU+4Z4lqiBTjDCJ2K5U18UnrI+AigzlzEUSB3Hz2HxQJfPpW4yfQyt1WQK3xgsPUnijpcuOn90Npbn9WgVwOLDYglKQIrOUhQK1H5nPE4t6UHqeez3vJ7bmH9ImKZQB+kBpYI6RfTV8dwzxiSwedvtgmBvErlabTKBn7E5Ykxn37OZTUPRwAeFl/aNU9nzjQArpGsKwO2uDUI0VH8HJIXDN0bVXhUux5x44gp6ULKfff+Exlln92i7HDwccUVwW1JbbhEW5cSWUzRq/bNkWPha0CONgdlG3M3bS5og6IWpf5RLl2WK7+yYZpym46HX+Q0QI6uFhAGlMBMgDKchhlHiFK9sz4u9UnArnMa5e9zBGJCcK+9rwMecSGBWXXNEgpS6WGEqFxm1gxJmRgp3zAiIBYgkWWYRHMTNSkw2/frY/xsorvjspZtlJI0lxrfsJ1esWcg8G30j9rDJQ+2BaTo60YgE/yzQ6QqTIoTJVFUho54Fdmr9A65A67liehKLR30mZKue36Uc8pgAJplvqwXjiyb8aZajQtNMxM99zp9+FrMdOm+xfXStVoGxqM0+vdAbIuFjcr1o+5VtZjbaG2gRT6oVpp67jcos/uOXYrSVwggrFPV8ndpwvI+Q/MQZngaqm3nYa2w2ESC5cdTLowJk7g4mVE/kRJgNI0lVR+6NE7nYju3ICMCnUNhEB6//uAo9MSouFDbBZKs08fbgi0OPMSr2EA0r996EDFl0/nXBdNuUsPSFcyZk+tzeZIhznNqHtsPxKVprhUegoj5KDZFtTkuqqXcFabOwRILBso1m2pvzFPV8ndpwvI+Q/MQZngaqm3nYa2w2ESC5cdTLowJk7g9qI+UQAsh85qWrBw/GFQxrM7YBTaBNnC94pyrjFn8BB8ndeCYdbP9YGI/KOdjMvBHLVhPmGjRQRrIIy3MmKxzy3EcrXz5V+eXiezt+Jc8xaMtkbE7mjZehayHPh0xJcrU7MwNGl1C3/frzllcLbf+wrvjspZtlJI0lxrfsJ1esWcg8G30j9rDJQ+2BaTo60Yrn1acRPxc+DUH2hGD/dt7eGiSuDjyc2AiqV9KKDsTs/Kue36Uc8pgAJplvqwXjiyb8aZajQtNMxM99zp9+FrMfdiSX9hKFyg4zS+hmHHWRpZssTyEVL5TTtlsoH/whgIMmnQMK1qz1B1OK5BQZnoFZhojrjw0SRme2V9U6TcIHxP4zoCrClFLVEfr8Et/wmQDS/wYQO4CT9CSXHaqDcuzRGuAN/OIv7PVrKQLPzqWwV8HH7qag6B6SDkFcHoIQUw4L1OrRB3gJV0p4mERK0S6/TvY4yX9jOSEB9+UMzgPit7jI7KdsEHLl1XdUK/7nNDle12b/Mz0DN3u16IeGf06szN5jdCNLM+YG/MLGo+L1RJSAt5ZsUUQn5PV3kLNCjroZy2vwXZ1FzYKQqDyFmBUobBroYYM4zEsC79aC/6US+5qT0nhShad9E4bINs6SaSy86pHL+YmsMMS7cqDzJ+fqVnb7f9uHvdI602jrNBUSjWHpZideT66fr0pvqy6EmY5OEIOPyU511Y1offfcJnt6J9IbD3eRHP69yVTj3LuDr3hfvdQOf6vL1r00lDJZp/0mGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagLthezFjqpzhhqAoQZ40M4tjCX9r7BFQpeCuAfvc0gPhJhhpu7vEkYpNGAuqciOYVwOj1BO2PSJesb2Nw/aBWoIhqgpoOWa77r9cwrpkSewaaASHoYiIX6C3fKO1Ib90aDF1kIWLIGPUjpcC30HKRh1Isnvg3AAlg87q7sgwtfLm4kHKWytIJhafCJ+dehTyqBM6VmPm+Y9iff+B/aLh3fXrTeZbs7zzHVvp07zAS2F4LY+hIzB7FXnpT31tpqcbbLzqkcv5iawwxLtyoPMn5+vOn2s1kQJaRCb//+H3gwxs5StHhe9nI7/YPllB7n+v4T9WeHjDD73j77R+ZqqCQxRQ7DJCc9v9Vu1xUjkvinMLVVhlaD/RhoF72CucZ0HY0K4ZlGPF/5hdQRHCopHbfEBDoOrkEkjLNZDJGa2h+fG8O3Iknv88yZK4K2vOhXX86cIWQEFnWE4a1ResMhXQ40MhmowvYXZLW9Q0/N/puDd7EshXM/8Hkvwz47jtwtd96Vtpxo63u/fbhIPzy6CinfoUNc/FKsDq04Oz22UPnAaO36tMlnIoQbfY2l8/3DEmWBzAFYjzzFb3lzDlFVFaRx0VsN0digF2vkeOtLuzOMvKJW5l4fxU/eBPTcQx5ip5tio2n7F/8UttVfiR1GrFuyT+M6AqwpRS1RH6/BLf8JkCounIpmIP1/BnRh9k6eFTks2nf1pSFJy8xPuGsvSCTySuGZRjxf+YXUERwqKR23xCuS0uwQ6xpmUXujSGJK/sX7/6VJXJaBiD9cJzzkY+Q4Xond0lGq7+JsMKGOoRAsFhL3ypJ81C9AiFLX1cgICNiYkk3ucLis5WC4y4s24dCpkXbj0P4AZVSOzgt18NDqfdatwFUj/LCI+xZ7RF0+SIxncf5ld0V2Nez9fH0LNsIjqVokvWHj5t9WNgtegQfAASelhG3ySmLH4Y6RDraZ3a5BC0XAiDQyWjHW8pwhI8iI1cxnRiPB+N/+Ir8xgolz8mNZIhoUov3VT3zfG6/G+auPxkSjEKaQ8G2kel5MqUOrL1QTGIHx6MoWcHFD/XyExoRyR+0CC2eYE8Hf/VF+nemVh0izqvkfEwR6pVD0JaSBNOzOBibwy8LfbDCcJ6bfebBs7H8l2qx4v+iaFk/u6RrFTSGbYpGdOx1BTSuW/For1J3RdKTYANBB4rmbR1Nfux2AAuwaIk06GTSm6pVh12vgn62eokieuePJyiLiWvZXQr1k343lXAyxnekjnJFCypwniM4oveXZKdo9GtvqzlleYD+SGyd79vIgKZSbl930XpatThTcvQOYIWWyssKU3+LUHNNv5tYUYnloAMaDomWLVAd7SjEvN2EO4/vXCHWViiYNECy7Um0e4WtLu0pSqG/8nk2OXohhKNRT3NiyqHBJ17SNtzxJfHfCZxJAh1psLqW/1rUq7S7RffgDnBc7MNJhhpu7vEkYpNGAuqciOYVwOj1BO2PSJesb2Nw/aBWoKK9yDKl++trT52bEk16emFA1jInpXDznv9oHr7VtVyU7jI7KdsEHLl1XdUK/7nNDmMeo/bsxrna9xweVv5gxUifcCLr6Z5Qwwu3OIN3BQ7pJqR6fjV4bLYpSWPJv6Fj95h8YXPnV06sYjPMFEKcG+RNySgT7IwhpvjsK7CfTNN/7Pg08sIeDzdt4/kqFHdRBJ883ZrlifN/KDLKuok6GzFYm39YYV5K84G0UBP0TVM9boZlPxP3QW2b5lVq1gRnXPedSpXG4EKQYVWlUqa+sC8vOqRy/mJrDDEu3Kg8yfn6JkUA6kYStX0U7gRCyT0RRjhqKXmGUuR1Hp8YpHdcjd5YNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S3Z9dQwjLxUvsnFkVB8iSzzfK/Qwh7GXZnAxjaIIvuzjBh6G5vNhwdOQ/sToTU4Udn+xhNjwXdyEXML6cnhT0LWCwKj9Ps5bBvJubxFJXW7cgwbj+HUka1DU4dM90C9TrX+sMvk0CPwPKlKm+/oeqsLbdvQt/nzsy1/O6bVT9x+tErHnwH5a97rvqp5Y7TJpQP9NS9vN3sXWPKl4ZaQvq17lUZ5Osc1OgqU7y1Gt6wZ0uk0/v3FBO6jZHTYTFQipi4G+Du7ZJ//D3X1wROHI+IGbtIBxhzepnMmaFYS0vZjbUAray2R5PfAwcqzJSfEvKfb4jwKHlugzLJAL4k9ZA66CXMvdqlqFAqtTbyXVw0tufHrRrNzKyVXbSL9aHf38CYQYV2+035cR4qZYyHXeO8H/s05mk4KbAF6vwptTt4zVBiQCMmqAUW7skayK3jIrVF8jh0OVkGnj4JIw8nxkWE77/d1juyBOQHb8Q5/7NYlp/sHbPlvM2Azs+iCvSRL5BML12DDDt8QXVQoMvAYViOE8d8SW9X35hXUXFH4WGsXnhY5IYNrnKZppaHHtrTtLDXX3fgS1hxy395KaL44iOTmHe34yCTln/ZWnK8E+81mjAoNTcUF2MEUCLpHzEtWcI2thv5B7qAxmTlH7b2FEFCzfK/Qwh7GXZnAxjaIIvuzjyTqADHdYthVqcskr1eJJrj8BX+nEV1XREjhhwkK9pqhcWrSWFE1C2kfTJZVxJCgPSYYabu7xJGKTRgLqnIjmFcDo9QTtj0iXrG9jcP2gVqAmweYrEDRVNnchdTDcrMtnt907COuvnk7gpF2QHz3qtwxdZCFiyBj1I6XAt9BykYeZmyOryFH5AAykHVB4T893IORb6jaQH7PN+XaLVk4/Uy86pHL+YmsMMS7cqDzJ+fpSJQPDUfruTQCULGIz5E1MdoEn2fCtJ5spSKF05T8PmBIfaQi1nmEEAlJa96/A5w4EBWzN0O5l8wHeYqU3p17RtBfsexqnzihq+dOeZzfQtLa8lX6b++fLImJXUfWKwTYWRZFD5A4wg2AswPNNaHIdVtMNzR32BZSznKiZAJHeI+wiU6p4EgDKa+asU8j5GszPo/PeYa9hXJVqxijD4ZgzTdLPzB05rLY6gs9U60gkocNhV0bsEavPjc7JusV9tuh2AAuwaIk06GTSm6pVh12vhQn/lr7nglPR3Eb+BCtnPiHIlulnMLPWQ3QPK5/AvS+sadQLZPHsxtgd8IU0p1DeT5Dax/b+Ei7ycGSgwcsXooozRkoZe7PhC3r4W1ry4A6nsWTTe/yoA5ionEaQqCq1rtuJxl3/XHYJTAkioScYErw+EM9hTjzu/ra1r7Rijk/wcfupqDoHpIOQVweghBTDNA2YwdprpSRYbAQhRabD1Bvwes5mN3KZl91LnGehWsRYNssv63MLq96TzAFC1x5ij4kASygx5xnBnW1I/YjiQHlyEnTHAqnF6V2NtWxryt+fKu8jEst/dmBdtlF2EF0Heid3SUarv4mwwoY6hECwWCqhVX6E98NtxVy2pi/ikJz+74P9AtPPpVL2Bb2Wz8idqdXjKsBy4ig1N8edelZFBBleudZWyfeVXSNcYyQxmIsLMHRX5bdorW2PHVBP5EdHeid3SUarv4mwwoY6hECwWM+64bODqu/xyoutWHXze5gWKeDF/ao0xpdeTxL43cInqdXjKsBy4ig1N8edelZFBAjr/UGJ26LXHfpa/vin9stelJGu+C5//aU96KGM2zddcFbldc7Vmqriy21NGC0m/9VCPDZpY6xCSAoilHo0VZvQ024VKSV8Bl5Fr+J/fQlCYzCZQ5ka/ogHtUyJBqm4pz9Zvujn4iO8E13nw4U8d3hhKBx3uVzh/5nVPVafM36M9L1uZWiM3YM/2xcsPM/AuqnV4yrAcuIoNTfHnXpWRQQXzexm6YvPDc2lKq6pjlfdM+nWdDeQ+3HfQSZh/j7DEkmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFaglfJBZsTlL1vhUcI25xQDTWCcYmj76HpgIOm3+zhLrhM+ra8Of2BZPbW2jkPsJ6zptdPrv/1T/bIVIwiWCCtHoOimPm2OzaIVEaf/Z2ov3GUvOqRy/mJrDDEu3Kg8yfn6SW/fXKKj6JcEPcNP0tjISvm+OeEo4Ze7g4jIAOZs7dGp1eMqwHLiKDU3x516VkUEF83sZumLzw3NpSquqY5X3ar/xROrGxqwbj+WHQSNNXpPzRK1lSCYljpAL1vBiPe6AYIMJaadrcgr4bpZYiyeJYfZurZpEHURS/774D+I46os56EOD+FGLu8B0vzBNKawJ3Dt/3XNZI8e9gdyAQW4QRhG1Gn0885y65fDz4RbwbURkX8Hnsmu3Sk8ce2GWKJiPq2vDn9gWT21to5D7Ces6eY+Xmh6VwN4fO+wVmHcePa4rPxZkaRjO1QS9U6erXLti691Obc1hpZRV0yL0jDeTNL4dBmfk9MjhcNiwXAahXrZpgrdsZlLaaHCmf1xhHnHFdM/Xq3kDR9e31c8NoLno6aJ6W51EjmB8oo5eFw0lkHs7QrsvTUrNGTeWo0NekupurmAdtOsTcpzHdAkyBfX+sVZOTxUn5JPJjaSvEW1vcB9csnKhk9iAkx8fJlO87oo3yv0MIexl2ZwMY2iCL7s48k6gAx3WLYVanLJK9XiSa64p/FMzsCT/OgseBTJRk84WgB8KIlZsICKQHZvzi/XZK371UI4wQqemONKg5v9RMnOLN7FhJ17tUD8lYp8PoiEZ3YPAJpAzV4IdgAopchrp3ond0lGq7+JsMKGOoRAsFhELMMUmWafr2e0ScV61sbynM/gnmw5BFf6KgVXKI0katQPQSx4MUAPgJN3lT0MCtBCuT9QJ8r0pawu0Hxd1YXiZVT2kMCwNA8ykdZ19l5iI98r9DCHsZdmcDGNogi+7OOjsHg6jZ4gUlFbdGa4KR7XrOr1ATsOkjsDS0uM5YSkrQEFRVdoqAjSSscAYIlFglrVZ6o9g9PtmwJRZIl5QQHFN1O4Y8NE4ME8sQt4LZF6oW9u4dPlwjaL6U3Ptcfj5oRJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRcFCaSpmjMrHCUkdSuIM+rWI1grkuTJSd8qMWJXwbwPc7CJTqngSAMpr5qxTyPkazGuKXAtO5kqEmtSyovlGS7eMCm2zO1791h2lk8Ab1BP5LlaevUMtlKVxuy7KlIHxIG6Mmi1adQk+1ubHTByNyn4w3g6rBkUQ5dJ1R6rd5+mss+uAH8Po4U3OzPQV6sn0Tod7fjIJOWf9lacrwT7zWaOH8EQbRSSiMtQuw6b9qosQke/4234+cjilFW+OGji7hdVCPDZpY6xCSAoilHo0VZvDhWbhp8fDe9JoAk/WtmxZZpQMty0BKbsZE3GK0Gvcl8N3IRTgtys+dStq+haDyXYcsnsUnmuKJhKMew3M6SMyReaI/tLW/xRJ6Rhoww/5BvBx+6moOgekg5BXB6CEFMNalosR2/HmGWOPyzh02JlLHv52VepHrazov2qOKXrQK1g2yy/rcwur3pPMAULXHmL6NFcnG9gj6gOgeyhnaJmejIhExj0GH6IEIfyvratKR6nV4yrAcuIoNTfHnXpWRQSIm5VbAsb9QdbGMLtbWP7L/fVMyj4r3/ibvCjCwTgQOanV4yrAcuIoNTfHnXpWRQRRNkK9roQLqZ127U8lWU+brKiiKNXPpFz7LitTJ2YdSdOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXi9mvg7HjCB2ZB729fJ1PPksCKab2A92zS12JHXESLWImRu1fbP3wuEvshpN5eFal5VJ191zeeirL+OSvgPRwXfCsH4i4/wgBkRN7taXQcyO3l00jTwnn3RAjpbze0Q9C3rZ7SPziPqh+jUA0mQvFrdgXpGBPHvEoJBydh4ymSggM2Xetcu2rP8ONkVwiZz76C627AzFDi6o2Ij4KSS7KpZYV3RXFuQSg6vgCkpfenTFs7+XYDiTEMmivA7ps9V1ibtbJXADbApfzkcFmWaeaKjJiNISfDeaPwpzNqdLR6wk2s3LLJTHiSNXKB/CZN8+Iet9Px+FiG+rVVo4ua5SUFmp99iAMAGTUDVkty/bdA6sZ6J3dJRqu/ibDChjqEQLBYD6X66m5s20rGYTd/U8ACRX5u3tpyefS71kJ80z4EovnQmPfE5dP2GOrJAH2VZWT7ZyS4odSoTaWlTEs4nQAioc99IZ+gG9+/nEgVTMmzcCHPW/spACLEbKQKwqHco94gP1m+6OfiI7wTXefDhTx3eOzkFQqjFZyDbtPzHdMAuxT45O083fRksVOOc51uHAov07M4GJvDLwt9sMJwnpt95sGzsfyXarHi/6JoWT+7pGsK/1EtuPt+UWc6+tW7VqAD8iKZwGDpHqgkbtGcaQGeD6pvuFU5wHJanJB9KLBPjFxDJA1d0k/6TPw2aLUX4EM+aAlwU5aAZ/7lODfs+ngCGcNhV0bsEavPjc7JusV9tuh2AAuwaIk06GTSm6pVh12vYjUmOiXrtz0fN8hS6FouApxD1JxMIPqOB5uRAkwfKiSMFSRsYvNzU1uFSv7zAPmb39N/Hxd/YWJtzMBPqKuv6wBJLFcyYZ1R/LGlVAJahhvl00jTwnn3RAjpbze0Q9C3dkXaqiN5T+ny9idbhUsz50yd3VK46bpXLVlZ0hX2WXep1eMqwHLiKDU3x516VkUEhqMWK16f7TQ2fYmuln0XKZPdbW57chfrEpJWC9aFT3iSuP6KoZkjPb3tNm392BoRIBjc2ejAjkZ2VIZFdk8TOFp4Yhlcz4gpKWSi3fDQIq3riHKRD+FFSU4fS+gx/48MTTh291hchHIsDHAUwyUXLptj7mfguKp7KrheFGFciAgxdy746+oKT/tc1tAVO7jkyQ6rZkpPvrTrFWaY33g8AZcZAfhxWqgjJIGO1HqABGP7ZjpJgyzVEZwbjW6CQ7mgpGDk3Sb8/rflSc38jjsugCFnYhbbDN+Xgp7fAAxUvUzi7owc0UwQw3ymspe38Pb2WDbLL+tzC6vek8wBQtceYuXcWg26dlY9N4QoZU44thon9TtVOFcyp2GPW5JxKwYC14DpHO2fEAhcVZBu3L+FI4SSFVTlKL+Wzj4fU5qZPGfDDMek6ytjuXpM1q9Gt/NWsR6j1uE2bnUBGahEZWrcXz/4ktpwNnB+mE/NaoOtbjlWGLta9TPF4IdzJRPxJJGlxYXBUOH8f6bMB9fAHx1Cd4d7fjIJOWf9lacrwT7zWaNsAn5oGRV4GqC+O/x4HODLL264F/dnXU3z1yslcgIcNT6trw5/YFk9tbaOQ+wnrOnxTPZg34RTNZl4vNoJY6rkg7pdwrUERgUqwtNaO2VpMD+M6AqwpRS1RH6/BLf8JkBpWjsRqGIbbsAXJvDnUEjwTDnA+RKDimLFCd8fH0ABAHond0lGq7+JsMKGOoRAsFiw7AmDN0E9UHQ5QfvyWn7Dk+sZiprK5471IeSrsaFl3qnV4yrAcuIoNTfHnXpWRQRWJQd34ADBEAoIFTZgVfWEqQbd5jPt075I0D9wwcMOU0rDIdiIp6QqeiscJjcPOthnEtl3Y3ingn4sWwt36+ECeLptOsFvqofgkFDxbeT+OIdp2PjAAPM3nFiDGG6GuozTszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4n1CICjCbbiuGetm2xymaPhYNssv63MLq96TzAFC1x5i5dxaDbp2Vj03hChlTji2Gu2HmKCKE/BPTBntN1bNzYuPWCSkFAM9TMQTXed3sDDImHxhc+dXTqxiM8wUQpwb5DEK1c/NqFwX1VE1btmanVCybqyB2CHiuv6XrhfmeA3X3cP5ZN5AaSWzdSohjJB9rr5INvXtVloFDKVlNsvs4vL2mUs0UlDlgxWLQsITyTP5kbtX2z98LhL7IaTeXhWpefPyRVuqytdA63QCTrIC25qXNtrVNCnl5t92wu2c01Z6K4ZlGPF/5hdQRHCopHbfEPLroy0ptZu4VKVAu5vnbhPEVu532uO5Qru9XAxiKYzoAf1e7SlQOf+ngNzhzs0TwppTQ9u+eoDa/t3Yf2ggsByBTcJl2iVGAg7MjAaXZCEuqdXjKsBy4ig1N8edelZFBPp1y+rfEZ3d3FoPOqGN03irrVXRc2VZqgpXV2wSfgIDSYYabu7xJGKTRgLqnIjmFcDo9QTtj0iXrG9jcP2gVqAMWrl5Ko+trBdjfuu8IDU4WwLmjeEz0Wud8GMYHbqliXj8osuw+4E6yOpUTt7kM/9zdYkFakaqXRaZnM6Ss2JEjONdIsoIPE3+MOyANeMl5S86pHL+YmsMMS7cqDzJ+fpBq4zMsxbwkl+TuixUcH/IOcYv/Ow24hj60irs6dV8Zj9Zvujn4iO8E13nw4U8d3jhlLG+PcnvdNwTLgAhoqBI6rH0WS7kFMproxLotcDCbPBx+6moOgekg5BXB6CEFMPenyYD0RxsUqHDqjW2trcq8ybfVhDePIrkuycv8IdZD5U0ZINfqGx7tJOy6WU26tMmsUkqbKdZc10UXEfpIeURJGIUsgVjNPeDDgCkYNHPslg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jot94OfeR7pMKl3WEWDH9tJ883ZrlifN/KDLKuok6GzGYfGFz51dOrGIzzBRCnBvkOKRhXGD2ELcg4wF2/0miy8UwQnDBvzJYAsOLJxOAo3CBvg7u2Sf/w919cEThyPiBX5tQOL2AnKriRyxhNuzLVDLbefYdrf8wsCloPei9kHTJDqtmSk++tOsVZpjfeDwBY+/BvdqzQNWxhMs5jrlDKRXkZR22iyB+1wlQExdB7FNJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRTlVCGsO/wmk2SXvS6NEssSfPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuWW0IuQ3SauD94eCBagfZ+5yms5SJOIKXqjq51kjA5Y+WEhWJPcXkDZsDCfDYN2XmpqZEK+DIEGs1zGXE2e49F6e+ICMF5dUXyX89/EXpUYc8HH7qag6B6SDkFcHoIQUw71N8eRep2zgf7H/Ep8h3jPRHStrKInypf7lSXMlJRCI1A9BLHgxQA+Ak3eVPQwK0EK5P1AnyvSlrC7QfF3VheL1MOkPrJ3P1GJnF+PuQqNA3yv0MIexl2ZwMY2iCL7s4zQh8e7TQ8WhK9uS07MmryDSYh0JTZ+/Ka62LVBty33dsR6j1uE2bnUBGahEZWrcX5G7V9s/fC4S+yGk3l4VqXnfIM87F+zSAna9E8eBsoz/Np4hMWuIhy+yhwcSftaQQ5sGdNE1rfKSF5pPx0l8iv9WThhObKIJL0zO/5Tj0NEJyA9fXCin8WHCbPDLM7IriqpvuFU5wHJanJB9KLBPjFyIOtPipp4oDbb8HImJA9sr7fSSQe6qqv20ElGl8chR62BFQb3LKlxfS5FBKBSEuMCDKSGs2r/6qjGOPO3xcDCyaQPs8Kei7qnp1852+7bsPfRNagwJR0AOJ4U2ZNP7Oe7dw/lk3kBpJbN1KiGMkH2uWUKkV7zn548cSGnesJAg32+fizIeUBAKaglIl3h5OJzG3rgn8nixsN/JeBWbsUy5LfLnxh2umcrn8ZFNH6dlX+jQ3WqYi8Ggp3eab80kXZ+bBnTRNa3ykheaT8dJfIr/jo2FWjCMHw2tHwiHUoOwCXgW2pGrM9xvYf1Qfe3GoMF6J3dJRqu/ibDChjqEQLBY6sgORUhmdFpBKX95HK7PIw7JdJ5XaG2+K+4ObZKNCyxYNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S1m+D6wPKM2nVK3SeXLAYZsmpHp+NXhstilJY8m/oWP3EHyx7QmYpj3LkQ8DSSJPuRcxN2JM5//dWs9A0mxr9hmc+5TDJIuK7XOZeM5Ea3dZrugR6LxsSz6yB5oQKZn+ppWuutn4/WUJxaolkPK2ikfH2l04QjIcCwfQMveUhW9eZxLZd2N4p4J+LFsLd+vhAlhwibSDjRefa+EGrWudZ1h/Xk4/6F1sgk28+cQGUDkXGnQvxs3hBKUnHs1G664Ty2YIDkNWTQbBAuvkO9DxaFReUQ1kBOF6y+R0L6CZI58x9EJEvNnNI981+g7Vxkd4SSGJZ/8KVo1n5WYDSStmzESnvVPFeu32xvkHIFMc/V9ReeCIX/JfGeKEuHR2lUpMK4AOkIsEnVWHLuw4TeLj9jd51DEuhwCVt/Kr1uvMmgmZT6Ogz77JeeQQZRpNqjXfEuPgsCMn9iXR2ovxbjs2qLRMH20NvPdLrMf82zJPSQTGcgjBeKADcYUn24I9SzYlbrRIgvCoUTi4UkJiLOsTgTWbGMhKVl89wX2/NRTdmdegsXp0/GEQmJZQhvyvwyW5RbRIgvCoUTi4UkJiLOsTgTX8/OEaq3yMyTb/0KV8FZhaTdkQl8OTv7LC9z1DuA+XNK90aGjLr4y6QjLGpkU+hRBmB7Io+Ga8r+wPnCcpkePOFFi4uiqvWsSpX7Yb5Kgbwf+o2QDTjcnM87Ci8imItM9hAJ12c8YNGkZhEE4A8dKjsdM/NP26UoqfyTH2W2NcsXond0lGq7+JsMKGOoRAsFiEGcX2gq7Mez9OtMrOKp5zKmYBT70VEW5ujSTxxPHw5Fg2yy/rcwur3pPMAULXHmJucKI955VRowBUTVNAVLdLhHYWnHNFpsZecv8LOPaA1mkiHDBvrrp/Zygffskv/cB6J3dJRqu/ibDChjqEQLBYjyksNHlQ+PGogXfuHoYAQjuY4s21auEmk0hci7ClyFGp1eMqwHLiKDU3x516VkUEqnP6segBb2b0gy9g+wwi9nrAV3ZdSzIYUMlhap+bLnbTszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4mNyWlaZHKJWvbt57rdN55OfPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuUDvoJjJcADsHFVVX2WWVfHRjUQU9yjRsCzaFkv37+aw1WeqPYPT7ZsCUWSJeUEBxThgtJKpGb4ZOCN/jdy1H1OHqm6t4v+zZAKrBvfWmXvFv5zwBS3jzC3aa3lt0kNsaLzxghVHSijjyHe6XLOjDgG6X7xJi0YDJrvDHRSU35bI1WeqPYPT7ZsCUWSJeUEBxVqfvMBRaZDzlGguyXh9bdPHmriduh7L6arPJPotKpEkkp2KexA6RGrMPnpo33AdvaMOTzvJVVUBCtV8F85kQeMcozw5akaR4x2RJdlcIBxLWEhWJPcXkDZsDCfDYN2XmgifEps9esGqvBJ3zlvcBST3Q21YNYY/n/Vyowj5SXCeWDbLL+tzC6vek8wBQtceYm5woj3nlVGjAFRNU0BUt0t1z9hno26uo0UcyH08Kpymn0ZsSmOcV/lMtIafDpl8a09rxtGx3D3hM4T8qnRtqWu8HvDDemtcbjbTjRYSb6HRs/Eo7bpxt6GfpUNtcMQdnh7UE+/TRr7m/hKciQeSLeH7FL8zM6sBzBWW1hqV/VEgXXyuN6vs35PA4gS1mNilcsktTNhUIupQLLl4P9FMxgOnV6UhMjJmSqoBnDR6YX60EmHZ+mxMXte+VaMbfRii9FhIViT3F5A2bAwnw2Ddl5rAN45+qzJcAJ9nTVUHAFaiIkw1XQaPmpmqseUTDVptIlg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jc/aK0yc+tIytQYc8vfD0GV/ZGJvJc+OWfBHE0MJiZBI0LC/nLol5nBY26xVrN74iOPPcA93P3+k5uqpUqRNrqnUcD9HvpVjRYapqLI0DTqphyUyXsWI70q3Ofl35B2BnMkFRgNEo1VUOIW2MrjGtlIBfswGSPH2OmUwwIcX0DldVy0HBuNL2ic1rsY9VOA7UVU5NyyPRNLn0KED6Zd8syexqvmj3QQJ6O3RaUH2pvKTwcfupqDoHpIOQVweghBTDR6hNi2BOT3YrpV+76yFu/2ai1JYN5b61edu+2Dp8jRRKwyHYiKekKnorHCY3DzrYZxLZd2N4p4J+LFsLd+vhAtcu/P+HVTuF0g1zr5311WKpScJ5zTo1TYXB/eXetCW1Z8D0Duz5hmJEk+rSK2/4g9xii0NZ9mipCtVwCTmh7fn/oe/8D1Q0IUq+YLI8cvCXZBj5Ou76nhfqHn6rbPyEVjuT5HKTqBiTD4Vs8gJs9MC9fGWoXBHl96sjkcyj6KOacSnueLHqtlV9ccolWckY8HDSLsCcrCEQqDis+Hszc9bNIlXw4aibe/LOSslOU2rATs8N0haSnu8bEgU69VfwPnGFRGNc0JlTbXVwpqLTAHoqeVpmLF1hFcPIwcseBAXunzzdmuWJ838oMsq6iTobMRB8se0JmKY9y5EPA0kiT7l3Y9dly3wer1ignmXxb04gTA+H13//sBYgzGFUId3IeYRDMPIZqnjUrEGi+zG/jOeqRzfr7oUSVP0JUgKvqPp7VSs1fPRisMHD0V87DAyB1nnuirFR2H5T7hL5eY1KwQ4JhvF5FUngU0hEZOpuqNbRuI/O3mAwmbtNVcHKzz9AKwxdZCFiyBj1I6XAt9BykYc7ko/9QeYh950Asjbw0b2OB2TVycFQT+RJPH4ZeQmH0+XTSNPCefdECOlvN7RD0LeHvJdFl7JxfkHXNihNKJTDI1gACBhoDYE0SOaYbcunrleQICjhzKyO7mR+m9T64Wy3uLUAm1PpgqRZNd3NOazrkECDeb5awR4Lb3p6sNzExYcpGk6hf6cavv7o+6jER5z99w8DiKEVA+OCVDtE0C8O9oZ7kFdOttCLtIM6aEgeAvDxvgFTacO23OzTwNZAkEmsZw0sYLYmurOcd3sXJQELdKGdGHCYal03BkqfOy2LBvBx+6moOgekg5BXB6CEFMOpp/TFl61glswcqHBosNF3jEubhNWB9ay8NLWkesQzWFg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jymbTlVP4MK96fhMk3DPOjkmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagXx27OTpnG7janVcywfPD4pYxzjnfJ9IZSAIYWUgBCbCHe34yCTln/ZWnK8E+81mjcq7rzfiId5MJW0QxBiRCY8efozJigcB8XXVQ9w2pyxRBrVYJiQ9GhWTwbmnwyCGHX3SEBzNUXi3MlfkkHgXNw+Zqo3WQUE2oIFvzwYpyqmVnE8ek1Lo1gV5UbK5E0ovrq6l3uGe//vUWHdERNnZRX3z07zCu8zLfCgsamVeTp98LXq1gnhKfzzaVZbnrMcG6zg17ZLR0Gt/KyRgkJyTe4kJqZ1W5/U9swfFzrOIaJ4j2kOpBL1FjnCiBa+Rwi7JIUc9BnrVO7eSuRV7knwg/DAk31tHhnU8DecdgCHqnmTYKBFa3ajH2NjR+68g6BstPFucdZYiMaRFZFxcAYNSWJ6EImAArnCqRUj9803pO9DIcx4Rz+87Q/KrRXppLGDSg6xrXOGcTJYnAZ/cqgbGaEr2WVzJOt/n6ymQcoeNz4aWhd0muTlf7oCiXALIcHwvOQa1WCYkPRoVk8G5p8Mghh190hAczVF4tzJX5JB4FzcP5aOsamSmhC7CWooksIHrpFdM/Xq3kDR9e31c8NoLnozZaI2MaTzcJfnnZoRpUfby9V5CvTJjL6hDVREv2bbDAT6Ogz77JeeQQZRpNqjXfEpoz2pQ2uu2HF/MVAWQC6sWjMF1/9+t6HscdBKkId1Ex+bw3X+m0oEHHL5pCZwhvbFSNnto8c5dBJvKdBn4f7zqDdrbi+wUgJU3x5CfRhs92yBMHP+kiATnw6OtqFNa9b4+ZP02KzsYLTBoVbHIFquhyuPtZ+0yBJh4aliXBJOVAXj/AUbbw3uLcZQ9I+2Ew1nMYxcoA2477b45QZT8QYDR6BuCjccoMnveoUnsZlEVEkvCBoYoTI0HeNSd+RSq9gK5YV8WnFLi+uQdJryujkqiTUnKVKZVg9P4vNgXf6vc8habbjawGGX2diEVBfoiluTKRzFtHADYFu6ufdNnsoJ2wA43yVc2aL50xKY5tRGZ3NHJzPf++reaGGBJRQ7/HAUgHlhQsH8uAaxnEe3K4kHgqOJsIQt9hdn+UaBqoXoAfyhhr/xqhs8eY+R2IAjvPmOH/9ypo0GVrrkHCm+9iiS+pVKQ5QaalEcoafAoPP4N81+a15ktjrFJLT1YCA9z/C2UQAYmfFuHcKZu1iHU0MInIb340EKmLaU0weEAbZZ+C71024z8TZ4dOkynbSxv9lJ/ZCbV7sSh0OAkYfGZjzm9nE8ek1Lo1gV5UbK5E0ovrq6l3uGe//vUWHdERNnZRX0Rx6U/lWFUkzwuH94Q2jGgV0z9ereQNH17fVzw2guejNlojYxpPNwl+edmhGlR9vAdm2TVApbe5xEnuLbEJvJ4ykcxbRwA2Bburn3TZ7KCdsAON8lXNmi+dMSmObURmdzHxmPg+vC5AFaGyBYc+Ifz2kOpBL1FjnCiBa+Rwi7JIUc9BnrVO7eSuRV7knwg/DEwOQT7XUZ9wHArtxEaCUOz5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOgFWn9oGxTkGReuxNH0wKLzT7LrgsLw6eauA60gofG6BqVSkOUGmpRHKGnwKDz+DfNfmteZLY6xSS09WAgPc/wvpYjtNK57gKIjHAmjnq36+6xrXOGcTJYnAZ/cqgbGaEr2WVzJOt/n6ymQcoeNz4aUUnOrtm6dI7RO1/mrCSza9cxjFygDbjvtvjlBlPxBgNHoG4KNxygye96hSexmURUTp5AmphQ0Lfat/xdhkmq2+ZxPHpNS6NYFeVGyuRNKL66upd7hnv/71Fh3RETZ2UV9Bb4yX+3A0ln+5xt3r1KdgFdM/Xq3kDR9e31c8NoLnozZaI2MaTzcJfnnZoRpUfbycJX3H3EXeHFTJDkwaLbNLMpHMW0cANgW7q5902eygnbADjfJVzZovnTEpjm1EZnfH+O9iq6fGGYd4nQtLnKJt9pDqQS9RY5wogWvkcIuySFHPQZ61Tu3krkVe5J8IPwyBpfy77k7eBoDI6auMywzn+bw3X+m0oEHHL5pCZwhvbFSNnto8c5dBJvKdBn4f7zr3YcePUV7fkQC892YJnrd8GRkvzNILH25cD9rjJyDsrKlUpDlBpqURyhp8Cg8/g3zX5rXmS2OsUktPVgID3P8LzzX9yXig3F2T1/wFEDhr4usa1zhnEyWJwGf3KoGxmhK9llcyTrf5+spkHKHjc+GlbuWMlzbgpXSaysAbzJN2xHMYxcoA2477b45QZT8QYDR6BuCjccoMnveoUnsZlEVEnGXvtB7UM0R3INkILFyymmcTx6TUujWBXlRsrkTSi+urqXe4Z7/+9RYd0RE2dlFflHS0Crf0du2RVaQyJ8qS6hXTP16t5A0fXt9XPDaC56M2WiNjGk83CX552aEaVH28+NUN+3zlnXZqYz7DFf0CFzKRzFtHADYFu6ufdNnsoJ2wA43yVc2aL50xKY5tRGZ3XPH47TGAgG9+lOXSonqA9PaQ6kEvUWOcKIFr5HCLskhRz0GetU7t5K5FXuSfCD8Md3ffjllKKuYk+Nzz8bJlrPm8N1/ptKBBxy+aQmcIb2xUjZ7aPHOXQSbynQZ+H+8692HHj1Fe35EAvPdmCZ63fAY8szjpGDwxU18QkYA8SUepVKQ5QaalEcoafAoPP4N8urbx6yLQIHES4PJp9tWhLP8e6gUr8orv8sB9yyu3Q2TrGtc4ZxMlicBn9yqBsZoSK5dT6sUtGnqn7iKHTXBfjonVLAIsNB15FWGjSBTltclBrVYJiQ9GhWTwbmnwyCGHFPkJwgawqdE9nykgZNwlc0jdkBO+d9RN6hAgbL5KltQV0z9ereQNH17fVzw2guejhSJAFKIz30akzCLuYNbNhEYMPIuyNPoEaNGUiBVLyChPo6DPvsl55BBlGk2qNd8SmjPalDa67YcX8xUBZALqxVXCJ1EgXPomshGwnQC7t7f5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOnn3Q4in6vAuf9kuuW2E4XpMgbOQYPL7SZxz2lC6Zp8cj5k/TYrOxgtMGhVscgWq6P2dObxPEL20EWLHryOU975+IvTwEFA8jUoCw+hcASUacxjFygDbjvtvjlBlPxBgNHoG4KNxygye96hSexmURUSd2UoNXpqLayt6ory/ew/IrlhXxacUuL65B0mvK6OSqJNScpUplWD0/i82Bd/q9zy4E3d2l9pktoblj04wrIJDMpHMW0cANgW7q5902eygnbADjfJVzZovnTEpjm1EZnfQnw3bWU3XoI1pyZzlKbkUSAeWFCwfy4BrGcR7criQeCo4mwhC32F2f5RoGqhegB+DPJEyM8uPJsHEGe9EQvVWBjyzOOkYPDFTXxCRgDxJR6lUpDlBpqURyhp8Cg8/g3zyv7EK//47S3w3D+9mbZP8s+grRdRFjHc7h802OKRtLchvfjQQqYtpTTB4QBtln4KX9PmnQEqrsCdsswOa83oQFyZyENehc5r3+3gxI23GAWcTx6TUujWBXlRsrkTSi+urqXe4Z7/+9RYd0RE2dlFfcqor6rSTLXNDnMA2/uFzwQterWCeEp/PNpVluesxwbpXhi/iEvZlKlUUjLa9ohaRRU5xTlnjCGx/VyBBZaGmgfaQ6kEvUWOcKIFr5HCLskhRz0GetU7t5K5FXuSfCD8MyejWe3o8iCHS2ShomW+KUQoEVrdqMfY2NH7ryDoGy08W5x1liIxpEVkXFwBg1JYnHO360rvM0v5bW58HpdjD2nStnNtbzknZ+PwbMQEAKfTrGtc4ZxMlicBn9yqBsZoSvZZXMk63+frKZByh43Phpci9zC95s7VgN7IlfLwBLS1BrVYJiQ9GhWTwbmnwyCGHX3SEBzNUXi3MlfkkHgXNw9RNjAjBUlZgBZW7yd+yFz0V0z9ereQNH17fVzw2guejNlojYxpPNwl+edmhGlR9vJPn/bueNJUYnqN2lszIHGJPo6DPvsl55BBlGk2qNd8SmjPalDa67YcX8xUBZALqxSXYNDAyKvIfYAYHqzyu3Mv5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOtvwKnl+gIxsD+ihdPz4wvx5z1387MMjWPBh/R9HhvfXj5k/TYrOxgtMGhVscgWq6HK4+1n7TIEmHhqWJcEk5UANweoWeH4ZhKOCFduub+abcxjFygDbjvtvjlBlPxBgNHoG4KNxygye96hSexmURUSmC7eQfqciAveqWKDOFnS4rlhXxacUuL65B0mvK6OSqJNScpUplWD0/i82Bd/q9zzfWjM+M3KH8JwQVWeF8tP1MpHMW0cANgW7q5902eygnbADjfJVzZovnTEpjm1EZneqFokwJBfiskgpoNUPXPIJSAeWFCwfy4BrGcR7criQeCo4mwhC32F2f5RoGqhegB/MUX9UN9TkymBj/9jKgb3bBjyzOOkYPDFTXxCRgDxJR6lUpDlBpqURyhp8Cg8/g3zyv7EK//47S3w3D+9mbZP8DWIMSYECByDSNAShpgSPm8hvfjQQqYtpTTB4QBtln4LvXTbjPxNnh06TKdtLG/2UQ2iL47Gk0vbm+7HQaYXTJGcTx6TUujWBXlRsrkTSi+urqXe4Z7/+9RYd0RE2dlFfiR5rKb/G6KhImwzGjHos9wterWCeEp/PNpVluesxwbpXhi/iEvZlKlUUjLa9ohaRCMq9pdj9+WOAyxwRwlBA3vaQ6kEvUWOcKIFr5HCLskhRz0GetU7t5K5FXuSfCD8MFdBlWVRJdreS2j7yGXdxPwoEVrdqMfY2NH7ryDoGy08W5x1liIxpEVkXFwBg1JYnSMX0/hxj77jIvp2tvIHMrfdTQks+cytcS3apzymhShnrGtc4ZxMlicBn9yqBsZoS13csJ9VWe+YryIzwr5Rs5ilnjM+Rdy5QPZMjzureA+pBrVYJiQ9GhWTwbmnwyCGHx5vfdnyiOmUVmd12k2b6eay43aiT9VkXHbGg8ggFGIIV0z9ereQNH17fVzw2guejNlojYxpPNwl+edmhGlR9vH1GESZwKYyM5yeKybvmtKlPo6DPvsl55BBlGk2qNd8SmjPalDa67YcX8xUBZALqxbJYuL+lzgeq6lZ9KynfTiT5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOgesqk+5OgfQC+sG6XrSfPKTcTlTaxJx0lLekiqFKGe1j5k/TYrOxgtMGhVscgWq6HK4+1n7TIEmHhqWJcEk5UC7slRumHuGN7Nab2exiG/ecxjFygDbjvtvjlBlPxBgNHoG4KNxygye96hSexmURUTBNylVdyztawE9mrazuIGbrlhXxacUuL65B0mvK6OSqJNScpUplWD0/i82Bd/q9zwzMOKKnBDbTA2+m8O1vxiAMpHMW0cANgW7q5902eygnbADjfJVzZovnTEpjm1EZnfmIONQ4ImA4n8XcT8aIQBoSAeWFCwfy4BrGcR7criQeCo4mwhC32F2f5RoGqhegB/29UWBZnDBM7XJrdn+tcP+skQCa9FTNrVLwkgw5FNdk6lUpDlBpqURyhp8Cg8/g3zyv7EK//47S3w3D+9mbZP8FjZJBXgrI7RqSSfN4ysBT8hvfjQQqYtpTTB4QBtln4LvXTbjPxNnh06TKdtLG/2UmO8YBK3wD6MLP5Ju4g8X/WcTx6TUujWBXlRsrkTSi+urqXe4Z7/+9RYd0RE2dlFfsqocv9a8qP5PLPLPfLCJFwterWCeEp/PNpVluesxwbpXhi/iEvZlKlUUjLa9ohaRJwQ/V69MYcZQjiToI9eD6vaQ6kEvUWOcKIFr5HCLskhRz0GetU7t5K5FXuSfCD8MiMMc04838mOWz/7rx0k4Wk09/lq5YrOEvc42QxTCPcIW5x1liIxpEVkXFwBg1JYnNwxe+UEbd5urgm4+n7ciilDw68QLbMUOsFYUpNCalcPrGtc4ZxMlicBn9yqBsZoSlxiIPODAvjnfQsukyHBqENUEW/kqIerdzClqKnBRuEbIb340EKmLaU0weEAbZZ+C71024z8TZ4dOkynbSxv9lJxQfNs6KJGD6to0P2588AhzGMXKANuO+2+OUGU/EGA0egbgo3HKDJ73qFJ7GZRFRHnu+Vy8HJl9bciK/9UZDO9BrVYJiQ9GhWTwbmnwyCGH7fZcjhy4asDS5+kPFlWEFfsQ8LzXyq+lxE2OhA20/a9nE8ek1Lo1gV5UbK5E0ovrMO3OodMJivsqK3eY30CAKS47m42E9xdIkaO/3uxD9lCuWFfFpxS4vrkHSa8ro5Kok1JylSmVYPT+LzYF3+r3PL4DhLzx4zQgG49yUykAmlEV0z9ereQNH17fVzw2guejNlojYxpPNwl+edmhGlR9vCzB6vyph8KnoA8eJO0Ati8LXq1gnhKfzzaVZbnrMcG6V4Yv4hL2ZSpVFIy2vaIWkXUGUWFmS4vxTAgm7GQbHZYykcxbRwA2Bburn3TZ7KCdsAON8lXNmi+dMSmObURmdyrEkOX/KZL2fPBsX0tzBvNPo6DPvsl55BBlGk2qNd8SmjPalDa67YcX8xUBZALqxaoMhYD0TP1xEBvCbFw1QBP2kOpBL1FjnCiBa+Rwi7JIUc9BnrVO7eSuRV7knwg/DAhRn7JjHelbYS2bfV22Bv1IB5YULB/LgGsZxHtyuJB4KjibCELfYXZ/lGgaqF6AH5U+djUIZdxGcMJncEFPDIX5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOqTAugIZbTei1LuqcuKDFzcKBFa3ajH2NjR+68g6BstPFucdZYiMaRFZFxcAYNSWJ9L48HeXTF9n3xXeDBXDd1LT7LrgsLw6eauA60gofG6BqVSkOUGmpRHKGnwKDz+DfN7fNnZqWz1OsEQ0AwnsfmYy0HTuJW7hUP+GeVhWGgx6j5k/TYrOxgtMGhVscgWq6BemUkAU6CyPEhibPb/N/3GAjdHtANKRh35+3PPKjJU56xrXOGcTJYnAZ/cqgbGaEpcYiDzgwL4530LLpMhwahDEnVfEfv9KPpLtEyDeTuspyG9+NBCpi2lNMHhAG2WfgrOXiUoRHFf7ktqxBzIR831m5eOWSDxSLLCcGTX/FUlFcxjFygDbjvtvjlBlPxBgNH+tXxeroav0GWSLVEGtpJGv0D7O0qMaHRmXddaRWkgjQa1WCYkPRoVk8G5p8Mghh190hAczVF4tzJX5JB4FzcPfG1SDtNvh2mL/WcpEiHdBZxPHpNS6NYFeVGyuRNKL6zDtzqHTCYr7Kit3mN9AgClFYEfskGZ1gOT1mwBiSwK7rlhXxacUuL65B0mvK6OSqJNScpUplWD0/i82Bd/q9zzE9xRMKT8Q3Q28A9uy8zFqFdM/Xq3kDR9e31c8NoLnoyU418umdjEMOAbMiysWjDhPNZQhvIBt1ID2I92q5T7NC16tYJ4Sn882lWW56zHBuv7X5IvqAJNtvv0uw3Sn3H/kRPykZUJPevOkYGDXpNQSMpHMW0cANgW7q5902eygnU/F4vX0dd9NP6Yg4CklMlVI2SsMaIrKTZbGaIAqxjmUT6Ogz77JeeQQZRpNqjXfEpoz2pQ2uu2HF/MVAWQC6sX3eoIjbbCwaEsPWH66rATfJSVFAyGSYh6oqzA3K5iKZ0K5P1AnyvSlrC7QfF3VheLS/vwqOpreYlwKyfKmPpikBr4eHWFT1Hk7oUH5DnsdiZG7V9s/fC4S+yGk3l4VqXn2HRJpa0H2ADps87QTlE8VoM/3ndYTPTdeC/JZZyKP6+XTSNPCefdECOlvN7RD0LdSjYDEj4GPec60Epazjd5aFrFoZbukH43J2NNU4StOsDWuPlK2VuDpntzDdfXq9G/Z/zvjrD3Ah12dR7HcmQo9gZ6bVtFO1KDIV454fysz2WaMV+iVl4XduHGDzfhc8qX6yj95vAK8LMLeCMXPHWPXQkWtz51jJWP30aoMDhgWDTu/m63UqY6B3XfpLjnranf2visacYzia9A3E1xi7L/im2U3GPzOyv3cPR23pUH7Dd/M2+p3RR7v/MRSu9oaCLolVTchPpH5gpa3KXL2nemftJ2UM4el6N/OoTreTH+WVrqXMlFSk4Ej6r3xCDNfsatUMs/g92sk2ZmkcnetOGcZ0OH2w+70NpRxD+HbILAZd6WbJoE1MVk8U0PEwUb51If0BQJzyEJOCTyTX2lIU3Vtzg17ZLR0Gt/KyRgkJyTe4u5fC3WqWtIN5d93e2z24aihozALVg0ju38kIiNfkufwqb4o13R9j6IN6koF+/XOc2kNh6dblgzMtHnCs5pdp3YeY27I4QdomsXFpOMZ75511Arv3eC6I63ZfmAZrz3GkepDvOn2/WpvzdQU1zFYIDVADvc6JEllJwziE7hTj2EumwZ00TWt8pIXmk/HSXyK/+25441/WvZ6uA9f4Li2gMc+BlNVvTXpZgYouANE5SFymwZ00TWt8pIXmk/HSXyK/1KNgMSPgY95zrQSlrON3lod/0V4mHDuBpCusZMiquAltryVfpv758siYldR9YrBNjzVOk/UPshWvKYs/pt9Hb+J8tVsw34GtsW45BZ2O1gyVP4W5/koIN7gtmUSUfA8S0EqNBtvwfyBajkv8Yios/tC/6FFoLHOXGqfoGJ68CzLeomDFm2/09tlqdWkaFgHyWr0G3evZd9pPdEOIp/oaIgfWODUlDzge01l2T5G7+QErcsD6anXvlpM66wAzEahJkXbJb64db9GI5+xBzGrPqLwU05I/aA+JYfwIqmyFm6phiaMRV4T7hoaAIf6mhD2YN6VlK3NU8hB3W8P9WkyIeLH1UBt5exZCKkTnmiV6T7SkzQV2uNS4UsCaMT+k3ZghUm+yW4VNPOt5ZfZx02x/6qUEurYI7/QOdqg9+7Ms4EGYUhoKm3mzeGgKKg2N88LkkqLbpqH6pgqBDgHeSy4sSG7AipzMWzDtn0F5+DPabMUiWUzEe9hXm+c1lVyDT6EQlOVT9hXHeMuqONLyxK90Z2VKY70jBY60cPRCI2llloeMe1z655emHVrCYu5okxH8QAKOdy+HIgaeKDV5BXG09exCd1Fvb7IH2UmL8eMSgraMrrmXVQg/mVplI8dn+cXCyHcVXK4Ti+0UTHTrtCFJeyS6qteMRdIZbHetY6NdSR0/3bbFWedIxYi+Q1LK2oC/TFk6iOvUnbGO1zJ3KlBUjZmC176XGhmjevquQWjq9guawJ9oMzqilrgtZszcgGcZYv613UiTPrV5tgkSnU+GGjDCVPqpgX+199RbNLFow3Ssmsg+Wsm+tmUhFr1ZyyvFWWeL+9ToMHV7urNYkvvS5vMKv7WrUPfZOZ1xCxTwG4dASyxArU+gDhpZKKGWTZb/jakP41Uy8toqtAbX68/oZUkH2dDW4K2swrX/AvuxcWM9Gpc78yjTBWEkcxH72oC0UcFzndYuL6RfZ796OLQzZm4bu5pN3qlBreDEl73YYHLVnULm21QTTBJZ+Pl9S4evizyVWsR96rt9ly1sXEqL+pVLb9RZH3M9UHJ54mVrXmM0zer3VuWFyMsovJnW19mRZ883ZrlifN/KDLKuok6GzEQfLHtCZimPcuRDwNJIk+5stRoceugfp7R0VoH3vKKprwOXY2sRAJf8VsmKDujFzmRu1fbP3wuEvshpN5eFal5VcKHTZFdZ6b6hQbB46cBpM/8/oqt94BRJdaRgAZ2GR+bBnTRNa3ykheaT8dJfIr/S5vf22uhXC4r6F2dVbRYYoc2gyTREbNb+DtDzr66K3kfs9cFrE5tiXW4CX+9j3gi6fDZzVJdEXGg5WHb94Q9zX4rsA6l/QSkeDqEWYVMboH3LnSSN6WM8+QD3bekoeguIdXNtZOZbbFRacJwOBdqvXlNnsDPuLo6wCzOrgN1EyYgBKxPn1xd+7I+tT13XYE6AvNMZatj4cQHxGUptzFn9Vn351fiCwZeX4RKJTpu0ia4H9UQG4musseEi8Nw+RG48UXwhUpsJCVLWjrT5GGCvj6ZSB5wfjgSoC0RE+VR2fTU1uHKaTLWGivC1P0kNtsSgPqXiExsKMqLNid9wUyNjwdK88moT+kGR4pb6af8sWFbsKzOEmN4EjFLtCdYIZS22GHcliD5y99gvNTbGLp568QohtGw1kcwgcHlfq94d4VCQZsLoU8DzLmjQDywFjPPSySN9c/huv7CH3XKhQO/0WpVCz77+h0Xy9o00cPOXtztoo/0394UO57Nf9RAQTO6FYWnHi88KqPjmVMthg5TQIkdana6AkkEKwAO8jKjQ5UHMGLFRHLGPVe8PdaK7oTyxogP4vGuNHo60//B+b5obbbn9NWIzC2Gh5x9ouTD0XgmiCfeqUkxAOZ0gGCP+LnLRPiErDQLhkvmgM139oAVe8C47A9r5AGnkP5XKNQR+6hZJgDi0YcKyEc3qzRY/l4G5bplDD6vn1FJymozv2SJ/fWgBaE8+oRLIvxmYFoQcAy2x2UApvuPvnxZT9IbThPfIkF0zqUxFQxrPxoeoxf3MWsOeyu1xKZqrN0HZV+LyQkBLLECtT6AOGlkooZZNlv+qWCNALyj81gQ6jqV1lJ7IvM/6q2xtI3AGoJXEcrlHk9D8Zx2xGkIBX0ozGKclNh94HbFq8JR0gGZHgHGtMn6KQBSmUIIhpdTy6gGc+FNTIK+eswzycUbKe16VDTikpw5LYXQFf0B+5mI4kjY6Dn4JaCHIIv4T9ED28bFpXS4ShofiMTHiNNLMHaKzjqYQOHawn26R0ml2xrC5mgRT9j8FzzTqWwswipTP+JeHlzQ1uG/7IXGsmrfnsrNrVWVITNqiqKiEmXKh0CSLrjhNcAfVEUXOYcQcZDdhNc2aosfnbLrHR653LUlLc+ZzO1GcKaadF08KiZ2wvvw77J1enip7OIv2gvEVZbxk/ak958TpHNQloCB0s0jm79QN232gd3QfmokHYTY++u73F0OLKGMJgsdmDqYaFiDTB6lnc9cVaIqDWGqVt5+yl8HaoCBzuZqX72eAvpzTnGM671tJXUvvvKvQSATe9KGL20xr6+z4Pk1PIIj+9P8pI71zC9VRlPFNwshU8IZq3hjq+X5LDfuXg1fTBcBjqZs6p9M4Hf18AuUrh/c/vmxRkTykB/FJ8exKeqnXfjPCmuFIhI5xXSMEku8uXVqWWtF3Rn0Pvc6j/eDR7PZbPa2Lyl2JGwU4lfy929i9z6su5h5kFxd8AFpXV58VLMPGXdDbUEaeGELt750mNS0OQgNukb5SlE5IgSBRGGuqeNajrEfNEsqsyY1rR28HYI9JxZdThS8vHuJ2cKqDJ2B8GFuXXhWnLkp0tOVE6vbkgPu63PpRbqU6OJD+X6W883kP8J5sReqt1IkqP649rcDyCOv11CxX4yIUqNL+JCeUx2dXSi8Y8OslSvGCno5AoImUhadzzJIDI+ARZ/8NsKLZ5IDqQ7In0J98LgByv36oRpq8Jv06Ky2Own2S1Xc4LMI4vKSKZWn0WY/YcogFL7GVusJf61U/hsOB1MWVgFSP6DyYGkwc3oH4gweSK/IufbWbLhpkpCkWksHlKlanMmYRluA3DGPTvVwS5SkbcgfiR24M1+tVfigEYNRlZ+Xw9XQxldmHlr6pPkJN1maHszUYYRRLrZyTn5F9cROZZDVZ15vvH7Iru7BqhoZ9n9jC5UbM4yA9KlBE5J0BNL3G0iAOChLoY8yzvraVgCeSySN9c/huv7CH3XKhQO/0WpVCz77+h0Xy9o00cPOXtxgX4e63PMjs6fXd9+PEFLgjqtEWWwoPeOqKT4vMJ1zDr4f+UCojDqTSOGr3iOs9ZYMPU5iOmtasYNkIcEUBNosnhqscy9pS6dm2+WQyTvXOLRT+nTl+CNV0MHMyCmliePeT2JZVmb9AodHysbW7+8zCbw15OujtzOVlVK6T56j7YpSEIw1dOm+16+RKM4mH2mvCrzKQUi6JS7nDdUFxl120Gc1Vx6m7wiqithFLBZ/uUeUxRK/l4IHGAM3d1qKyUauDIyJXwoEfjl6yMoUo/mIdIOx4Vr19e8lrctdlDITQ+pECDdgFSCWG8B1XkPRBeKfVXBjS6SD+dVYJEfuZbeFDGRORNjoSzRwCtTiatggsJynFDPJ2fIZ7FxHJF9JoCCp1eMqwHLiKDU3x516VkUE+nXL6t8Rnd3cWg86oY3TeNE1wWBu1Xp2AUx7TKEpvQ7RFHZtzFEWzDadCJpDGa0z6hAArcnTcTObmUdDFGuP3+S6Ou2NJ9c/JZm4tTQZXzbwcfupqDoHpIOQVweghBTDWpaLEdvx5hljj8s4dNiZS61dcG+909bkSyTZCfmNVYgnKA7YLDStHAyFZIMLsRgxK3jx1wE4e/oMmzJTqyGIq2flHhPobqME1E1zzjfuAOfKlZEi+pYJEYC4P0dKcUnckbtX2z98LhL7IaTeXhWpeWPO1uo3c+0VP1xFo3tDN0k5jgd8SucR2ypaz6DSfqzy9EJEvNnNI981+g7Vxkd4SRcEANCH8gylv2J3UQw4sbS3EjF4tTjVaDlfryxdO1Hpp7Fk03v8qAOYqJxGkKgqtYwqwfGoB2Yz5Kg6fOtGlD38LcLTmA3fpq7lqdge6sSCWEhWJPcXkDZsDCfDYN2XmlZ6is67BLi4XsQdg8xSvLb55h57UNvXnyq0SKUC3uil07M4GJvDLwt9sMJwnpt95kK5P1AnyvSlrC7QfF3VheKSq2VsOWucUafvAakRg7JEs+uAH8Po4U3OzPQV6sn0Tod7fjIJOWf9lacrwT7zWaNAiEhGV2WVw/09LTDFkae83FqUNdqmiMJt6WD5rJbLD8RAB9NogDz9n97mEJ2YKN5GgthV2pWTgUrcQjQk65r3gxGujlSqydMjXHIRRxhIEv1qckty1R+MT0QkKjKBNiPtWBJojSgwOvP3Av0gK1r/+kTeCTjjtYzmR10nm5dLJlhIViT3F5A2bAwnw2Ddl5pzg9VCM1TGvt0+gzvOdD1FFzmZqG+jMbU0kPJB2RRjadOzOBibwy8LfbDCcJ6bfebBs7H8l2qx4v+iaFk/u6RrTjL2IGmY4R+6NHYTCUz08Saken41eGy2KUljyb+hY/cQfLHtCZimPcuRDwNJIk+5e2/+4QxYHpX4kPpqsq1zqFSbemWVdGlug7/NHM6cvmWbBnTRNa3ykheaT8dJfIr/4k4ugV7R7JJzSQpS10Cl07Uu+0sIxLpjVkWYmGAjMGyqb7hVOcByWpyQfSiwT4xcj+ywLneidC5xmv6hKj6UAeQKty0zYEVVwLM21pmHwhh6J3dJRqu/ibDChjqEQLBYpJzyK1t36B8H6sWL1ntAjl7cYvRdyXT/Vjx9oAfCXqg1LVkbEUv2W68Xl6f4D2i3KOq+gvHIKnMWIZGeAQt8KRGrHq+S6XUdJ9VGQvk6FQ62vJV+m/vnyyJiV1H1isE2lwZbUW47xCPaO6gKXp4ijJQU5cQUJ7ARDGjXGRXpDhN876BU8XAWTT/hSWM41mqnV9Zu71qrWAuQkScseIOjcmWrrbAmDnNLt/cveAkfGJAooKWAOSaqMnzEvMs3+jNrlTRkg1+obHu0k7LpZTbq01p6irDRVMMR1i3wxg7P0IFP2IfOe7VBN8UY5zsumw5H07M4GJvDLwt9sMJwnpt95kK5P1AnyvSlrC7QfF3VheJfQTP0yTOXnAD4L+g8sDEWnzzdmuWJ838oMsq6iTobMRB8se0JmKY9y5EPA0kiT7kPv9WCTV+LGZEmvzm9DaDBHFEEa6EfyQL+xkGnNShnftVnqj2D0+2bAlFkiXlBAcVq7+exMi3TTgxUw4uFkxPrZGVTm+wpZr7eUJQYFUxX0apvuFU5wHJanJB9KLBPjFww7c6h0wmK+yord5jfQIApm3/kUkxJKYa5qVcusJ0wzl8jh0OVkGnj4JIw8nxkWE4BrcwTRsHzQGX5PaZPNH1w2ivJbA9dwmIj0MiYYL1O0/RNagwJR0AOJ4U2ZNP7Oe7dw/lk3kBpJbN1KiGMkH2usaJT+JVF6SnSdzB3F5SrVHBjxc0+raOxcM13klqhDRE25ZVqKDI7SPOnVqlhcIk8eid3SUarv4mwwoY6hECwWH/fJaxi6XRHeAWye5xQGxRjqd3uq6FcUyZV2dOja/1Gk2adjxiNSDIajgDvGm50RVhIViT3F5A2bAwnw2Ddl5olhm9lb3ngKxBAq9YLJUcPIkrdL0rTE8xJJifVfjkJYGM+G0lqXynU1exk3pjdXulFbDdHYoBdr5HjrS7szjLySF1qnTZBBDrEISriJyTEqw+FdimGs+1p6rrNBhFHqfUy515FgSCJGQyCTQrZUHcikbtX2z98LhL7IaTeXhWpeagjW5Bbj+PSfLPxJdNRK6E7s695JPMd9V+JT4wUzMANMudeRYEgiRkMgk0K2VB3IpG7V9s/fC4S+yGk3l4VqXnI4QAIpplbL2QH0SgfcGMj/bRjmDSgZDi+dpRZIVRz0kmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagZx+J2N0vAXahyPsgraW92LaPq5NmzYnCIidw9t57TjuxenT8YRCYllCG/K/DJblFtEiC8KhROLhSQmIs6xOBNTYlmat6DG58xgA7tSf40VXK+mpjjKWPm7E9Znh4BnRH8HH7qag6B6SDkFcHoIQUw2qQGwZdGOAMeToIISMcIxNYvFDHJC/rDWpjdeSdhBPmG5FkDYnn4hA0avNmWNMK7li8IorGoZamvIIk99+D+7cGVoA9WN/1PPHkGzW8cMiwOTvznBJxZBBg+fbgR+ELKDKRzFtHADYFu6ufdNnsoJ1WY8mDhjMcz7LaD4DW9zHhSoZj4+mCFKtKxgujH2MlEXIIwXigA3GFJ9uCPUs2JW5nEtl3Y3ingn4sWwt36+ECJ5mT3yOyhI2TT+2awcFp/AAZ/IW4vvtah6zg2/Gcmy9AQTtKEUShs0dw3TFOyNTIf2849CFk7BysSg4z/xJTGuWBqs0alJZ1GasAJki7FMEfeLd7jjenbuLqI8IqMounqGIja4sn/7HkSf1k/qm4jTll0VGeFmkXoxivmlTOP0iQp/rvXHxpU7e0Esp1JaPnW5Wvy0ICQ76PIWwxVsWzFByZo+QEBJIQKNW7Y4vQGz2SnYp7EDpEasw+emjfcB29CoxUpVZHUqjgnB7z8Q8rBcYbCIqzGs34M1Y4ItvznKziZAHGBEbsFyjpC0g6C0aE0R8Br5zKlLXJE0PjEXWCjkBRaS0eQwr1gQe4IaHxVQfwcfupqDoHpIOQVweghBTDR6hNi2BOT3YrpV+76yFu/664F1rtrrMxkDoTtZaRBXxKwyHYiKekKnorHCY3DzrYZxLZd2N4p4J+LFsLd+vhAvTWwETrzZCqzgCnNLwrVW4POONDPdUXQoLPMr6kLtHfp7Fk03v8qAOYqJxGkKgqtZRkYMZNEtnfwgsgFPJQh6UGsgpdxHhFDlLZndjPu9/uUh1k9mIqEIAN4CgfaJSMQkr8W4xN1ynveeTTiy/HMGVukTsbPusoX079+eG2iYhESYYabu7xJGKTRgLqnIjmFcc+R25Zp/g9LhruecfBcEUgQBqPoT2CzWDoUfr0HqKZ7KgNj7b4A5/vQwID1FRkKtRhIa4gUDTk2Qovv/2nFqY3lDvYL+pPPXbXPSZy1TL7x7Qn8yUxRgrjHEm3z2EOl3ond0lGq7+JsMKGOoRAsFhZ+2Rz4rP80qFcggsGv5fgynAE/uusQkjqdabezUPtYqnV4yrAcuIoNTfHnXpWRQT6dcvq3xGd3dxaDzqhjdN4eBsiDFQezfHtRtb5IZOFwUmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFag9KN0FFyjj+AjsU6ITlMEOxlCuzOXSiyXOownEQ50dvqHe34yCTln/ZWnK8E+81mjqeGBad7I1L4+3iKVazm71A5sZOypplINlK4A0KYNesXVZ6o9g9PtmwJRZIl5QQHFSo3QO8wPIvCCZWhihj4FHpJ15BZx7LENQn79F4RjfJmSnYp7EDpEasw+emjfcB298IJImq4bu5pL5Z/TwLxGKJeDGDVmhP1Ev6FAJuldU4P5stQWJi23xuESeXXheeOHR0BvtQHLHLE+Cv+vq27ug9q+oD7YH8DrBsBEMS0SQK7TszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4ggjPlQlwudhI1ZYDuzaD8lJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRZNES9upgDcSvtiYSiF3msppxOkYK7hM61jVmW1HRlxjdUrV0LSV5wxwnZvYEnDfsiMQxflCI9z+O1DDv6m7FR87OarTzZfR/P7I6yTrQLLMP/iS2nA2cH6YT81qg61uOSOtfQ4VVv5D8uoc6zQIW7Sp9BUp52dncVv98NMNFiJDb60D48NIAI13bSLYSE8z/VKNgMSPgY95zrQSlrON3lrDFEC6Ob99NPzvqUXNa0NoWEhWJPcXkDZsDCfDYN2XmvTMDltbzKuOiuzRIH93zFCZLi54ktv6UGrOuh2MitWXOQM/w95R72A3qpCueUCXdrr3NcjYARQF7WPElVSn+ClRkzAV69+O9Z+4dT34UTLv3fg8YMLDdk4HwciBbKfkVbv4TuLGzjEBqHOZchnOc6nvc9ndAKk9qrsw18zCpzhjtlWD5UNOA5qW3XQE29MPr7r5xf7/hgAn432rtcOCKrQjnmeA3C3u78rjkwL8l3V+MdcNpJ/U+O6TyOSzV3y/RstJgVi3CNgwedMkHQGCDLeKsASazqVM6qb+wQePw+5WwD4rFXHRWY1vkEbtwivxqxSFSX41SW/QUHwoSYG6ooTPRmpSTBf7ySzQxI+1lkgJ2vgkS7ao98IKHfrv9Qq0uFhIViT3F5A2bAwnw2Ddl5r/ECmO3Lmov5Av/ncV8l6ZwCw4bhb6Rn76XuzX2mQvxFhIViT3F5A2bAwnw2Ddl5pzg9VCM1TGvt0+gzvOdD1F6Noea7kFqjCGstO5cOOiZ+wiU6p4EgDKa+asU8j5Gsym/tX5m0tzwTkPzRYSOMgAhePuKqERI6bj0QY5JxOL70rDIdiIp6QqeiscJjcPOti0SILwqFE4uFJCYizrE4E1npDfrFTGkyn27ZMfD5L4JGrOOx1pcDkNlOYeiLQCAK0vsLJiY8LxnDTUYShSem5VESm3h8pppczau8yfNEmYkfCgtYXfbijyBL619GviYvIvOqRy/mJrDDEu3Kg8yfn6PrSbAXQSwLx6ZRj8SdMT0PLwUyXc/YC2JKZQE79sKKXwcfupqDoHpIOQVweghBTDqaf0xZetYJbMHKhwaLDRd2H6ycdpe0EvUiAObqeRvuaxenT8YRCYllCG/K/DJblFZxLZd2N4p4J+LFsLd+vhAtZM9C3oa8QMc+d0zmN3DWyfPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuXkYcNcO/M+U/XsC8oC4Dc1QRDVESh4iQG1Xk5dSS2T1yQ6rZkpPvrTrFWaY33g8AcwoyTmJrwg770xkgzYMwkj/M6HC2WIjxuDQrVFxpprakp2KexA6RGrMPnpo33AdvdEV4+6bLQHcgwcEyS310sADgFmHyhMzE+HRRLz0Rm5R8HH7qag6B6SDkFcHoIQUw41mwsHwsx3tvS31QHZ82SaH7an8m2vjvzjGo/60WCuASsMh2IinpCp6KxwmNw862N5Tz+uta5LmdZiTLxNOVRLkrnZORkx84XNa8YdHzwYVrBxFDL5ou7O/g+aNyhSoUNVnqj2D0+2bAlFkiXlBAcVZ+kxqN/T1aBmKGewrjg38bYyq0UcTZRBh3ZhJXe2wWckOq2ZKT7606xVmmN94PAHLN3Pbfh1fXyVywgy9RCjCkHnVlkK0LgCRfPD6qxAcGZKdinsQOkRqzD56aN9wHb3Smm/ZJtM4u3tOgdtMtPTWTnxcyKIv00Tjb2xxzmD9rfBx+6moOgekg5BXB6CEFMMceyrdXuXD05fUYPqCeQPJASgDEPNRG0gj/MpGuYMw2anV4yrAcuIoNTfHnXpWRQRBSlSw3IQ3BzNBYyeWmUALBFq6YH+oftnaXd9OJKlLYanV4yrAcuIoNTfHnXpWRQT6dcvq3xGd3dxaDzqhjdN4AhBO4aoWAZ3SvdIDza4Ja0mGGm7u8SRik0YC6pyI5hXHPkduWaf4PS4a7nnHwXBF4O3CX+SsqI7w00czg/bIF8b8wabBHieT1GB7ecQWt2Ldw/lk3kBpJbN1KiGMkH2uR02yYbilSzKCEdlEr3Y3wbWCTeSqUCnB8cDo/o/xPtabBnTRNa3ykheaT8dJfIr/rBYNTUNW9HcOiG272IrHZM8788FG2DYZb69kW3YJuMGqb7hVOcByWpyQfSiwT4xcDsQzcz6pIdNPFXhLQOlWICdENmjI3Z09AGS6NQB+R7ZYSFYk9xeQNmwMJ8Ng3ZeaRD1eh+jPKhxcgOYx8rlggL6y16QxLon2pIjubuEMBirTszgYm8MvC32wwnCem33mwbOx/JdqseL/omhZP7uka7kOGImlBKX6TysVGyK3Z5JMO+4wwuZn8Ptb16ac/qoL79qN+Hro3V0h8WxVTICXCD9AI9vyUwqr8+Fjuo01pfNjOueBl8M4Y9w1+qoWdIhRHLznB1Ib6KcBZugieqHVI1oyLgkuYcoGU92437vly3q6COQ5uwNugGnuqIu3W0+9tryVfpv758siYldR9YrBNsJmtVCrg8DdZxF0lQwEm2/wiyXnvDs1zz6S781zD9kmqdXjKsBy4ig1N8edelZFBPp1y+rfEZ3d3FoPOqGN03imT8DFnwdss/6APot9rfm1SsMh2IinpCp6KxwmNw862LRIgvCoUTi4UkJiLOsTgTU2V97yjXj3ph7JGOvEFhLyTPcjQwkSRSASQGxyHeXzLMk6gAx3WLYVanLJK9XiSa74xlAWavQ+x+lIq5DyJnqXuI/O3mAwmbtNVcHKzz9AKwxdZCFiyBj1I6XAt9BykYfP3YihZJCBGZlYurvAKfFBu9mr0WQZros2T1yncLUPuZKdinsQOkRqzD56aN9wHb0incOGvAemMVhQxgX7zd831G7/T3W1hvtUuyJKMEiPD1hIViT3F5A2bAwnw2Ddl5pUdHc80LvbG26m+desd2cgbIvJYv4cJDeJWYCS6mYOYEmGGm7u8SRik0YC6pyI5hXHPkduWaf4PS4a7nnHwXBFNaendzOFWbqNG/m9fh4fDCtB4x4RD7jIhPfOtSzPA/DVZ6o9g9PtmwJRZIl5QQHFNzsdCqGKMnGfjhzBzP+7GRWS2RUl+aaHKwG+NmKFRGDl00jTwnn3RAjpbze0Q9C3/QOlKLV3v7LKiOEsK1k1GWKZYQKifxHrCc22DPnBaB62vJV+m/vnyyJiV1H1isE2PFMP3Li9irOojQ9yhcBLPMybk9dsaqYPs8KRdnGrMHjwcfupqDoHpIOQVweghBTDV7Shc5m6YOW6CX/CPS4EY9KJ99FEGocpdftNHyQB/LruvGfmssTGwFA2xGXJCQ1BEYUYCYwy+7+8udRaqI3cnXULuIamN6XLpUUEqxbddOGdZQpkdEr1VjYrFWNOCzndLzqkcv5iawwxLtyoPMn5+uA3Q04BRK1FMSPxZYn78JI61zHT6zHGIbBCdvO5hZnC1A9BLHgxQA+Ak3eVPQwK0MGzsfyXarHi/6JoWT+7pGuK2jU38RWpZePSU3SjpjNoGUK7M5dKLJc6jCcRDnR2+pG7V9s/fC4S+yGk3l4VqXntF613SzKK8qW0s7ldJxjWOwlEPXjJUPA5uqs40dyQbNVnqj2D0+2bAlFkiXlBAcVuDrwD1gAW8qoYeZ+4aNkKOrmC6GjmRsFbWfPN9BrNxS1/8j5JzRVSg3RGBijGXSAPyqLHZbUV9WKk71/1vvOsdj5eVaQYP0zKVKVW/TS6a/Bx+6moOgekg5BXB6CEFMNHqE2LYE5PdiulX7vrIW7/Af+RxjJgzcxbY6mWGK7fyHepSMClkvz9H+ME/ZoyMVpTK2iPFNR2pmtfRK0zs9JOEwLCbekpKHm3jYWNaLdwzEmGGm7u8SRik0YC6pyI5hXHPkduWaf4PS4a7nnHwXBFRRy3EUR+/vWf8updE1UTWaaHxD0orFAqakfJ1HoJ0N3VZ6o9g9PtmwJRZIl5QQHFTyeD29w8OduBdQOIKiaF/3DkdB338BBP5XmuG3TX9Mux+7Q1VRjvRLmL5uIM98gjD8qix2W1FfVipO9f9b7zrP5UjhQOUohYskLuXQRW+5Cp1eMqwHLiKDU3x516VkUEF83sZumLzw3NpSquqY5X3cWoHcLF6iSzzDDDmVkztgbUD0EseDFAD4CTd5U9DArQQrk/UCfK9KWsLtB8XdWF4rCkE3lrDVIEUmvbg8h8b6oSiH3BO8fQYBh+Wb04EX0See6KsVHYflPuEvl5jUrBDnROgK83hWUV9ie+5r7NENCfPN2a5YnzfygyyrqJOhsxmHxhc+dXTqxiM8wUQpwb5InHuTi7t9+HmB2FOlFhhC8k5dOWh2iE1pA2R/85Md0+OH5JPKvGoNS9MugWFooofxQ7DJCc9v9Vu1xUjkvinMIvXLbT13t7Bew6Z92OzFOC+kMjgka8Il1Xsr4L26tSQ7uvKelvJSBSmQLUNO6o+GsvA7L7+4hvYs5ErznJMAJ+hGKA5+OP+3HKTkORxzafHFgXcbE86vNx73mxWJEbiRX+HSHXofUJJBFFrjjfiO+jTs8N0haSnu8bEgU69VfwPuxCz4OUoag0fB3+TEwMVZkKpmeIeOY88iGhWOEVNPmdypWRIvqWCRGAuD9HSnFJ3JG7V9s/fC4S+yGk3l4VqXl1BPjh5v7PlKO+kj86NZnDeKlC/SWlmD6h9lukolmG/9Vnqj2D0+2bAlFkiXlBAcXsmwrO8EldhVu1lbmv323v/sgz1D6bygvNmAKo/aGob93D+WTeQGkls3UqIYyQfa7DS0ws0YpQvdkt8EnOUCd/pdLGyy8tZKVLAk5rxiYfNpsGdNE1rfKSF5pPx0l8iv9SAFm+e4mjjbzHF+tqaJuCvWdw7He7dba+q4jsoVMAVl4oPB9Sq/W/VXHHJS5EPSa2SDNn3RXsgL4ZOkBNcRLRZclb/HyhDVn42poq5IZC7Uz3I0MJEkUgEkBsch3l8yzJOoAMd1i2FWpyySvV4kmu/1RZFb8QMyeK/IKmAWgeZvpA2kALGiqKMLWN7kaww07JDqtmSk++tOsVZpjfeDwBomhYiS+OAv7Kz/0KS1hmiGXwjWkJjZUKAAgq4o73dS2nsWTTe/yoA5ionEaQqCq1yaSS/XJYi7sgqAMClVFk8AjO2oiVlTuLmoQFbCGDCBF6J3dJRqu/ibDChjqEQLBYtOsy0hSBNuFYDdJ0AK1SsuSnLNCIFcbYJiJVKnc3d/HLSYFYtwjYMHnTJB0Bggy3YMxjtft00gV10I12+CuiSnHDaKLWGrBwHfv+p9fXMrQypTWCj/I4WjSwYAlVToJU1WeqPYPT7ZsCUWSJeUEBxZK9l8Z3dyqWSnI0fEuh/bOsppAGBr9linzWGp3yikzxP1m+6OfiI7wTXefDhTx3eKt/eQW5c7V7X34Bcdbyg+dxtFvQbh7Pq+5HiPfyl5qeqdXjKsBy4ig1N8edelZFBFE2Qr2uhAupnXbtTyVZT5veM+GWvBvBZzxesgpQIJq+07M4GJvDLwt9sMJwnpt95kK5P1AnyvSlrC7QfF3VheI6ucNAxDT+2/ddy2VQRqeUSYYabu7xJGKTRgLqnIjmFcc+R25Zp/g9LhruecfBcEVDe481EfdDYS7iHEXJ459s6oeRPa4feNmJMClRSVKGqt3D+WTeQGkls3UqIYyQfa5UCuseBdHhosryYpoNw1hpYp45Lj+20GmH+7p8sOvx4y88TCs06T+AvaUcEuFokRv7lklgmrtRzUzLpmYqV46bE8oJFIj03Bt5k1degKP2kZsGdNE1rfKSF5pPx0l8iv/c2d+jT6KZes5CgkCxSIe489hTL7UPd7XNiEyX225ZbjGM3vIg2QmzVBRhJ/6oBFSd32eBH+GIJ9KVRdM8RXiTVBfMHWVUUSTyOYTxJeQtH7a8lX6b++fLImJXUfWKwTabB0NwoiLAQhPkzUfr2B/ZN69x75HZSncNfOmQjl2uWKnV4yrAcuIoNTfHnXpWRQT6dcvq3xGd3dxaDzqhjdN4AbYm5MDIwe9Xa/TXhjHOgtOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXijhBjEsNsiceeCKbp9vsljtOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXi4HeiCavJmjmgMiFTAYxZokrDIdiIp6QqeiscJjcPOthnEtl3Y3ingn4sWwt36+ECpmJDiDbxTjWeC34jVtN1KkmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagkMwLWCIud1d4w/9XIsKs4szF7fME230+CYywZIoAQwWBvg7u2Sf/w919cEThyPiBw8jk/Od9qjOhIgJUgUSZOJI1deWKt/faUDJbiHgjdMaCTCGSAOwCLyx6A70zfhzpfWS3KQmnM30xFaIcVMasT3RRmjZBBZeaQXRqOyoQSyogk9oAoHGh7xs1qXd/7mAuXXE6/QY3E+efOaq61jN1rLgb1/Ex3hUFRe0GBO1ybUWTmTZNZymg+5wb0UNeG9eDSqmXAD2KaGqLLbzxD0I/ySyS9CgArTA8eLTijGEQYCTTszgYm8MvC32wwnCem33mwbOx/JdqseL/omhZP7uka981mMT/keiMvjPzyxvn/hca06uCvt/z3trbfJ4PqegetEiC8KhROLhSQmIs6xOBNSeoa2A/dRH49qMh2RhdbUFM9yNDCRJFIBJAbHId5fMsyTqADHdYthVqcskr1eJJrkx2eHwuaJ7BPatbDs55qmZywYfLp+IzTBPiH9Ho7ISt60kIuLRs80fGuMDgl3seZqq6Z/QEI0hfb6YW4hjkn+QGn5hEvXGhHb1NQdn+PQmDkd/bL9KHNTeetWHGQG0EGGLDDJg//CEWV7yruck9nM7lpRdEKPBXYi+absil/kQA5dNI08J590QI6W83tEPQt8IHHKjdANRfbMuiYKqUJY79t7DGGg5SptPuCiAj9fkt3yv0MIexl2ZwMY2iCL7s48k6gAx3WLYVanLJK9XiSa7DEOgd2ogIgnpbli+HQcxS3l6XvUj1HnMralHANho3lml1Sd9KQkTCTtxBBA03NsrbLT6ij7OIYI+orY3ADqF6x1IWicjZCKDS/sLdxUafp3gPVDXTGxW6FgpIGUvDgiWGu2Gm/PyqDoDag4OXKbNXaZ1u2+9s3BTMZSdV5y2r3dOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXiJ68S82pJPqFBEaI6TXTDIy5Wnr1DLZSlcbsuypSB8SBujJotWnUJPtbmx0wcjcp+sIquKiGsA6skL3RRqH5vnyppeVqzU/IEd5m7BU9ZVcndw/lk3kBpJbN1KiGMkH2uCK5VO3hxATD4S9P3tMBtVJIU+QwB1dF1ZuB3HujzSPjl00jTwnn3RAjpbze0Q9C3QuIfDluUdkTAjfN/T9uFHLTVkNfRqsxz/kQeQhl+t9LLSYFYtwjYMHnTJB0Bggy3qgTyr/u0hI9yWNKWl2QlqY79C1sr7YonAFhum1jiBG3TszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4j1NSWIZwlu/sBCdbLMYHjd1934EtYcct/eSmi+OIjk5kbtX2z98LhL7IaTeXhWpeVYSoqjVtAUV+gxu5HS+BpR+ErndeiULzLnnvjI/jLGU1WeqPYPT7ZsCUWSJeUEBxTGDHl2i5tRXufCb+DJ3QQ/ozAsr8LtN9T2qX57yzXyN5dNI08J590QI6W83tEPQt54KkxSBlQ45lhMItpHWiGS+NE91s7RUNz3r+5ZYcy1Leid3SUarv4mwwoY6hECwWBvAStuyd7CqHOk1Tl85+SxN72jB0zW0iyoDORdmuPf1WEhWJPcXkDZsDCfDYN2Xmo/d5a+3rKVzNVzuFtYTJUxA54szXVMZna06iTiBDFpYp38WCdeaU4SdBUBtyeGN0gn6cCU1IuH7Jnoy5m2IcpkLgWJi78Y7xPX7n1hNtsm8B7E/nHsF+DMhdMLFd2b6Yod7fjIJOWf9lacrwT7zWaMbTpdfGwh9pxiPtN10tCEJkS/opEcJNgoC0g50iYktknond0lGq7+JsMKGOoRAsFhAGEiR0+aRfTFQg+LHisWWGZZVo1EoVmhXYTetLsDfPp883ZrlifN/KDLKuok6GzEQfLHtCZimPcuRDwNJIk+5XCK6oLNPZ/56FS0WEwfZBbPYtwpaWv4OvJDmGY4bUzNkNfxyOGBHjACRMK8FkToIFDsMkJz2/1W7XFSOS+KcwlfyabADn3WQprF7mZOzQFag1qomUFxgSYZByhLHo3mLsDIMn+JlUH/vv1jd4syqYG6dfPrWk1EQsEaH7ags1H3dFr8++ypMRIfF7D0MXbuHo8BWOKZJ5rNwUQmtINPvwAhI6R7iXjpqSpLsPoQtqfWp1eMqwHLiKDU3x516VkUEQUpUsNyENwczQWMnlplACzTM5VG+WgZOSQXAanYkG8rwcfupqDoHpIOQVweghBTD3p8mA9EcbFKhw6o1tra3Kh4c6f3O0i8Gc2cyP6YSmWHayDkAU6c4cAhf62EbVMncfi2ZWRFZGE+1vf8naeol7SnW7HXZPDCyovcbnUPfRU2fPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuVsLBoHy5Db40h2HWesbjTWChtcRyUsmRrsm8tV3LDoSRWw3R2KAXa+R460u7M4y8kT+3+SEo+ic++gvryE28FvlGsHQXNLpeUrU00IZdqlRyCBcaV7nFZ6sTUGEiSznWLuvKelvJSBSmQLUNO6o+Gu0wLxjt8X2Pf1XRuPMwrQj1WeqPYPT7ZsCUWSJeUEBxftpiEPVd+QfFs+kdJs4ljp4WpkHbIY3OopNmnyIear+fVCDbpURTZp2hfCJwBmsj0BTJEz7u6aXt9xAKLAa6hT11PEdfutKuz7VGp+Vk8jAqm+4VTnAclqckH0osE+MXKupd7hnv/71Fh3RETZ2UV/T8VPt2QgyE3jw0m4VmfMhNS1ZGxFL9luvF5en+A9otyjqvoLxyCpzFiGRngELfCkad9JTSlWWQ+nUy+zDadYteid3SUarv4mwwoY6hECwWHkXfXg2OI+AgfyVKaJTGfWA5bSc96BtzUKd346A/K/Mli9nz0awT7oc7VQnxZb8hvuLP8JQA0/Tne80Iof5IBQhGiw/Jr6VMR4ZQxnldgB+qm+4VTnAclqckH0osE+MXBl6HJiyyXRCwhEkkxSyuhmfMKuqIEhIZrVGFQTP8664qdXjKsBy4ig1N8edelZFBAjr/UGJ26LXHfpa/vin9svhk7h5Mxy8YI2sWjAc2QwWSqzmsyL3e3O4Q5W7xPT/YeuYDvn4IEuV2kB1Ai5fSPPZagaOOCefT4zY2NTOBW5XXAr9s3e8x5cK4/Iewa+Wq8j7apZcwwr/kBCNijrE8QsZjjXx29nfLJ4HMRHZtSP67CJTqngSAMpr5qxTyPkazNAs2QXLCiUiZJ9mKPMOqMq3XcrgsmkqgGkI3xafUIo/M3Ic8Gfj/5JGSVOKOMkpQ8c+R25Zp/g9LhruecfBcEXN6AMZnT0A1buYoA0bfSH2dF47+jQ4oWZZ5ZtoLuTfoz6trw5/YFk9tbaOQ+wnrOnLkZHKRjXf6GNG212q0gP2aTEM5F/2uBTYVc9K3/AcdwterWCeEp/PNpVluesxwbovE4AtaP6snSaCMrnNZpBNjhzXX+WjcyQNnhatgIaBhZ+W7SRqq2BslHDPZaN5IQSNfVMiVjI6ziZ9KvkVdgm4tEiC8KhROLhSQmIs6xOBNQo477EZTgF0y18TH+IZpO27bFmoyRnq8xLomhJTghZ2qfRdI9qzHFzRdMJ3C126gT4hnLvcRl4ln/Xu5puguPcPVxIfFVqnmGmkm850qpIDZBj5Ou76nhfqHn6rbPyEVqlQSPmJYdFD1lk4g2lKH4zt60+g95bqxcN8WhMS/UGk/WpyS3LVH4xPRCQqMoE2I8H1m5bwTVjJhqvK+1NUKcCQSDhXeaU5Bvu+Zigg4qw25d2K0Nh26eAq/8MXG74sdDfs/GAz8Gih3WMA9eq0Xs1PdxTwAoTA7v/0c854NOGbYSly1rcfdlK/5rLn+L1436Q1qhFJYLA+Kp4ZvSXvvE0Ly5CCISuABx4v60suA9+Iv4KINrgOJ8i/YL//LPHNO8kOq2ZKT7606xVmmN94PAFmukgIRsYKCnkBl24yRuRAojrFdzOn0eLGEreQPP2R4j+M6AqwpRS1RH6/BLf8JkCXbFPD/jZn6B1Q83Ottr3oY/pZpuBJ+7Lwcvbg20dPA/Bx+6moOgekg5BXB6CEFMNHWa2hC2Wfb8ljgSuZ858uKwssc64BbiSvAgrFAwBR5S5Wnr1DLZSlcbsuypSB8SArZUmjpsIIYPhaPeWzj8FGEVopaxci00sA1JfhJERb6F2M4SRl90kptVdMU+aoFPaBvg7u2Sf/w919cEThyPiBC6jM+15wxPNV6nOAifvPuSaken41eGy2KUljyb+hY/cQfLHtCZimPcuRDwNJIk+51tq71yMHNklvAJhbNvFybemy9qe3qQ1O8/aFZwExVk6Ru1fbP3wuEvshpN5eFal5mlOmIwhybuw4DIEqsbtnHWb8GKeD3/e4rByThlG39x+P9ja26nZ7Hamg1//BLAG0kagB+55xz1CptLHDxN/+2RaKzrfEg4GB98RYm5r71SVJUU9e+iGPiAIjqNoS1ug+Yd8zLe70mpNaQ01IVEvsVGSjQ0TT+yXup96kVUShnwZYSFYk9xeQNmwMJ8Ng3ZeawHORFqsyvIX/5RRBMpvUr/Y8QGg5bwpXKr+5IKzMOhjwcfupqDoHpIOQVweghBTDB9egKYgG1JKOTGUOg8h8iPRzZ2N9S2xpaN/9OyhMB+rTszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4vTkOuK10tQLP0yVlY7Ck+ZJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRYtgscMc8xYYlhjX2LcHDXefPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuVmSLiqk8451aPU/U1wHRkn3V6SSpx7CezfuQFCeCo4V3cP5ZN5AaSWzdSohjJB9rtt3BC2PAHKN2afsZxwhN5EhVrPZ0zAbasNn2x6tXPt11WeqPYPT7ZsCUWSJeUEBxTBrz1nzZJw2Ke6iMxhHBXfAtOPWTVW1nFA3fJfQRMSaZDX8cjhgR4wAkTCvBZE6CBQ7DJCc9v9Vu1xUjkvinMLH6FLOsjiOL5GJHk6YCLCjyQ6rZkpPvrTrFWaY33g8Aeqi92PFvRbynwBtHphwIrTTvE47Fo5cDm4wZzIFMdkKeid3SUarv4mwwoY6hECwWF1mKmTerlY9PnpFh8Yk7UdtTbFIp/gMT9igcmc5SY+deid3SUarv4mwwoY6hECwWN7It2dAYLogEK8NNIryAqn1hLbEGZTS699kjp3OwY7ee4kFmVYu3XQWg2ty50n1cfVLxbjScxyNTixTNzf/eNfunkKTxl3+T0EV3x7lLmqTP24KXUWXpPiVD7qVHlXlTj+M6AqwpRS1RH6/BLf8JkBeS+OKuBDlEvrQb3nsmuao/DYOkMc0ZokFemIpFRH7G3ond0lGq7+JsMKGOoRAsFgwBMi4e968lXXQ/yIS5+USaG6mP7llHzRwgHKTz+fVTVg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jQyukoHo3y0bGd55kxQ06FPuyXHJrS1bUlUJyMV1M/K1UfzZJ2QTyt7DGZ4oh50gRT19ujBm6NNW+htUzDOqhUQdaOHYEHCAdG48Ln/vn4RBafCmGM8RP9fXcG39ZQ2ERPq2vDn9gWT21to5D7Ces6VSAEmFoH53mR2IYVjGmHiRDOjms8ujxAWkJY601cALokpTC3wnwc6kJklthm8BkbHL8vZmlWiTw9SVW7tK8QZKMLlX9uz8wyTE32QQl6pQl1bucTMaGkKxzxTPwO7YeXtujZPNLBDaKnJ6kOjAL7L2YIjCV1Sd9jsmEMWARBIzJXaCs5HqiTXZrokceJBL00dOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXivT1sHMRYWOfqPpgjHqL3S0mGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagIpj/kgCzfoCgrcdsCGCuQuOjp/HkUyB1uBzcaTd34FPd0Vd38OD7WW1qv9nBS5wZaM4yUeVqTzSVVU5OhvDQ1lcLUtjddU8E/+4g6sAR8nxeOsQGyeJlc+77F131pPzPg7lnHf3WwmpYydUj3dUVTTUIMuPOLBuA4rByrdgx5HCZ3CZjN+WjMSUC3j2Y7I7xO0ShjcSPgmO608REPQHzyqTUNp/dTJtob46oQWrd77+SnYp7EDpEasw+emjfcB29qO5Gv9+mx85pkEEd+EjtOhnXB3AhKvzUdJrz2xZOi/6P9ja26nZ7Hamg1//BLAG0Nt4IPXW85SK+K4mFh0pWTURGho/7k1MRifElOFdLHUpJUU9e+iGPiAIjqNoS1ug+ZRyP5cmy/W2Iei9QmPoQ7OIhO/OLf4el+FObmetN+YzkUX2+SrebDNGDMDvhtvbdgnG6jCYRgOZJ2GUDDrOXGy2MjOEPuwN3Sr8DnNpxGW9YNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S8lZG+AvwlS+86rNwVuKPCgmpHp+NXhstilJY8m/oWP3mHxhc+dXTqxiM8wUQpwb5A4bnT7+1VSFF+9D4+46TofbRaqGqTAKhA38t1cQwXWlEPTPityyU4QtdUaNERym54vhf1VQsr/vCPAg5R1jWq9J02ldRw8nZhrXU4heQbFhO/l2A4kxDJorwO6bPVdYm8+rEjH7AyWvnWK+mAbb7bqz+XwRymCo8Bluhgqc9tby/zyuP1toDrc5rGXBD9anXJ5SG/ZDDBghwPqvRl96ToIotD3Et8K6ze8k3M4JG4XBH6q4IpSXFmDGrTf9W/yQAqpHN+vuhRJU/QlSAq+o+nsJfqs352FT+xIeK8lgZdrmk6/S833wkSwCjs/jT3lR8mTarbfWmLPEAc1d8Fkc4hHsIlOqeBIAymvmrFPI+RrMnkm7Yuo0wSuPXRrE6KsFOY9OUFo2yKNCVNPKULc4+bTDYVdG7BGrz43OybrFfbbosSN6LXNIb98Um9B32GFSpqEFRFAcPrWdkYnGwFLPYo363SfFON1Lxrg16Z/v5JxacGLlZv3aJrDz8r/2cZk//voTRRYf++nZZIEedPoQY/zNjhlyVVxUQwRWkBsaS+cf5WST1jVKvvIsFhpLv1lYA+RASGz4WKtqxet/M+7TCulT5EO7CD49uuEw6OMhbFLwWEhWJPcXkDZsDCfDYN2Xmgb1cVa5F34ySpT99gxjS4l5yyogYbNAFyt2IKuEfJoJ7CJTqngSAMpr5qxTyPkazI2B6B52Dlr4tu/3UV7luRBW7b1n6o5Z04QOFnSR8nkP8HH7qag6B6SDkFcHoIQUw21W2fYZGXhedD7FW7NDvfmY6ots7yi24Fp+ZoRYwhk+07M4GJvDLwt9sMJwnpt95kK5P1AnyvSlrC7QfF3VheJTW8WeotvwqqIs0mTDg8aelR38IcDLuAna1MXNpXez1kwSNbGQftLTjigpg+NRtpIQvjM9Gyd306UxJbUbwnvcd5LeCRSnL6eq6FCDcnebEt3D+WTeQGkls3UqIYyQfa73HHWpk8vd/FI0l7UteKb8DQviZjMcmSYPCekDCjephFhIViT3F5A2bAwnw2Ddl5oGMj0yOHxa6LbTlJRWd2VL51sR2lweK+hhV8PAVcodZanV4yrAcuIoNTfHnXpWRQT6dcvq3xGd3dxaDzqhjdN4084IhHqZngIB0JCI81e2M9OzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXimSuk4LIGcMzEEDNsMnq7l4k5P98VzELzLlM94cU2rceRu1fbP3wuEvshpN5eFal5R0CefYaQtaHTHGo3RfApTwOe0wlIhCgtxB9/t1cBCYd86AU1FC+XdYNwHhs40ArcNzp6PUdPLlgCqYObH6SWg/4Y1aCzVQ7ZXKdrIpfTEy2kGVE9nS7/gn/tTBPjWs1bsLFyeyTcUu4vUbt2RwkMt9pfkA6S/6Feg9si7MkEXVWSnYp7EDpEasw+emjfcB29F8sK0CGU3DAxZUmrPiDa1VbiCf/331NaTmHwNmUwyIfLSYFYtwjYMHnTJB0Bggy3klhHn4rdQs0WqqSXJllBjk0lbbkA7Hb8QxLoDc1CnScvU71Y64l1zO/0yznkAJkTpUBSdOaTIDZd+utmKcrynqhCdVAxBCEjqoWFjA91II9KwyHYiKekKnorHCY3DzrYMJxRUGDPCklRWJhwC0EVPuZA4KMBLmIqy4wKv9nS/f6GdyQ4K4OpFFk0ItCGGLw5C16tYJ4Sn882lWW56zHBujC6Z/9QMl1rLCDeORFmoU/ymZCYscO7Gbg+6lCN+6+YWDbLL+tzC6vek8wBQtceYhMkUbJTpM1K8r0UAcTB7SOKgsRmzwH/jokPxtoA3H8LuETqjdOIlJtreNaqx28xyZMs17Ko43qGa4q9LalVSCguEiVVfjvGAj1CJodpFrvYSsMh2IinpCp6KxwmNw862LRIgvCoUTi4UkJiLOsTgTWsqjLA2/A1x1CmxthyeJfJJqR6fjV4bLYpSWPJv6Fj9xB8se0JmKY9y5EPA0kiT7mSHaa/b+YyFTQppUFvASyEUbW2l+3qJM4r17+U6HTrdMkOq2ZKT7606xVmmN94PAHNO+w5A1PA+meg+XHvAI0ie7TLp/z7wvn+67WtYMICWZKdinsQOkRqzD56aN9wHb3YoEPf6Gsv7/B6W8DRIHstiOiTcibJbauaXNlRCVWB2Nvgl0ZmRFMYuuj5as9cPquJpRVBvBDTnD6jtquxNZ+KEhQIzWBzxyYFZcLmvUqipXond0lGq7+JsMKGOoRAsFh58JLkeS5VJ0wEv+4jEXFb+7cZAmepIEtuMd7v0Z8xywterWCeEp/PNpVluesxwboaupfY9KU4pfrJrnz6ftPFN/epRYxieGcZkcbS6vMxnqd/FgnXmlOEnQVAbcnhjdLCpAjskai5yD3GDAQ/2+29aQDUrWXZJoImLotdErutx5883ZrlifN/KDLKuok6GzEQfLHtCZimPcuRDwNJIk+56qzQ4A0xE3aHIZI3PNLDlJUkgvfr4RLdSWwvNZZxScErhmUY8X/mF1BEcKikdt8Q81G/vEhb4YiQmK/3//bg7Q6SzX6DTetGowoHdA+0tAd6J3dJRqu/ibDChjqEQLBYpJzyK1t36B8H6sWL1ntAjgoB97GbeUY4qc5yeXKQXrpYNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S55Jng6P8pP7Sl3mejjY4LUmpHp+NXhstilJY8m/oWP3EHyx7QmYpj3LkQ8DSSJPuc8NlH0rNGQmgnA8lSrqwz6ndT3/QksOnDG4NQspg70vgb4O7tkn/8PdfXBE4cj4gbWJ7T+++tmhLieSpt1fUP44E0pJdEZmblb/ggU0FOumO/l2A4kxDJorwO6bPVdYm5fg/VCWCeFNr1Pahe7syhBxRFSukgt3aXwVgaa40RyrvICFw8FdYG9/dCSYagHOgqpHN+vuhRJU/QlSAq+o+ntkE/xX1kiU6Kw3KVPpzJKbtklyEbIcG2kvkg9FSAogsBQ7DJCc9v9Vu1xUjkvinMLQwBFZ+fEqbcuPe+3ZpD5xyMKsgzxROiir2ICLgG7tNMDo9QTtj0iXrG9jcP2gVqBVdq5q/u0C3iXiFSFa3xP6vVeJP7/IvTJlMtn12voQ+9W5jFGQwj+s8nAu2hx4IK7VkVgE7Be3ULNY+t2HWag9DiQjVUb8PlqfUUXXazoI/8kOq2ZKT7606xVmmN94PAERz+8fkRlh9q7JMWP7c4E87tqzupqEh68/dI+UmGRlxKnV4yrAcuIoNTfHnXpWRQSGoxYrXp/tNDZ9ia6WfRcp+AEdVLsgXCKb07ok8NeJUtOzOBibwy8LfbDCcJ6bfebBs7H8l2qx4v+iaFk/u6Rri/aldxj36peKavvgBlfsPTNyHPBn4/+SRklTijjJKUNWqq4V1ZW18ovKxOebhZw1QvQmnZovxO3YEXEu6yDupFg2yy/rcwur3pPMAULXHmJucKI955VRowBUTVNAVLdLhUVZUVzW20mXG0EBUkTeslg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jhQ4i8HABxr8QwsOk4AmyDOZh8vo7YwFFMYnP2RaNkV8b/gnIeuNy8bSk4JdkA11pTKcb9SNY2cyj20LRnPQF3WlKIQNqBi6EH+Naps09zl4/jOgKsKUUtUR+vwS3/CZANL/BhA7gJP0JJcdqoNy7NPQsgBR9ISEIw2vAqDKiLD0rhmUY8X/mF1BEcKikdt8QxbYHV94JWaXiA7xycORXCLpHjLEEd5VhSseeSmsOG3iSnYp7EDpEasw+emjfcB29CoxUpVZHUqjgnB7z8Q8rBXn+cloNdzPduSV5Is6kcuZYSFYk9xeQNmwMJ8Ng3ZeajujXwnpenTZazkGIZhQLxGErpZIx/4l0fKrwClyxFs16J3dJRqu/ibDChjqEQLBYdm4GNGQgXZ8WA4lCsFLw8TDvQKcjjUpNtbBskODXSo5YNssv63MLq96TzAFC1x5iEyRRslOkzUryvRQBxMHtIxwxx/83OmZjwp4Nkyny+bdJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRWqqHroQk8qZhJYzuHsNPrSfPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuS2rCBfcsOhSWT63nPlSURTOtAX2u7TEGHCqO2rX1pgE1WeqPYPT7ZsCUWSJeUEBxX9dkdTC0n9wh+0Bj4qrtsQTfhPConPzywi+clRBfdEwP4zoCrClFLVEfr8Et/wmQM/nW0tj3uv9zPZNNfGKBkVpy7lAWskMt1eB+IF8/u2T5EKCo8fKm4y0XObHAvh198cC7ijQ6CLSIcVsGbSgtvTmW0CjSZ90uhyp5j6f9uSQVUmrhTT9me7ilxb1FbZvdnN3szcKRVRRadbaYtAdkb8A0eqB9K3ClnvaAWpPqhMj5d/sXcsekstbjfmdUcC5Aw/yyPHsszH/L29+DvkGZQkBgDdtsyvtBk4Nh0/h5+yFb4e/Fb5RHJHF8ZQqvigH4jgoSCtjEDTD4cwQ35XcYB+Bmh9IWYjL5ud+mAtTRRP/enEFyHoy0IAe5VyXUMtVqrEjei1zSG/fFJvQd9hhUqZslndSfltGfz0CMRN67oKdRTrtmyb3NFzXqh5hf8ye/N3D+WTeQGkls3UqIYyQfa7RBxuKha71Po4RqPyERaJfG2HWtIGY9+rcL5mNH17hOvt7NrlluU7+Sf3z2vlvC0+9/hunnapG/NrvdrxGIm0U/OXoZKAKSMwAXHBFTnNtxcDo9QTtj0iXrG9jcP2gVqCga3a5YJXauBBOR8A18nxZa/+PRMYE4PPzZpzihUlDLt3D+WTeQGkls3UqIYyQfa6oJi7nieiQRAooWoWM6gUSzxukwtkLYXWGf10mq2uw7RYf3rcJn/lbwenQwvFmQy7A9bhkJadu3RgjdfHwEso2yK21eIoLgzQIY8bGEOD68skOq2ZKT7606xVmmN94PAFg0DbOHa018Z5FhzSR9niikzYonsIVCGozPht9nkBSEJKdinsQOkRqzD56aN9wHb2keVuO6juG3/Dnm6Sily+ZwORLCJGcoTLX1XPjf9U6XErDIdiIp6QqeiscJjcPOthnEtl3Y3ingn4sWwt36+ECwfHfTcVGQn5yyj+4h9ir+TLnXkWBIIkZDIJNCtlQdyL3Tq6+bngg5kDTjA9li+a7vc3zCz2KztYjBWXqFYLtpSkSaC4Cr5zxQxHo/rl49lZYSFYk9xeQNmwMJ8Ng3ZeakR7Z+tT3YJqWn+TI9mLjCdBfoeVvYRhLaYGTZOY5JgCp1eMqwHLiKDU3x516VkUE+nXL6t8Rnd3cWg86oY3TeEbkro41HGZYLjohrSPnJo7DYVdG7BGrz43OybrFfbbosSN6LXNIb98Um9B32GFSpv5ZVZ7AW7LDfAwI77JMVy41ooQDXck9X/jVARta0D+EFh/etwmf+VvB6dDC8WZDLjeJku8yrMTV9y0StZGIuoCPL8UzkOJtj4RK1wLkm/m/mwZ00TWt8pIXmk/HSXyK/xwCAVvulCkNFMK4vdPe3jeYOVcqW/LLbb/3uld3xzhaqm+4VTnAclqckH0osE+MXCL8k8FCWWuUVeHpUM91DmB/yFon0t9Oxow99W63MjaCWEhWJPcXkDZsDCfDYN2XmgUIvYFV/h3CYV48tlD3FPP1vpHGkbpVwFo3rJlDviybWDbLL+tzC6vek8wBQtceYm5woj3nlVGjAFRNU0BUt0s20l82ancDuCkymcZNLnwv3cMsp5znua27QCCtSCOSEyg0hmlilIGAwsJoT8mlO3djTPDyCSChGRMyKIf6YBstcGAdlNJborbP8xLGJJ95LSPalDhqRirewYtxhC/RehszfePLtrKmZtfi7z0n8Va/M2PQj2sLvO6CT6F2x+9P/w8d5KMf0gXk0hoqmn3qwu+IZpqM13W0q7xmKhS4jRVYGdb63LSFqhY4C35CqqJzCMtJgVi3CNgwedMkHQGCDLdV08gcsZhYBJ4ABNyhjKcmoOkMXKmwbIIic99En7Ev2anV4yrAcuIoNTfHnXpWRQSP4/Obr+FNj2EiLn6Njig1E6qJptV6O5sI1zpu9Vynmlg2yy/rcwur3pPMAULXHmLLs6YJLMZFAg22bRTvnw/8aE8KEmZIsDtpw1iwTNXVLFg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jwjEHwOGxiqdVqdV9lIXBAdOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXiTovrSLcdF9DlptTbIVnsWOZrnNY0HkpYBHixkKUj0vQV0z9ereQNH17fVzw2guej810Nlv1sRY1VgLJ6hVDmyms961GofKwi6KpRRGIFrpJYNssv63MLq96TzAFC1x5iEyRRslOkzUryvRQBxMHtIwpbrV9mplaBPAasVhuwRGtKwyHYiKekKnorHCY3DzrYtEiC8KhROLhSQmIs6xOBNfh1FhOLlGwLucMdvJ+wQ5H4znCxt8Ds0uLbyunv+ZK3v5zwBS3jzC3aa3lt0kNsaLzxghVHSijjyHe6XLOjDgEs5MjN1obbuG1Pr3UWkgVAujlE1c9lA3DT2UJnHD2bUJlPodPwEn84xkDjXyGRAme9iqxuPbGRFXpwImyUQxq6qm+4VTnAclqckH0osE+MXLJBJnqq+JcS1cASWuhdMFOxqmOazoYMjxmvZOoXh8e9eid3SUarv4mwwoY6hECwWB3vBajUjIl6f7NLYF35XTzCkeDQP3cHj4Il1bo3KuJLFBPdId2OMR25LIQPpOYDMjza0IXS3JbOw3Xlo2Iyu4qqc956kLXscESRvStGfJnU8seqQe8VMSFHKFnIEYxKFAl0wj1k6/5Id670lY9lVRwT9lJtv7SgRQvs39kfcguE35bqiVIwVe3siKBnFHr7ABee+cRpOjjPl6AWFf9xbVLoo1A3mRj7yqeCOaQiPM64Lc1wX06FkWd/iXNkMTzqyGSKkBFEW2XEhl55/ojywbq1GOjfdUvI/U1diT4oHCtUTqs9r64fNCET3g1zhrr0mEHC5hdtZ2MibvJIhPVlM5JB+KY2XMWQmSo3g3GWOTt+oCg5qMItg2uo+RmEef77ag0Lzw71sZUQlYT46PTdQP8FRutbxjzBPEo2lAZHJwFSLbVVQ7SOQUJ0wFfbwfMyprHcWWXP2PuuddFMxLJ2JN8EYXyeUSnL8yqy/f+OYMf+LyZxn7kNj2rSITAZk9Dp+N5RSSBbljRhU6N9aAktlp1WoBfgTFtIE3Zr4QLsyrD91+kwfvh1IbLoI1nj/LZfB5nrUmUqm/PhFozS4SsW4ot4D1Q10xsVuhYKSBlLw4Il1pswaRI04Z6bdTgnLpQ2otKVzjuHPj/Tx9mOZxCjTfUKGj2WVLfVPcj1sZZ3vL3n1m1o5SrCQbFbg21PCFROjlO2xgMITHaNYfjSCMMb8nSXsbLX+DXbJWgQXae9vp8EBGF8nlEpy/Mqsv3/jmDH/oQJUhqEH6b4w3atTDH+r8grWVPvFUawSsrF3PnCR2eyI9qUOGpGKt7Bi3GEL9F6GzK9HtxUwBJ3OvN1qBlJoEcLeKu/S0+3p10CjZxQ93c2gCLTgS62x+3X3bnFB+g3HQ8NNsaigDVikx5cnbD6ojb3Q1wyZrt2QPedUld9GpnCrUtnovVhoLH3xA+jnEPmPgyhTsjzuGbGc6VKI7URdlvBUpEARtsk7cL6uMbcBZlIWDbLL+tzC6vek8wBQtceYhMkUbJTpM1K8r0UAcTB7SO17MbjzAS0CDvCnJVqHKUmlTRkg1+obHu0k7LpZTbq0/uyI93iBCYmWa+Z7nDX1quG0C5KAvFuT1SM557L8rO2nzzdmuWJ838oMsq6iTobMRB8se0JmKY9y5EPA0kiT7mNB/tjJ3zoIz4i5p5zXFAiLRu3RO3c/9k67A9yVe1tvQiS546ZZVQLe4H2a7pOUmQfUnkYfVbEDnc4ZWw7a6yLlAd23ZgGWVk2OVXbXHpvaw==
|
1.0
|
Glt7RyHmocTWQXh4CQT0By3KWR1Kj/WR/spy4+F7Y2SQcC13+DtDS/Dw1D8DeDC7u50A6KslorZMBPnemFl3J/UXMw9QpgFLG5SEq3w+E/mswE4EcfeLXPF75b+ysE8fcWHe6lXR3RrXl1inIAASfw/V2deikXheIFI9IDuxEME= - JuZiQzs98LCkwQ8Uju7zCbp2KefIdJBxAH/GX7CRvmvel0XGLHHK2hprZXqG7/Qp23idmcBGRuhKxIOuDA0hLw7AlhTh+iLllRwOFWBNzjkjVaOrVvgmz9BWOrb2qxrc7ZJ5dKxOJ8RypmEyGCZqaHY+9ijgRpjOTNUrKOee1wR2NbIykPONWmtGE/I+Jmu7zmDrikp/Io4Y+OJcurV3b5vFYQ5iveYP1+QUXo8ODLSqRzfr7oUSVP0JUgKvqPp7d2T2WRMvHQbyNffLTcxP+j+9Tth8Ti4qlWWyiTiGapEGRFZI6qT7X9DfJkXwGSc1OH5JPKvGoNS9MugWFooofxQ7DJCc9v9Vu1xUjkvinMIh5Q5WCrrsU7krUSC0kHnSeffkOiaTM6uWVpcMqqCXk+B+zDjLH9rWZERKj4vSYiobDpE/MVh+06EcQAxU1VagNTFTS/m7f2tydyLChhjwc/pKhlm0R6juUnBAOTo872AYM7p1XzipucwvF9OkTVU47CJTqngSAMpr5qxTyPkazBtRYDPVUIv3pfUbt1C1x+Uc6wbqlzIFNstQY+Iz+U/QU07FNDfk+gxBkHqc0TYkWQT12srC0WLcCGTUlWqLZ+2mvVstgsP4OppIAq0af5H/MuC8UFlhKtxNiX/x+bYF5vBx+6moOgekg5BXB6CEFMOpp/TFl61glswcqHBosNF3r8dirVa3Xy6jMeVVY+yneNOzOBibwy8LfbDCcJ6bfebBs7H8l2qx4v+iaFk/u6RrTsY/5UcHXHczMeybssQlMy5Wnr1DLZSlcbsuypSB8SBujJotWnUJPtbmx0wcjcp+pvbPzICk0aBwQB5BCeT8ITkaJ5mYFmKPurCNQ3UdbmV57oqxUdh+U+4S+XmNSsEOLWEnLmfy5RiBTYZiaIFOfW0tFiPH/NgED26PPPX1ItnLSYFYtwjYMHnTJB0Bggy3vrRaheRIJrRhbMLmlelntZyWzQOm45hAX8ttwEX9frHC4JMLM33Uj9RFBZAuzHMw9EJEvNnNI981+g7Vxkd4SV85F1VqcY0fYUnpddOVjc0PVo+yMsGhCsw4lIGKKLJdnA19qRbngXlnQ+URrOtqWHrDU0wmZsBwUh+97LZJ63QZOOzEH3dj/ksoxxicZgbkp38WCdeaU4SdBUBtyeGN0gn6cCU1IuH7Jnoy5m2IcpnIaZRKApBPFzqL3JFgafzrnzzdmuWJ838oMsq6iTobMZh8YXPnV06sYjPMFEKcG+TkG32r0dqhB6cUI05Jbh0rzpvyNJnVMmRe4/Vxand39+o8RV+M6Zt0IdDHcI/DrdJ57oqxUdh+U+4S+XmNSsEOO/I9wBvSeDOfv+B/GmxUKkPfs/0cpmkadOvBdfTNw3BEM0nBaeiy8R7sQ+lmSv6rYefTevoZryqSV63YFekEhbMNmwLRnoyNjIKD6sCJaYhfI4dDlZBp4+CSMPJ8ZFhORTUFtD7xvIccROAF622A4s5Vlyq5uBMhkclKE6exTeFYNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S7hgEFsM8Q+rhrMzV5k8dZBJhhpu7vEkYpNGAuqciOYVwOj1BO2PSJesb2Nw/aBWoF3Wl+HbBZkWHnO1ZvTeUHqQxNEyz7Bdnedfx9JzDSjcPq2vDn9gWT21to5D7Ces6Y4mewISwDxm2M3P+B/IgJDNLVn+bPr1MXL3ElC4PuoePsZv192ZAoVKtphpYO/Vbc3r9Eswt8JgohlQNrc/K/Ymh/wWaH/zPCvotBBl3Qu1edhMo4M2X1Bgf97uYkXfXhtgBfUfRR+9k7HwAOjS+/T+QbZtunnKec8p/oRcvYmrBRRZK2TPEigsCTs/byfc1hDLAwY+2sf8zR2tbe6I3d4+rE07PbqaCw5XBmi7SzYEKugXiy3k8/Rg+aWumZvZVi86pHL+YmsMMS7cqDzJ+fru6RK2/iergcwaQo0O+DQ2aoeaQoWfXi/7+/rvrdD35/Bx+6moOgekg5BXB6CEFMOpp/TFl61glswcqHBosNF3d+tDg60bI7Nfe+0F6Wp4XtOzOBibwy8LfbDCcJ6bfeZBbHmwaVUSrhqGP2MLH4Q9wC5gUncCuk5ceFC4/ABt1kmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagMPnYwzy5FvSaGUsybv62JSCbfRhjBqvcR8UlrQIysbQANOFBbhYmc0AxKBfytjphuv4y3jeYqcet3gLS6pNrx9ReqLhjevvLzkxSHtXj9EXl00jTwnn3RAjpbze0Q9C3Zy8rJu+ZHL0iTuOk73ab3WKi7zlnYqHqt9rnnh+Gg4pYSFYk9xeQNmwMJ8Ng3ZeaSqlxuXaZmLtKxPndGrXokGhXN72Z9ZgcCXrE3Og9+QRYNssv63MLq96TzAFC1x5iv4QSofPEc3wl50vE6XI1Q/o2HyqrO4w47Og9uKk7uFKYjlgOpXApd52n9TTPJFBZtoO7z4CXWYyWbaDIJ9YwBp3mGU9jN5YPaHytgVjzxPxpcnlq1718+7t7NeJnVteFjxCE5fcK0/IavHLG7D7ol9sz4u9UnArnMa5e9zBGJCcJIvN9vStY3A2UbqnirRP/06g/HAnJK8RxAgcaMqIegPuZb+ymtbR8Cstvj3OJnrwK02lF2TzRHy0BpA3hggLg/zQefcma4VEr5KgoJOXATxrzkwyNhvraMaOe/t+txyQXncd248YVsBnAAAi32q3kirzqscXrG+uvBOaM7ptPYr2EA0r996EDFl0/nXBdNuVTLiN5cCcP8tLD7DzEWpkisX3F3B/LAusojTszMcf34pJmJrJmrRZXfCkmTJVV3ul5mypqMCD5aH/a5DUoOjM3CtNpRdk80R8tAaQN4YIC4P80Hn3JmuFRK+SoKCTlwE8HBkfgkbuRQAhEG4mqy2Iwtq4VJ8ubDq2PouiQqmIJrdYZujLjfyKazntDCTobswbVTCroEJwbnGFvtyo1UNLFcv7vpwfvku9j6kCTtkuBbWKGUpLwXUjJg31qO1xrnrIjQVDNHcPtOlWB5kiGve7+MZZZ/douxw8HHFFcFtSW24RFuXEllM0av2zZFj4WtAjKTG5NVJOmFyNv4QXQd6YS31LAkG8RpDP3UaHOyv2jyOt4KvcNtUm89vi+D+c8I4EKaM+qr2ydNN7CpmP6MyNEl+775FVJ3kHHNA9lkxW/RiOwsViYb1Lu8NjJ27XepPkRy7gKsseeyv7Z+AgSQTbIMZZZ/douxw8HHFFcFtSW24RFuXEllM0av2zZFj4WtAjGD8LS9TGLN3Zln/aybHgMq+RN3X0uW0glMGl5GShYznNyWidSsIc5fSI6LV481LYKaM+qr2ydNN7CpmP6MyNERxp+V21lGFj7hUe0XOstcazp4JRRrzlKfv15RUflLHMB78kE3IYfad928d/LzgvK6uFhAGlMBMgDKchhlHiFK9sz4u9UnArnMa5e9zBGJCfiV/NCFWayAAoaGZdQadU+4Z4lqiBTjDCJ2K5U18UnrI+AigzlzEUSB3Hz2HxQJfPpW4yfQyt1WQK3xgsPUnijpcuOn90Npbn9WgVwOLDYglKQIrOUhQK1H5nPE4t6UHqeez3vJ7bmH9ImKZQB+kBpYI6RfTV8dwzxiSwedvtgmBvErlabTKBn7E5Ykxn37OZTUPRwAeFl/aNU9nzjQArpGsKwO2uDUI0VH8HJIXDN0bVXhUux5x44gp6ULKfff+Exlln92i7HDwccUVwW1JbbhEW5cSWUzRq/bNkWPha0CONgdlG3M3bS5og6IWpf5RLl2WK7+yYZpym46HX+Q0QI6uFhAGlMBMgDKchhlHiFK9sz4u9UnArnMa5e9zBGJCcK+9rwMecSGBWXXNEgpS6WGEqFxm1gxJmRgp3zAiIBYgkWWYRHMTNSkw2/frY/xsorvjspZtlJI0lxrfsJ1esWcg8G30j9rDJQ+2BaTo60YgE/yzQ6QqTIoTJVFUho54Fdmr9A65A67liehKLR30mZKue36Uc8pgAJplvqwXjiyb8aZajQtNMxM99zp9+FrMdOm+xfXStVoGxqM0+vdAbIuFjcr1o+5VtZjbaG2gRT6oVpp67jcos/uOXYrSVwggrFPV8ndpwvI+Q/MQZngaqm3nYa2w2ESC5cdTLowJk7g4mVE/kRJgNI0lVR+6NE7nYju3ICMCnUNhEB6//uAo9MSouFDbBZKs08fbgi0OPMSr2EA0r996EDFl0/nXBdNuUsPSFcyZk+tzeZIhznNqHtsPxKVprhUegoj5KDZFtTkuqqXcFabOwRILBso1m2pvzFPV8ndpwvI+Q/MQZngaqm3nYa2w2ESC5cdTLowJk7g9qI+UQAsh85qWrBw/GFQxrM7YBTaBNnC94pyrjFn8BB8ndeCYdbP9YGI/KOdjMvBHLVhPmGjRQRrIIy3MmKxzy3EcrXz5V+eXiezt+Jc8xaMtkbE7mjZehayHPh0xJcrU7MwNGl1C3/frzllcLbf+wrvjspZtlJI0lxrfsJ1esWcg8G30j9rDJQ+2BaTo60Yrn1acRPxc+DUH2hGD/dt7eGiSuDjyc2AiqV9KKDsTs/Kue36Uc8pgAJplvqwXjiyb8aZajQtNMxM99zp9+FrMfdiSX9hKFyg4zS+hmHHWRpZssTyEVL5TTtlsoH/whgIMmnQMK1qz1B1OK5BQZnoFZhojrjw0SRme2V9U6TcIHxP4zoCrClFLVEfr8Et/wmQDS/wYQO4CT9CSXHaqDcuzRGuAN/OIv7PVrKQLPzqWwV8HH7qag6B6SDkFcHoIQUw4L1OrRB3gJV0p4mERK0S6/TvY4yX9jOSEB9+UMzgPit7jI7KdsEHLl1XdUK/7nNDle12b/Mz0DN3u16IeGf06szN5jdCNLM+YG/MLGo+L1RJSAt5ZsUUQn5PV3kLNCjroZy2vwXZ1FzYKQqDyFmBUobBroYYM4zEsC79aC/6US+5qT0nhShad9E4bINs6SaSy86pHL+YmsMMS7cqDzJ+fqVnb7f9uHvdI602jrNBUSjWHpZideT66fr0pvqy6EmY5OEIOPyU511Y1offfcJnt6J9IbD3eRHP69yVTj3LuDr3hfvdQOf6vL1r00lDJZp/0mGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagLthezFjqpzhhqAoQZ40M4tjCX9r7BFQpeCuAfvc0gPhJhhpu7vEkYpNGAuqciOYVwOj1BO2PSJesb2Nw/aBWoIhqgpoOWa77r9cwrpkSewaaASHoYiIX6C3fKO1Ib90aDF1kIWLIGPUjpcC30HKRh1Isnvg3AAlg87q7sgwtfLm4kHKWytIJhafCJ+dehTyqBM6VmPm+Y9iff+B/aLh3fXrTeZbs7zzHVvp07zAS2F4LY+hIzB7FXnpT31tpqcbbLzqkcv5iawwxLtyoPMn5+vOn2s1kQJaRCb//+H3gwxs5StHhe9nI7/YPllB7n+v4T9WeHjDD73j77R+ZqqCQxRQ7DJCc9v9Vu1xUjkvinMLVVhlaD/RhoF72CucZ0HY0K4ZlGPF/5hdQRHCopHbfEBDoOrkEkjLNZDJGa2h+fG8O3Iknv88yZK4K2vOhXX86cIWQEFnWE4a1ResMhXQ40MhmowvYXZLW9Q0/N/puDd7EshXM/8Hkvwz47jtwtd96Vtpxo63u/fbhIPzy6CinfoUNc/FKsDq04Oz22UPnAaO36tMlnIoQbfY2l8/3DEmWBzAFYjzzFb3lzDlFVFaRx0VsN0digF2vkeOtLuzOMvKJW5l4fxU/eBPTcQx5ip5tio2n7F/8UttVfiR1GrFuyT+M6AqwpRS1RH6/BLf8JkCounIpmIP1/BnRh9k6eFTks2nf1pSFJy8xPuGsvSCTySuGZRjxf+YXUERwqKR23xCuS0uwQ6xpmUXujSGJK/sX7/6VJXJaBiD9cJzzkY+Q4Xond0lGq7+JsMKGOoRAsFhL3ypJ81C9AiFLX1cgICNiYkk3ucLis5WC4y4s24dCpkXbj0P4AZVSOzgt18NDqfdatwFUj/LCI+xZ7RF0+SIxncf5ld0V2Nez9fH0LNsIjqVokvWHj5t9WNgtegQfAASelhG3ySmLH4Y6RDraZ3a5BC0XAiDQyWjHW8pwhI8iI1cxnRiPB+N/+Ir8xgolz8mNZIhoUov3VT3zfG6/G+auPxkSjEKaQ8G2kel5MqUOrL1QTGIHx6MoWcHFD/XyExoRyR+0CC2eYE8Hf/VF+nemVh0izqvkfEwR6pVD0JaSBNOzOBibwy8LfbDCcJ6bfebBs7H8l2qx4v+iaFk/u6RrFTSGbYpGdOx1BTSuW/For1J3RdKTYANBB4rmbR1Nfux2AAuwaIk06GTSm6pVh12vgn62eokieuePJyiLiWvZXQr1k343lXAyxnekjnJFCypwniM4oveXZKdo9GtvqzlleYD+SGyd79vIgKZSbl930XpatThTcvQOYIWWyssKU3+LUHNNv5tYUYnloAMaDomWLVAd7SjEvN2EO4/vXCHWViiYNECy7Um0e4WtLu0pSqG/8nk2OXohhKNRT3NiyqHBJ17SNtzxJfHfCZxJAh1psLqW/1rUq7S7RffgDnBc7MNJhhpu7vEkYpNGAuqciOYVwOj1BO2PSJesb2Nw/aBWoKK9yDKl++trT52bEk16emFA1jInpXDznv9oHr7VtVyU7jI7KdsEHLl1XdUK/7nNDmMeo/bsxrna9xweVv5gxUifcCLr6Z5Qwwu3OIN3BQ7pJqR6fjV4bLYpSWPJv6Fj95h8YXPnV06sYjPMFEKcG+RNySgT7IwhpvjsK7CfTNN/7Pg08sIeDzdt4/kqFHdRBJ883ZrlifN/KDLKuok6GzFYm39YYV5K84G0UBP0TVM9boZlPxP3QW2b5lVq1gRnXPedSpXG4EKQYVWlUqa+sC8vOqRy/mJrDDEu3Kg8yfn6JkUA6kYStX0U7gRCyT0RRjhqKXmGUuR1Hp8YpHdcjd5YNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S3Z9dQwjLxUvsnFkVB8iSzzfK/Qwh7GXZnAxjaIIvuzjBh6G5vNhwdOQ/sToTU4Udn+xhNjwXdyEXML6cnhT0LWCwKj9Ps5bBvJubxFJXW7cgwbj+HUka1DU4dM90C9TrX+sMvk0CPwPKlKm+/oeqsLbdvQt/nzsy1/O6bVT9x+tErHnwH5a97rvqp5Y7TJpQP9NS9vN3sXWPKl4ZaQvq17lUZ5Osc1OgqU7y1Gt6wZ0uk0/v3FBO6jZHTYTFQipi4G+Du7ZJ//D3X1wROHI+IGbtIBxhzepnMmaFYS0vZjbUAray2R5PfAwcqzJSfEvKfb4jwKHlugzLJAL4k9ZA66CXMvdqlqFAqtTbyXVw0tufHrRrNzKyVXbSL9aHf38CYQYV2+035cR4qZYyHXeO8H/s05mk4KbAF6vwptTt4zVBiQCMmqAUW7skayK3jIrVF8jh0OVkGnj4JIw8nxkWE77/d1juyBOQHb8Q5/7NYlp/sHbPlvM2Azs+iCvSRL5BML12DDDt8QXVQoMvAYViOE8d8SW9X35hXUXFH4WGsXnhY5IYNrnKZppaHHtrTtLDXX3fgS1hxy395KaL44iOTmHe34yCTln/ZWnK8E+81mjAoNTcUF2MEUCLpHzEtWcI2thv5B7qAxmTlH7b2FEFCzfK/Qwh7GXZnAxjaIIvuzjyTqADHdYthVqcskr1eJJrj8BX+nEV1XREjhhwkK9pqhcWrSWFE1C2kfTJZVxJCgPSYYabu7xJGKTRgLqnIjmFcDo9QTtj0iXrG9jcP2gVqAmweYrEDRVNnchdTDcrMtnt907COuvnk7gpF2QHz3qtwxdZCFiyBj1I6XAt9BykYeZmyOryFH5AAykHVB4T893IORb6jaQH7PN+XaLVk4/Uy86pHL+YmsMMS7cqDzJ+fpSJQPDUfruTQCULGIz5E1MdoEn2fCtJ5spSKF05T8PmBIfaQi1nmEEAlJa96/A5w4EBWzN0O5l8wHeYqU3p17RtBfsexqnzihq+dOeZzfQtLa8lX6b++fLImJXUfWKwTYWRZFD5A4wg2AswPNNaHIdVtMNzR32BZSznKiZAJHeI+wiU6p4EgDKa+asU8j5GszPo/PeYa9hXJVqxijD4ZgzTdLPzB05rLY6gs9U60gkocNhV0bsEavPjc7JusV9tuh2AAuwaIk06GTSm6pVh12vhQn/lr7nglPR3Eb+BCtnPiHIlulnMLPWQ3QPK5/AvS+sadQLZPHsxtgd8IU0p1DeT5Dax/b+Ei7ycGSgwcsXooozRkoZe7PhC3r4W1ry4A6nsWTTe/yoA5ionEaQqCq1rtuJxl3/XHYJTAkioScYErw+EM9hTjzu/ra1r7Rijk/wcfupqDoHpIOQVweghBTDNA2YwdprpSRYbAQhRabD1Bvwes5mN3KZl91LnGehWsRYNssv63MLq96TzAFC1x5ij4kASygx5xnBnW1I/YjiQHlyEnTHAqnF6V2NtWxryt+fKu8jEst/dmBdtlF2EF0Heid3SUarv4mwwoY6hECwWCqhVX6E98NtxVy2pi/ikJz+74P9AtPPpVL2Bb2Wz8idqdXjKsBy4ig1N8edelZFBBleudZWyfeVXSNcYyQxmIsLMHRX5bdorW2PHVBP5EdHeid3SUarv4mwwoY6hECwWM+64bODqu/xyoutWHXze5gWKeDF/ao0xpdeTxL43cInqdXjKsBy4ig1N8edelZFBAjr/UGJ26LXHfpa/vin9stelJGu+C5//aU96KGM2zddcFbldc7Vmqriy21NGC0m/9VCPDZpY6xCSAoilHo0VZvQ024VKSV8Bl5Fr+J/fQlCYzCZQ5ka/ogHtUyJBqm4pz9Zvujn4iO8E13nw4U8d3hhKBx3uVzh/5nVPVafM36M9L1uZWiM3YM/2xcsPM/AuqnV4yrAcuIoNTfHnXpWRQQXzexm6YvPDc2lKq6pjlfdM+nWdDeQ+3HfQSZh/j7DEkmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFaglfJBZsTlL1vhUcI25xQDTWCcYmj76HpgIOm3+zhLrhM+ra8Of2BZPbW2jkPsJ6zptdPrv/1T/bIVIwiWCCtHoOimPm2OzaIVEaf/Z2ov3GUvOqRy/mJrDDEu3Kg8yfn6SW/fXKKj6JcEPcNP0tjISvm+OeEo4Ze7g4jIAOZs7dGp1eMqwHLiKDU3x516VkUEF83sZumLzw3NpSquqY5X3ar/xROrGxqwbj+WHQSNNXpPzRK1lSCYljpAL1vBiPe6AYIMJaadrcgr4bpZYiyeJYfZurZpEHURS/774D+I46os56EOD+FGLu8B0vzBNKawJ3Dt/3XNZI8e9gdyAQW4QRhG1Gn0885y65fDz4RbwbURkX8Hnsmu3Sk8ce2GWKJiPq2vDn9gWT21to5D7Ces6eY+Xmh6VwN4fO+wVmHcePa4rPxZkaRjO1QS9U6erXLti691Obc1hpZRV0yL0jDeTNL4dBmfk9MjhcNiwXAahXrZpgrdsZlLaaHCmf1xhHnHFdM/Xq3kDR9e31c8NoLno6aJ6W51EjmB8oo5eFw0lkHs7QrsvTUrNGTeWo0NekupurmAdtOsTcpzHdAkyBfX+sVZOTxUn5JPJjaSvEW1vcB9csnKhk9iAkx8fJlO87oo3yv0MIexl2ZwMY2iCL7s48k6gAx3WLYVanLJK9XiSa64p/FMzsCT/OgseBTJRk84WgB8KIlZsICKQHZvzi/XZK371UI4wQqemONKg5v9RMnOLN7FhJ17tUD8lYp8PoiEZ3YPAJpAzV4IdgAopchrp3ond0lGq7+JsMKGOoRAsFhELMMUmWafr2e0ScV61sbynM/gnmw5BFf6KgVXKI0katQPQSx4MUAPgJN3lT0MCtBCuT9QJ8r0pawu0Hxd1YXiZVT2kMCwNA8ykdZ19l5iI98r9DCHsZdmcDGNogi+7OOjsHg6jZ4gUlFbdGa4KR7XrOr1ATsOkjsDS0uM5YSkrQEFRVdoqAjSSscAYIlFglrVZ6o9g9PtmwJRZIl5QQHFN1O4Y8NE4ME8sQt4LZF6oW9u4dPlwjaL6U3Ptcfj5oRJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRcFCaSpmjMrHCUkdSuIM+rWI1grkuTJSd8qMWJXwbwPc7CJTqngSAMpr5qxTyPkazGuKXAtO5kqEmtSyovlGS7eMCm2zO1791h2lk8Ab1BP5LlaevUMtlKVxuy7KlIHxIG6Mmi1adQk+1ubHTByNyn4w3g6rBkUQ5dJ1R6rd5+mss+uAH8Po4U3OzPQV6sn0Tod7fjIJOWf9lacrwT7zWaOH8EQbRSSiMtQuw6b9qosQke/4234+cjilFW+OGji7hdVCPDZpY6xCSAoilHo0VZvDhWbhp8fDe9JoAk/WtmxZZpQMty0BKbsZE3GK0Gvcl8N3IRTgtys+dStq+haDyXYcsnsUnmuKJhKMew3M6SMyReaI/tLW/xRJ6Rhoww/5BvBx+6moOgekg5BXB6CEFMNalosR2/HmGWOPyzh02JlLHv52VepHrazov2qOKXrQK1g2yy/rcwur3pPMAULXHmL6NFcnG9gj6gOgeyhnaJmejIhExj0GH6IEIfyvratKR6nV4yrAcuIoNTfHnXpWRQSIm5VbAsb9QdbGMLtbWP7L/fVMyj4r3/ibvCjCwTgQOanV4yrAcuIoNTfHnXpWRQRRNkK9roQLqZ127U8lWU+brKiiKNXPpFz7LitTJ2YdSdOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXi9mvg7HjCB2ZB729fJ1PPksCKab2A92zS12JHXESLWImRu1fbP3wuEvshpN5eFal5VJ191zeeirL+OSvgPRwXfCsH4i4/wgBkRN7taXQcyO3l00jTwnn3RAjpbze0Q9C3rZ7SPziPqh+jUA0mQvFrdgXpGBPHvEoJBydh4ymSggM2Xetcu2rP8ONkVwiZz76C627AzFDi6o2Ij4KSS7KpZYV3RXFuQSg6vgCkpfenTFs7+XYDiTEMmivA7ps9V1ibtbJXADbApfzkcFmWaeaKjJiNISfDeaPwpzNqdLR6wk2s3LLJTHiSNXKB/CZN8+Iet9Px+FiG+rVVo4ua5SUFmp99iAMAGTUDVkty/bdA6sZ6J3dJRqu/ibDChjqEQLBYD6X66m5s20rGYTd/U8ACRX5u3tpyefS71kJ80z4EovnQmPfE5dP2GOrJAH2VZWT7ZyS4odSoTaWlTEs4nQAioc99IZ+gG9+/nEgVTMmzcCHPW/spACLEbKQKwqHco94gP1m+6OfiI7wTXefDhTx3eOzkFQqjFZyDbtPzHdMAuxT45O083fRksVOOc51uHAov07M4GJvDLwt9sMJwnpt95sGzsfyXarHi/6JoWT+7pGsK/1EtuPt+UWc6+tW7VqAD8iKZwGDpHqgkbtGcaQGeD6pvuFU5wHJanJB9KLBPjFxDJA1d0k/6TPw2aLUX4EM+aAlwU5aAZ/7lODfs+ngCGcNhV0bsEavPjc7JusV9tuh2AAuwaIk06GTSm6pVh12vYjUmOiXrtz0fN8hS6FouApxD1JxMIPqOB5uRAkwfKiSMFSRsYvNzU1uFSv7zAPmb39N/Hxd/YWJtzMBPqKuv6wBJLFcyYZ1R/LGlVAJahhvl00jTwnn3RAjpbze0Q9C3dkXaqiN5T+ny9idbhUsz50yd3VK46bpXLVlZ0hX2WXep1eMqwHLiKDU3x516VkUEhqMWK16f7TQ2fYmuln0XKZPdbW57chfrEpJWC9aFT3iSuP6KoZkjPb3tNm392BoRIBjc2ejAjkZ2VIZFdk8TOFp4Yhlcz4gpKWSi3fDQIq3riHKRD+FFSU4fS+gx/48MTTh291hchHIsDHAUwyUXLptj7mfguKp7KrheFGFciAgxdy746+oKT/tc1tAVO7jkyQ6rZkpPvrTrFWaY33g8AZcZAfhxWqgjJIGO1HqABGP7ZjpJgyzVEZwbjW6CQ7mgpGDk3Sb8/rflSc38jjsugCFnYhbbDN+Xgp7fAAxUvUzi7owc0UwQw3ymspe38Pb2WDbLL+tzC6vek8wBQtceYuXcWg26dlY9N4QoZU44thon9TtVOFcyp2GPW5JxKwYC14DpHO2fEAhcVZBu3L+FI4SSFVTlKL+Wzj4fU5qZPGfDDMek6ytjuXpM1q9Gt/NWsR6j1uE2bnUBGahEZWrcXz/4ktpwNnB+mE/NaoOtbjlWGLta9TPF4IdzJRPxJJGlxYXBUOH8f6bMB9fAHx1Cd4d7fjIJOWf9lacrwT7zWaNsAn5oGRV4GqC+O/x4HODLL264F/dnXU3z1yslcgIcNT6trw5/YFk9tbaOQ+wnrOnxTPZg34RTNZl4vNoJY6rkg7pdwrUERgUqwtNaO2VpMD+M6AqwpRS1RH6/BLf8JkBpWjsRqGIbbsAXJvDnUEjwTDnA+RKDimLFCd8fH0ABAHond0lGq7+JsMKGOoRAsFiw7AmDN0E9UHQ5QfvyWn7Dk+sZiprK5471IeSrsaFl3qnV4yrAcuIoNTfHnXpWRQRWJQd34ADBEAoIFTZgVfWEqQbd5jPt075I0D9wwcMOU0rDIdiIp6QqeiscJjcPOthnEtl3Y3ingn4sWwt36+ECeLptOsFvqofgkFDxbeT+OIdp2PjAAPM3nFiDGG6GuozTszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4n1CICjCbbiuGetm2xymaPhYNssv63MLq96TzAFC1x5i5dxaDbp2Vj03hChlTji2Gu2HmKCKE/BPTBntN1bNzYuPWCSkFAM9TMQTXed3sDDImHxhc+dXTqxiM8wUQpwb5DEK1c/NqFwX1VE1btmanVCybqyB2CHiuv6XrhfmeA3X3cP5ZN5AaSWzdSohjJB9rr5INvXtVloFDKVlNsvs4vL2mUs0UlDlgxWLQsITyTP5kbtX2z98LhL7IaTeXhWpefPyRVuqytdA63QCTrIC25qXNtrVNCnl5t92wu2c01Z6K4ZlGPF/5hdQRHCopHbfEPLroy0ptZu4VKVAu5vnbhPEVu532uO5Qru9XAxiKYzoAf1e7SlQOf+ngNzhzs0TwppTQ9u+eoDa/t3Yf2ggsByBTcJl2iVGAg7MjAaXZCEuqdXjKsBy4ig1N8edelZFBPp1y+rfEZ3d3FoPOqGN03irrVXRc2VZqgpXV2wSfgIDSYYabu7xJGKTRgLqnIjmFcDo9QTtj0iXrG9jcP2gVqAMWrl5Ko+trBdjfuu8IDU4WwLmjeEz0Wud8GMYHbqliXj8osuw+4E6yOpUTt7kM/9zdYkFakaqXRaZnM6Ss2JEjONdIsoIPE3+MOyANeMl5S86pHL+YmsMMS7cqDzJ+fpBq4zMsxbwkl+TuixUcH/IOcYv/Ow24hj60irs6dV8Zj9Zvujn4iO8E13nw4U8d3jhlLG+PcnvdNwTLgAhoqBI6rH0WS7kFMproxLotcDCbPBx+6moOgekg5BXB6CEFMPenyYD0RxsUqHDqjW2trcq8ybfVhDePIrkuycv8IdZD5U0ZINfqGx7tJOy6WU26tMmsUkqbKdZc10UXEfpIeURJGIUsgVjNPeDDgCkYNHPslg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jot94OfeR7pMKl3WEWDH9tJ883ZrlifN/KDLKuok6GzGYfGFz51dOrGIzzBRCnBvkOKRhXGD2ELcg4wF2/0miy8UwQnDBvzJYAsOLJxOAo3CBvg7u2Sf/w919cEThyPiBX5tQOL2AnKriRyxhNuzLVDLbefYdrf8wsCloPei9kHTJDqtmSk++tOsVZpjfeDwBY+/BvdqzQNWxhMs5jrlDKRXkZR22iyB+1wlQExdB7FNJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRTlVCGsO/wmk2SXvS6NEssSfPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuWW0IuQ3SauD94eCBagfZ+5yms5SJOIKXqjq51kjA5Y+WEhWJPcXkDZsDCfDYN2XmpqZEK+DIEGs1zGXE2e49F6e+ICMF5dUXyX89/EXpUYc8HH7qag6B6SDkFcHoIQUw71N8eRep2zgf7H/Ep8h3jPRHStrKInypf7lSXMlJRCI1A9BLHgxQA+Ak3eVPQwK0EK5P1AnyvSlrC7QfF3VheL1MOkPrJ3P1GJnF+PuQqNA3yv0MIexl2ZwMY2iCL7s4zQh8e7TQ8WhK9uS07MmryDSYh0JTZ+/Ka62LVBty33dsR6j1uE2bnUBGahEZWrcX5G7V9s/fC4S+yGk3l4VqXnfIM87F+zSAna9E8eBsoz/Np4hMWuIhy+yhwcSftaQQ5sGdNE1rfKSF5pPx0l8iv9WThhObKIJL0zO/5Tj0NEJyA9fXCin8WHCbPDLM7IriqpvuFU5wHJanJB9KLBPjFyIOtPipp4oDbb8HImJA9sr7fSSQe6qqv20ElGl8chR62BFQb3LKlxfS5FBKBSEuMCDKSGs2r/6qjGOPO3xcDCyaQPs8Kei7qnp1852+7bsPfRNagwJR0AOJ4U2ZNP7Oe7dw/lk3kBpJbN1KiGMkH2uWUKkV7zn548cSGnesJAg32+fizIeUBAKaglIl3h5OJzG3rgn8nixsN/JeBWbsUy5LfLnxh2umcrn8ZFNH6dlX+jQ3WqYi8Ggp3eab80kXZ+bBnTRNa3ykheaT8dJfIr/jo2FWjCMHw2tHwiHUoOwCXgW2pGrM9xvYf1Qfe3GoMF6J3dJRqu/ibDChjqEQLBY6sgORUhmdFpBKX95HK7PIw7JdJ5XaG2+K+4ObZKNCyxYNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S1m+D6wPKM2nVK3SeXLAYZsmpHp+NXhstilJY8m/oWP3EHyx7QmYpj3LkQ8DSSJPuRcxN2JM5//dWs9A0mxr9hmc+5TDJIuK7XOZeM5Ea3dZrugR6LxsSz6yB5oQKZn+ppWuutn4/WUJxaolkPK2ikfH2l04QjIcCwfQMveUhW9eZxLZd2N4p4J+LFsLd+vhAlhwibSDjRefa+EGrWudZ1h/Xk4/6F1sgk28+cQGUDkXGnQvxs3hBKUnHs1G664Ty2YIDkNWTQbBAuvkO9DxaFReUQ1kBOF6y+R0L6CZI58x9EJEvNnNI981+g7Vxkd4SSGJZ/8KVo1n5WYDSStmzESnvVPFeu32xvkHIFMc/V9ReeCIX/JfGeKEuHR2lUpMK4AOkIsEnVWHLuw4TeLj9jd51DEuhwCVt/Kr1uvMmgmZT6Ogz77JeeQQZRpNqjXfEuPgsCMn9iXR2ovxbjs2qLRMH20NvPdLrMf82zJPSQTGcgjBeKADcYUn24I9SzYlbrRIgvCoUTi4UkJiLOsTgTWbGMhKVl89wX2/NRTdmdegsXp0/GEQmJZQhvyvwyW5RbRIgvCoUTi4UkJiLOsTgTX8/OEaq3yMyTb/0KV8FZhaTdkQl8OTv7LC9z1DuA+XNK90aGjLr4y6QjLGpkU+hRBmB7Io+Ga8r+wPnCcpkePOFFi4uiqvWsSpX7Yb5Kgbwf+o2QDTjcnM87Ci8imItM9hAJ12c8YNGkZhEE4A8dKjsdM/NP26UoqfyTH2W2NcsXond0lGq7+JsMKGOoRAsFiEGcX2gq7Mez9OtMrOKp5zKmYBT70VEW5ujSTxxPHw5Fg2yy/rcwur3pPMAULXHmJucKI955VRowBUTVNAVLdLhHYWnHNFpsZecv8LOPaA1mkiHDBvrrp/Zygffskv/cB6J3dJRqu/ibDChjqEQLBYjyksNHlQ+PGogXfuHoYAQjuY4s21auEmk0hci7ClyFGp1eMqwHLiKDU3x516VkUEqnP6segBb2b0gy9g+wwi9nrAV3ZdSzIYUMlhap+bLnbTszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4mNyWlaZHKJWvbt57rdN55OfPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuUDvoJjJcADsHFVVX2WWVfHRjUQU9yjRsCzaFkv37+aw1WeqPYPT7ZsCUWSJeUEBxThgtJKpGb4ZOCN/jdy1H1OHqm6t4v+zZAKrBvfWmXvFv5zwBS3jzC3aa3lt0kNsaLzxghVHSijjyHe6XLOjDgG6X7xJi0YDJrvDHRSU35bI1WeqPYPT7ZsCUWSJeUEBxVqfvMBRaZDzlGguyXh9bdPHmriduh7L6arPJPotKpEkkp2KexA6RGrMPnpo33AdvaMOTzvJVVUBCtV8F85kQeMcozw5akaR4x2RJdlcIBxLWEhWJPcXkDZsDCfDYN2XmgifEps9esGqvBJ3zlvcBST3Q21YNYY/n/Vyowj5SXCeWDbLL+tzC6vek8wBQtceYm5woj3nlVGjAFRNU0BUt0t1z9hno26uo0UcyH08Kpymn0ZsSmOcV/lMtIafDpl8a09rxtGx3D3hM4T8qnRtqWu8HvDDemtcbjbTjRYSb6HRs/Eo7bpxt6GfpUNtcMQdnh7UE+/TRr7m/hKciQeSLeH7FL8zM6sBzBWW1hqV/VEgXXyuN6vs35PA4gS1mNilcsktTNhUIupQLLl4P9FMxgOnV6UhMjJmSqoBnDR6YX60EmHZ+mxMXte+VaMbfRii9FhIViT3F5A2bAwnw2Ddl5rAN45+qzJcAJ9nTVUHAFaiIkw1XQaPmpmqseUTDVptIlg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jc/aK0yc+tIytQYc8vfD0GV/ZGJvJc+OWfBHE0MJiZBI0LC/nLol5nBY26xVrN74iOPPcA93P3+k5uqpUqRNrqnUcD9HvpVjRYapqLI0DTqphyUyXsWI70q3Ofl35B2BnMkFRgNEo1VUOIW2MrjGtlIBfswGSPH2OmUwwIcX0DldVy0HBuNL2ic1rsY9VOA7UVU5NyyPRNLn0KED6Zd8syexqvmj3QQJ6O3RaUH2pvKTwcfupqDoHpIOQVweghBTDR6hNi2BOT3YrpV+76yFu/2ai1JYN5b61edu+2Dp8jRRKwyHYiKekKnorHCY3DzrYZxLZd2N4p4J+LFsLd+vhAtcu/P+HVTuF0g1zr5311WKpScJ5zTo1TYXB/eXetCW1Z8D0Duz5hmJEk+rSK2/4g9xii0NZ9mipCtVwCTmh7fn/oe/8D1Q0IUq+YLI8cvCXZBj5Ou76nhfqHn6rbPyEVjuT5HKTqBiTD4Vs8gJs9MC9fGWoXBHl96sjkcyj6KOacSnueLHqtlV9ccolWckY8HDSLsCcrCEQqDis+Hszc9bNIlXw4aibe/LOSslOU2rATs8N0haSnu8bEgU69VfwPnGFRGNc0JlTbXVwpqLTAHoqeVpmLF1hFcPIwcseBAXunzzdmuWJ838oMsq6iTobMRB8se0JmKY9y5EPA0kiT7l3Y9dly3wer1ignmXxb04gTA+H13//sBYgzGFUId3IeYRDMPIZqnjUrEGi+zG/jOeqRzfr7oUSVP0JUgKvqPp7VSs1fPRisMHD0V87DAyB1nnuirFR2H5T7hL5eY1KwQ4JhvF5FUngU0hEZOpuqNbRuI/O3mAwmbtNVcHKzz9AKwxdZCFiyBj1I6XAt9BykYc7ko/9QeYh950Asjbw0b2OB2TVycFQT+RJPH4ZeQmH0+XTSNPCefdECOlvN7RD0LeHvJdFl7JxfkHXNihNKJTDI1gACBhoDYE0SOaYbcunrleQICjhzKyO7mR+m9T64Wy3uLUAm1PpgqRZNd3NOazrkECDeb5awR4Lb3p6sNzExYcpGk6hf6cavv7o+6jER5z99w8DiKEVA+OCVDtE0C8O9oZ7kFdOttCLtIM6aEgeAvDxvgFTacO23OzTwNZAkEmsZw0sYLYmurOcd3sXJQELdKGdGHCYal03BkqfOy2LBvBx+6moOgekg5BXB6CEFMOpp/TFl61glswcqHBosNF3jEubhNWB9ay8NLWkesQzWFg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jymbTlVP4MK96fhMk3DPOjkmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagXx27OTpnG7janVcywfPD4pYxzjnfJ9IZSAIYWUgBCbCHe34yCTln/ZWnK8E+81mjcq7rzfiId5MJW0QxBiRCY8efozJigcB8XXVQ9w2pyxRBrVYJiQ9GhWTwbmnwyCGHX3SEBzNUXi3MlfkkHgXNw+Zqo3WQUE2oIFvzwYpyqmVnE8ek1Lo1gV5UbK5E0ovrq6l3uGe//vUWHdERNnZRX3z07zCu8zLfCgsamVeTp98LXq1gnhKfzzaVZbnrMcG6zg17ZLR0Gt/KyRgkJyTe4kJqZ1W5/U9swfFzrOIaJ4j2kOpBL1FjnCiBa+Rwi7JIUc9BnrVO7eSuRV7knwg/DAk31tHhnU8DecdgCHqnmTYKBFa3ajH2NjR+68g6BstPFucdZYiMaRFZFxcAYNSWJ6EImAArnCqRUj9803pO9DIcx4Rz+87Q/KrRXppLGDSg6xrXOGcTJYnAZ/cqgbGaEr2WVzJOt/n6ymQcoeNz4aWhd0muTlf7oCiXALIcHwvOQa1WCYkPRoVk8G5p8Mghh190hAczVF4tzJX5JB4FzcP5aOsamSmhC7CWooksIHrpFdM/Xq3kDR9e31c8NoLnozZaI2MaTzcJfnnZoRpUfby9V5CvTJjL6hDVREv2bbDAT6Ogz77JeeQQZRpNqjXfEpoz2pQ2uu2HF/MVAWQC6sWjMF1/9+t6HscdBKkId1Ex+bw3X+m0oEHHL5pCZwhvbFSNnto8c5dBJvKdBn4f7zqDdrbi+wUgJU3x5CfRhs92yBMHP+kiATnw6OtqFNa9b4+ZP02KzsYLTBoVbHIFquhyuPtZ+0yBJh4aliXBJOVAXj/AUbbw3uLcZQ9I+2Ew1nMYxcoA2477b45QZT8QYDR6BuCjccoMnveoUnsZlEVEkvCBoYoTI0HeNSd+RSq9gK5YV8WnFLi+uQdJryujkqiTUnKVKZVg9P4vNgXf6vc8habbjawGGX2diEVBfoiluTKRzFtHADYFu6ufdNnsoJ2wA43yVc2aL50xKY5tRGZ3NHJzPf++reaGGBJRQ7/HAUgHlhQsH8uAaxnEe3K4kHgqOJsIQt9hdn+UaBqoXoAfyhhr/xqhs8eY+R2IAjvPmOH/9ypo0GVrrkHCm+9iiS+pVKQ5QaalEcoafAoPP4N81+a15ktjrFJLT1YCA9z/C2UQAYmfFuHcKZu1iHU0MInIb340EKmLaU0weEAbZZ+C71024z8TZ4dOkynbSxv9lJ/ZCbV7sSh0OAkYfGZjzm9nE8ek1Lo1gV5UbK5E0ovrq6l3uGe//vUWHdERNnZRX0Rx6U/lWFUkzwuH94Q2jGgV0z9ereQNH17fVzw2guejNlojYxpPNwl+edmhGlR9vAdm2TVApbe5xEnuLbEJvJ4ykcxbRwA2Bburn3TZ7KCdsAON8lXNmi+dMSmObURmdzHxmPg+vC5AFaGyBYc+Ifz2kOpBL1FjnCiBa+Rwi7JIUc9BnrVO7eSuRV7knwg/DEwOQT7XUZ9wHArtxEaCUOz5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOgFWn9oGxTkGReuxNH0wKLzT7LrgsLw6eauA60gofG6BqVSkOUGmpRHKGnwKDz+DfNfmteZLY6xSS09WAgPc/wvpYjtNK57gKIjHAmjnq36+6xrXOGcTJYnAZ/cqgbGaEr2WVzJOt/n6ymQcoeNz4aUUnOrtm6dI7RO1/mrCSza9cxjFygDbjvtvjlBlPxBgNHoG4KNxygye96hSexmURUTp5AmphQ0Lfat/xdhkmq2+ZxPHpNS6NYFeVGyuRNKL66upd7hnv/71Fh3RETZ2UV9Bb4yX+3A0ln+5xt3r1KdgFdM/Xq3kDR9e31c8NoLnozZaI2MaTzcJfnnZoRpUfbycJX3H3EXeHFTJDkwaLbNLMpHMW0cANgW7q5902eygnbADjfJVzZovnTEpjm1EZnfH+O9iq6fGGYd4nQtLnKJt9pDqQS9RY5wogWvkcIuySFHPQZ61Tu3krkVe5J8IPwyBpfy77k7eBoDI6auMywzn+bw3X+m0oEHHL5pCZwhvbFSNnto8c5dBJvKdBn4f7zr3YcePUV7fkQC892YJnrd8GRkvzNILH25cD9rjJyDsrKlUpDlBpqURyhp8Cg8/g3zX5rXmS2OsUktPVgID3P8LzzX9yXig3F2T1/wFEDhr4usa1zhnEyWJwGf3KoGxmhK9llcyTrf5+spkHKHjc+GlbuWMlzbgpXSaysAbzJN2xHMYxcoA2477b45QZT8QYDR6BuCjccoMnveoUnsZlEVEnGXvtB7UM0R3INkILFyymmcTx6TUujWBXlRsrkTSi+urqXe4Z7/+9RYd0RE2dlFflHS0Crf0du2RVaQyJ8qS6hXTP16t5A0fXt9XPDaC56M2WiNjGk83CX552aEaVH28+NUN+3zlnXZqYz7DFf0CFzKRzFtHADYFu6ufdNnsoJ2wA43yVc2aL50xKY5tRGZ3XPH47TGAgG9+lOXSonqA9PaQ6kEvUWOcKIFr5HCLskhRz0GetU7t5K5FXuSfCD8Md3ffjllKKuYk+Nzz8bJlrPm8N1/ptKBBxy+aQmcIb2xUjZ7aPHOXQSbynQZ+H+8692HHj1Fe35EAvPdmCZ63fAY8szjpGDwxU18QkYA8SUepVKQ5QaalEcoafAoPP4N8urbx6yLQIHES4PJp9tWhLP8e6gUr8orv8sB9yyu3Q2TrGtc4ZxMlicBn9yqBsZoSK5dT6sUtGnqn7iKHTXBfjonVLAIsNB15FWGjSBTltclBrVYJiQ9GhWTwbmnwyCGHFPkJwgawqdE9nykgZNwlc0jdkBO+d9RN6hAgbL5KltQV0z9ereQNH17fVzw2guejhSJAFKIz30akzCLuYNbNhEYMPIuyNPoEaNGUiBVLyChPo6DPvsl55BBlGk2qNd8SmjPalDa67YcX8xUBZALqxVXCJ1EgXPomshGwnQC7t7f5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOnn3Q4in6vAuf9kuuW2E4XpMgbOQYPL7SZxz2lC6Zp8cj5k/TYrOxgtMGhVscgWq6P2dObxPEL20EWLHryOU975+IvTwEFA8jUoCw+hcASUacxjFygDbjvtvjlBlPxBgNHoG4KNxygye96hSexmURUSd2UoNXpqLayt6ory/ew/IrlhXxacUuL65B0mvK6OSqJNScpUplWD0/i82Bd/q9zy4E3d2l9pktoblj04wrIJDMpHMW0cANgW7q5902eygnbADjfJVzZovnTEpjm1EZnfQnw3bWU3XoI1pyZzlKbkUSAeWFCwfy4BrGcR7criQeCo4mwhC32F2f5RoGqhegB+DPJEyM8uPJsHEGe9EQvVWBjyzOOkYPDFTXxCRgDxJR6lUpDlBpqURyhp8Cg8/g3zyv7EK//47S3w3D+9mbZP8s+grRdRFjHc7h802OKRtLchvfjQQqYtpTTB4QBtln4KX9PmnQEqrsCdsswOa83oQFyZyENehc5r3+3gxI23GAWcTx6TUujWBXlRsrkTSi+urqXe4Z7/+9RYd0RE2dlFfcqor6rSTLXNDnMA2/uFzwQterWCeEp/PNpVluesxwbpXhi/iEvZlKlUUjLa9ohaRRU5xTlnjCGx/VyBBZaGmgfaQ6kEvUWOcKIFr5HCLskhRz0GetU7t5K5FXuSfCD8MyejWe3o8iCHS2ShomW+KUQoEVrdqMfY2NH7ryDoGy08W5x1liIxpEVkXFwBg1JYnHO360rvM0v5bW58HpdjD2nStnNtbzknZ+PwbMQEAKfTrGtc4ZxMlicBn9yqBsZoSvZZXMk63+frKZByh43Phpci9zC95s7VgN7IlfLwBLS1BrVYJiQ9GhWTwbmnwyCGHX3SEBzNUXi3MlfkkHgXNw9RNjAjBUlZgBZW7yd+yFz0V0z9ereQNH17fVzw2guejNlojYxpPNwl+edmhGlR9vJPn/bueNJUYnqN2lszIHGJPo6DPvsl55BBlGk2qNd8SmjPalDa67YcX8xUBZALqxSXYNDAyKvIfYAYHqzyu3Mv5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOtvwKnl+gIxsD+ihdPz4wvx5z1387MMjWPBh/R9HhvfXj5k/TYrOxgtMGhVscgWq6HK4+1n7TIEmHhqWJcEk5UANweoWeH4ZhKOCFduub+abcxjFygDbjvtvjlBlPxBgNHoG4KNxygye96hSexmURUSmC7eQfqciAveqWKDOFnS4rlhXxacUuL65B0mvK6OSqJNScpUplWD0/i82Bd/q9zzfWjM+M3KH8JwQVWeF8tP1MpHMW0cANgW7q5902eygnbADjfJVzZovnTEpjm1EZneqFokwJBfiskgpoNUPXPIJSAeWFCwfy4BrGcR7criQeCo4mwhC32F2f5RoGqhegB/MUX9UN9TkymBj/9jKgb3bBjyzOOkYPDFTXxCRgDxJR6lUpDlBpqURyhp8Cg8/g3zyv7EK//47S3w3D+9mbZP8DWIMSYECByDSNAShpgSPm8hvfjQQqYtpTTB4QBtln4LvXTbjPxNnh06TKdtLG/2UQ2iL47Gk0vbm+7HQaYXTJGcTx6TUujWBXlRsrkTSi+urqXe4Z7/+9RYd0RE2dlFfiR5rKb/G6KhImwzGjHos9wterWCeEp/PNpVluesxwbpXhi/iEvZlKlUUjLa9ohaRCMq9pdj9+WOAyxwRwlBA3vaQ6kEvUWOcKIFr5HCLskhRz0GetU7t5K5FXuSfCD8MFdBlWVRJdreS2j7yGXdxPwoEVrdqMfY2NH7ryDoGy08W5x1liIxpEVkXFwBg1JYnSMX0/hxj77jIvp2tvIHMrfdTQks+cytcS3apzymhShnrGtc4ZxMlicBn9yqBsZoS13csJ9VWe+YryIzwr5Rs5ilnjM+Rdy5QPZMjzureA+pBrVYJiQ9GhWTwbmnwyCGHx5vfdnyiOmUVmd12k2b6eay43aiT9VkXHbGg8ggFGIIV0z9ereQNH17fVzw2guejNlojYxpPNwl+edmhGlR9vH1GESZwKYyM5yeKybvmtKlPo6DPvsl55BBlGk2qNd8SmjPalDa67YcX8xUBZALqxbJYuL+lzgeq6lZ9KynfTiT5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOgesqk+5OgfQC+sG6XrSfPKTcTlTaxJx0lLekiqFKGe1j5k/TYrOxgtMGhVscgWq6HK4+1n7TIEmHhqWJcEk5UC7slRumHuGN7Nab2exiG/ecxjFygDbjvtvjlBlPxBgNHoG4KNxygye96hSexmURUTBNylVdyztawE9mrazuIGbrlhXxacUuL65B0mvK6OSqJNScpUplWD0/i82Bd/q9zwzMOKKnBDbTA2+m8O1vxiAMpHMW0cANgW7q5902eygnbADjfJVzZovnTEpjm1EZnfmIONQ4ImA4n8XcT8aIQBoSAeWFCwfy4BrGcR7criQeCo4mwhC32F2f5RoGqhegB/29UWBZnDBM7XJrdn+tcP+skQCa9FTNrVLwkgw5FNdk6lUpDlBpqURyhp8Cg8/g3zyv7EK//47S3w3D+9mbZP8FjZJBXgrI7RqSSfN4ysBT8hvfjQQqYtpTTB4QBtln4LvXTbjPxNnh06TKdtLG/2UmO8YBK3wD6MLP5Ju4g8X/WcTx6TUujWBXlRsrkTSi+urqXe4Z7/+9RYd0RE2dlFfsqocv9a8qP5PLPLPfLCJFwterWCeEp/PNpVluesxwbpXhi/iEvZlKlUUjLa9ohaRJwQ/V69MYcZQjiToI9eD6vaQ6kEvUWOcKIFr5HCLskhRz0GetU7t5K5FXuSfCD8MiMMc04838mOWz/7rx0k4Wk09/lq5YrOEvc42QxTCPcIW5x1liIxpEVkXFwBg1JYnNwxe+UEbd5urgm4+n7ciilDw68QLbMUOsFYUpNCalcPrGtc4ZxMlicBn9yqBsZoSlxiIPODAvjnfQsukyHBqENUEW/kqIerdzClqKnBRuEbIb340EKmLaU0weEAbZZ+C71024z8TZ4dOkynbSxv9lJxQfNs6KJGD6to0P2588AhzGMXKANuO+2+OUGU/EGA0egbgo3HKDJ73qFJ7GZRFRHnu+Vy8HJl9bciK/9UZDO9BrVYJiQ9GhWTwbmnwyCGH7fZcjhy4asDS5+kPFlWEFfsQ8LzXyq+lxE2OhA20/a9nE8ek1Lo1gV5UbK5E0ovrMO3OodMJivsqK3eY30CAKS47m42E9xdIkaO/3uxD9lCuWFfFpxS4vrkHSa8ro5Kok1JylSmVYPT+LzYF3+r3PL4DhLzx4zQgG49yUykAmlEV0z9ereQNH17fVzw2guejNlojYxpPNwl+edmhGlR9vCzB6vyph8KnoA8eJO0Ati8LXq1gnhKfzzaVZbnrMcG6V4Yv4hL2ZSpVFIy2vaIWkXUGUWFmS4vxTAgm7GQbHZYykcxbRwA2Bburn3TZ7KCdsAON8lXNmi+dMSmObURmdyrEkOX/KZL2fPBsX0tzBvNPo6DPvsl55BBlGk2qNd8SmjPalDa67YcX8xUBZALqxaoMhYD0TP1xEBvCbFw1QBP2kOpBL1FjnCiBa+Rwi7JIUc9BnrVO7eSuRV7knwg/DAhRn7JjHelbYS2bfV22Bv1IB5YULB/LgGsZxHtyuJB4KjibCELfYXZ/lGgaqF6AH5U+djUIZdxGcMJncEFPDIX5vDdf6bSgQccvmkJnCG9sVI2e2jxzl0Em8p0Gfh/vOqTAugIZbTei1LuqcuKDFzcKBFa3ajH2NjR+68g6BstPFucdZYiMaRFZFxcAYNSWJ9L48HeXTF9n3xXeDBXDd1LT7LrgsLw6eauA60gofG6BqVSkOUGmpRHKGnwKDz+DfN7fNnZqWz1OsEQ0AwnsfmYy0HTuJW7hUP+GeVhWGgx6j5k/TYrOxgtMGhVscgWq6BemUkAU6CyPEhibPb/N/3GAjdHtANKRh35+3PPKjJU56xrXOGcTJYnAZ/cqgbGaEpcYiDzgwL4530LLpMhwahDEnVfEfv9KPpLtEyDeTuspyG9+NBCpi2lNMHhAG2WfgrOXiUoRHFf7ktqxBzIR831m5eOWSDxSLLCcGTX/FUlFcxjFygDbjvtvjlBlPxBgNH+tXxeroav0GWSLVEGtpJGv0D7O0qMaHRmXddaRWkgjQa1WCYkPRoVk8G5p8Mghh190hAczVF4tzJX5JB4FzcPfG1SDtNvh2mL/WcpEiHdBZxPHpNS6NYFeVGyuRNKL6zDtzqHTCYr7Kit3mN9AgClFYEfskGZ1gOT1mwBiSwK7rlhXxacUuL65B0mvK6OSqJNScpUplWD0/i82Bd/q9zzE9xRMKT8Q3Q28A9uy8zFqFdM/Xq3kDR9e31c8NoLnoyU418umdjEMOAbMiysWjDhPNZQhvIBt1ID2I92q5T7NC16tYJ4Sn882lWW56zHBuv7X5IvqAJNtvv0uw3Sn3H/kRPykZUJPevOkYGDXpNQSMpHMW0cANgW7q5902eygnU/F4vX0dd9NP6Yg4CklMlVI2SsMaIrKTZbGaIAqxjmUT6Ogz77JeeQQZRpNqjXfEpoz2pQ2uu2HF/MVAWQC6sX3eoIjbbCwaEsPWH66rATfJSVFAyGSYh6oqzA3K5iKZ0K5P1AnyvSlrC7QfF3VheLS/vwqOpreYlwKyfKmPpikBr4eHWFT1Hk7oUH5DnsdiZG7V9s/fC4S+yGk3l4VqXn2HRJpa0H2ADps87QTlE8VoM/3ndYTPTdeC/JZZyKP6+XTSNPCefdECOlvN7RD0LdSjYDEj4GPec60Epazjd5aFrFoZbukH43J2NNU4StOsDWuPlK2VuDpntzDdfXq9G/Z/zvjrD3Ah12dR7HcmQo9gZ6bVtFO1KDIV454fysz2WaMV+iVl4XduHGDzfhc8qX6yj95vAK8LMLeCMXPHWPXQkWtz51jJWP30aoMDhgWDTu/m63UqY6B3XfpLjnranf2visacYzia9A3E1xi7L/im2U3GPzOyv3cPR23pUH7Dd/M2+p3RR7v/MRSu9oaCLolVTchPpH5gpa3KXL2nemftJ2UM4el6N/OoTreTH+WVrqXMlFSk4Ej6r3xCDNfsatUMs/g92sk2ZmkcnetOGcZ0OH2w+70NpRxD+HbILAZd6WbJoE1MVk8U0PEwUb51If0BQJzyEJOCTyTX2lIU3Vtzg17ZLR0Gt/KyRgkJyTe4u5fC3WqWtIN5d93e2z24aihozALVg0ju38kIiNfkufwqb4o13R9j6IN6koF+/XOc2kNh6dblgzMtHnCs5pdp3YeY27I4QdomsXFpOMZ75511Arv3eC6I63ZfmAZrz3GkepDvOn2/WpvzdQU1zFYIDVADvc6JEllJwziE7hTj2EumwZ00TWt8pIXmk/HSXyK/+25441/WvZ6uA9f4Li2gMc+BlNVvTXpZgYouANE5SFymwZ00TWt8pIXmk/HSXyK/1KNgMSPgY95zrQSlrON3lod/0V4mHDuBpCusZMiquAltryVfpv758siYldR9YrBNjzVOk/UPshWvKYs/pt9Hb+J8tVsw34GtsW45BZ2O1gyVP4W5/koIN7gtmUSUfA8S0EqNBtvwfyBajkv8Yios/tC/6FFoLHOXGqfoGJ68CzLeomDFm2/09tlqdWkaFgHyWr0G3evZd9pPdEOIp/oaIgfWODUlDzge01l2T5G7+QErcsD6anXvlpM66wAzEahJkXbJb64db9GI5+xBzGrPqLwU05I/aA+JYfwIqmyFm6phiaMRV4T7hoaAIf6mhD2YN6VlK3NU8hB3W8P9WkyIeLH1UBt5exZCKkTnmiV6T7SkzQV2uNS4UsCaMT+k3ZghUm+yW4VNPOt5ZfZx02x/6qUEurYI7/QOdqg9+7Ms4EGYUhoKm3mzeGgKKg2N88LkkqLbpqH6pgqBDgHeSy4sSG7AipzMWzDtn0F5+DPabMUiWUzEe9hXm+c1lVyDT6EQlOVT9hXHeMuqONLyxK90Z2VKY70jBY60cPRCI2llloeMe1z655emHVrCYu5okxH8QAKOdy+HIgaeKDV5BXG09exCd1Fvb7IH2UmL8eMSgraMrrmXVQg/mVplI8dn+cXCyHcVXK4Ti+0UTHTrtCFJeyS6qteMRdIZbHetY6NdSR0/3bbFWedIxYi+Q1LK2oC/TFk6iOvUnbGO1zJ3KlBUjZmC176XGhmjevquQWjq9guawJ9oMzqilrgtZszcgGcZYv613UiTPrV5tgkSnU+GGjDCVPqpgX+199RbNLFow3Ssmsg+Wsm+tmUhFr1ZyyvFWWeL+9ToMHV7urNYkvvS5vMKv7WrUPfZOZ1xCxTwG4dASyxArU+gDhpZKKGWTZb/jakP41Uy8toqtAbX68/oZUkH2dDW4K2swrX/AvuxcWM9Gpc78yjTBWEkcxH72oC0UcFzndYuL6RfZ796OLQzZm4bu5pN3qlBreDEl73YYHLVnULm21QTTBJZ+Pl9S4evizyVWsR96rt9ly1sXEqL+pVLb9RZH3M9UHJ54mVrXmM0zer3VuWFyMsovJnW19mRZ883ZrlifN/KDLKuok6GzEQfLHtCZimPcuRDwNJIk+5stRoceugfp7R0VoH3vKKprwOXY2sRAJf8VsmKDujFzmRu1fbP3wuEvshpN5eFal5VcKHTZFdZ6b6hQbB46cBpM/8/oqt94BRJdaRgAZ2GR+bBnTRNa3ykheaT8dJfIr/S5vf22uhXC4r6F2dVbRYYoc2gyTREbNb+DtDzr66K3kfs9cFrE5tiXW4CX+9j3gi6fDZzVJdEXGg5WHb94Q9zX4rsA6l/QSkeDqEWYVMboH3LnSSN6WM8+QD3bekoeguIdXNtZOZbbFRacJwOBdqvXlNnsDPuLo6wCzOrgN1EyYgBKxPn1xd+7I+tT13XYE6AvNMZatj4cQHxGUptzFn9Vn351fiCwZeX4RKJTpu0ia4H9UQG4musseEi8Nw+RG48UXwhUpsJCVLWjrT5GGCvj6ZSB5wfjgSoC0RE+VR2fTU1uHKaTLWGivC1P0kNtsSgPqXiExsKMqLNid9wUyNjwdK88moT+kGR4pb6af8sWFbsKzOEmN4EjFLtCdYIZS22GHcliD5y99gvNTbGLp568QohtGw1kcwgcHlfq94d4VCQZsLoU8DzLmjQDywFjPPSySN9c/huv7CH3XKhQO/0WpVCz77+h0Xy9o00cPOXtztoo/0394UO57Nf9RAQTO6FYWnHi88KqPjmVMthg5TQIkdana6AkkEKwAO8jKjQ5UHMGLFRHLGPVe8PdaK7oTyxogP4vGuNHo60//B+b5obbbn9NWIzC2Gh5x9ouTD0XgmiCfeqUkxAOZ0gGCP+LnLRPiErDQLhkvmgM139oAVe8C47A9r5AGnkP5XKNQR+6hZJgDi0YcKyEc3qzRY/l4G5bplDD6vn1FJymozv2SJ/fWgBaE8+oRLIvxmYFoQcAy2x2UApvuPvnxZT9IbThPfIkF0zqUxFQxrPxoeoxf3MWsOeyu1xKZqrN0HZV+LyQkBLLECtT6AOGlkooZZNlv+qWCNALyj81gQ6jqV1lJ7IvM/6q2xtI3AGoJXEcrlHk9D8Zx2xGkIBX0ozGKclNh94HbFq8JR0gGZHgHGtMn6KQBSmUIIhpdTy6gGc+FNTIK+eswzycUbKe16VDTikpw5LYXQFf0B+5mI4kjY6Dn4JaCHIIv4T9ED28bFpXS4ShofiMTHiNNLMHaKzjqYQOHawn26R0ml2xrC5mgRT9j8FzzTqWwswipTP+JeHlzQ1uG/7IXGsmrfnsrNrVWVITNqiqKiEmXKh0CSLrjhNcAfVEUXOYcQcZDdhNc2aosfnbLrHR653LUlLc+ZzO1GcKaadF08KiZ2wvvw77J1enip7OIv2gvEVZbxk/ak958TpHNQloCB0s0jm79QN232gd3QfmokHYTY++u73F0OLKGMJgsdmDqYaFiDTB6lnc9cVaIqDWGqVt5+yl8HaoCBzuZqX72eAvpzTnGM671tJXUvvvKvQSATe9KGL20xr6+z4Pk1PIIj+9P8pI71zC9VRlPFNwshU8IZq3hjq+X5LDfuXg1fTBcBjqZs6p9M4Hf18AuUrh/c/vmxRkTykB/FJ8exKeqnXfjPCmuFIhI5xXSMEku8uXVqWWtF3Rn0Pvc6j/eDR7PZbPa2Lyl2JGwU4lfy929i9z6su5h5kFxd8AFpXV58VLMPGXdDbUEaeGELt750mNS0OQgNukb5SlE5IgSBRGGuqeNajrEfNEsqsyY1rR28HYI9JxZdThS8vHuJ2cKqDJ2B8GFuXXhWnLkp0tOVE6vbkgPu63PpRbqU6OJD+X6W883kP8J5sReqt1IkqP649rcDyCOv11CxX4yIUqNL+JCeUx2dXSi8Y8OslSvGCno5AoImUhadzzJIDI+ARZ/8NsKLZ5IDqQ7In0J98LgByv36oRpq8Jv06Ky2Own2S1Xc4LMI4vKSKZWn0WY/YcogFL7GVusJf61U/hsOB1MWVgFSP6DyYGkwc3oH4gweSK/IufbWbLhpkpCkWksHlKlanMmYRluA3DGPTvVwS5SkbcgfiR24M1+tVfigEYNRlZ+Xw9XQxldmHlr6pPkJN1maHszUYYRRLrZyTn5F9cROZZDVZ15vvH7Iru7BqhoZ9n9jC5UbM4yA9KlBE5J0BNL3G0iAOChLoY8yzvraVgCeSySN9c/huv7CH3XKhQO/0WpVCz77+h0Xy9o00cPOXtxgX4e63PMjs6fXd9+PEFLgjqtEWWwoPeOqKT4vMJ1zDr4f+UCojDqTSOGr3iOs9ZYMPU5iOmtasYNkIcEUBNosnhqscy9pS6dm2+WQyTvXOLRT+nTl+CNV0MHMyCmliePeT2JZVmb9AodHysbW7+8zCbw15OujtzOVlVK6T56j7YpSEIw1dOm+16+RKM4mH2mvCrzKQUi6JS7nDdUFxl120Gc1Vx6m7wiqithFLBZ/uUeUxRK/l4IHGAM3d1qKyUauDIyJXwoEfjl6yMoUo/mIdIOx4Vr19e8lrctdlDITQ+pECDdgFSCWG8B1XkPRBeKfVXBjS6SD+dVYJEfuZbeFDGRORNjoSzRwCtTiatggsJynFDPJ2fIZ7FxHJF9JoCCp1eMqwHLiKDU3x516VkUE+nXL6t8Rnd3cWg86oY3TeNE1wWBu1Xp2AUx7TKEpvQ7RFHZtzFEWzDadCJpDGa0z6hAArcnTcTObmUdDFGuP3+S6Ou2NJ9c/JZm4tTQZXzbwcfupqDoHpIOQVweghBTDWpaLEdvx5hljj8s4dNiZS61dcG+909bkSyTZCfmNVYgnKA7YLDStHAyFZIMLsRgxK3jx1wE4e/oMmzJTqyGIq2flHhPobqME1E1zzjfuAOfKlZEi+pYJEYC4P0dKcUnckbtX2z98LhL7IaTeXhWpeWPO1uo3c+0VP1xFo3tDN0k5jgd8SucR2ypaz6DSfqzy9EJEvNnNI981+g7Vxkd4SRcEANCH8gylv2J3UQw4sbS3EjF4tTjVaDlfryxdO1Hpp7Fk03v8qAOYqJxGkKgqtYwqwfGoB2Yz5Kg6fOtGlD38LcLTmA3fpq7lqdge6sSCWEhWJPcXkDZsDCfDYN2XmlZ6is67BLi4XsQdg8xSvLb55h57UNvXnyq0SKUC3uil07M4GJvDLwt9sMJwnpt95kK5P1AnyvSlrC7QfF3VheKSq2VsOWucUafvAakRg7JEs+uAH8Po4U3OzPQV6sn0Tod7fjIJOWf9lacrwT7zWaNAiEhGV2WVw/09LTDFkae83FqUNdqmiMJt6WD5rJbLD8RAB9NogDz9n97mEJ2YKN5GgthV2pWTgUrcQjQk65r3gxGujlSqydMjXHIRRxhIEv1qckty1R+MT0QkKjKBNiPtWBJojSgwOvP3Av0gK1r/+kTeCTjjtYzmR10nm5dLJlhIViT3F5A2bAwnw2Ddl5pzg9VCM1TGvt0+gzvOdD1FFzmZqG+jMbU0kPJB2RRjadOzOBibwy8LfbDCcJ6bfebBs7H8l2qx4v+iaFk/u6RrTjL2IGmY4R+6NHYTCUz08Saken41eGy2KUljyb+hY/cQfLHtCZimPcuRDwNJIk+5e2/+4QxYHpX4kPpqsq1zqFSbemWVdGlug7/NHM6cvmWbBnTRNa3ykheaT8dJfIr/4k4ugV7R7JJzSQpS10Cl07Uu+0sIxLpjVkWYmGAjMGyqb7hVOcByWpyQfSiwT4xcj+ywLneidC5xmv6hKj6UAeQKty0zYEVVwLM21pmHwhh6J3dJRqu/ibDChjqEQLBYpJzyK1t36B8H6sWL1ntAjl7cYvRdyXT/Vjx9oAfCXqg1LVkbEUv2W68Xl6f4D2i3KOq+gvHIKnMWIZGeAQt8KRGrHq+S6XUdJ9VGQvk6FQ62vJV+m/vnyyJiV1H1isE2lwZbUW47xCPaO6gKXp4ijJQU5cQUJ7ARDGjXGRXpDhN876BU8XAWTT/hSWM41mqnV9Zu71qrWAuQkScseIOjcmWrrbAmDnNLt/cveAkfGJAooKWAOSaqMnzEvMs3+jNrlTRkg1+obHu0k7LpZTbq01p6irDRVMMR1i3wxg7P0IFP2IfOe7VBN8UY5zsumw5H07M4GJvDLwt9sMJwnpt95kK5P1AnyvSlrC7QfF3VheJfQTP0yTOXnAD4L+g8sDEWnzzdmuWJ838oMsq6iTobMRB8se0JmKY9y5EPA0kiT7kPv9WCTV+LGZEmvzm9DaDBHFEEa6EfyQL+xkGnNShnftVnqj2D0+2bAlFkiXlBAcVq7+exMi3TTgxUw4uFkxPrZGVTm+wpZr7eUJQYFUxX0apvuFU5wHJanJB9KLBPjFww7c6h0wmK+yord5jfQIApm3/kUkxJKYa5qVcusJ0wzl8jh0OVkGnj4JIw8nxkWE4BrcwTRsHzQGX5PaZPNH1w2ivJbA9dwmIj0MiYYL1O0/RNagwJR0AOJ4U2ZNP7Oe7dw/lk3kBpJbN1KiGMkH2usaJT+JVF6SnSdzB3F5SrVHBjxc0+raOxcM13klqhDRE25ZVqKDI7SPOnVqlhcIk8eid3SUarv4mwwoY6hECwWH/fJaxi6XRHeAWye5xQGxRjqd3uq6FcUyZV2dOja/1Gk2adjxiNSDIajgDvGm50RVhIViT3F5A2bAwnw2Ddl5olhm9lb3ngKxBAq9YLJUcPIkrdL0rTE8xJJifVfjkJYGM+G0lqXynU1exk3pjdXulFbDdHYoBdr5HjrS7szjLySF1qnTZBBDrEISriJyTEqw+FdimGs+1p6rrNBhFHqfUy515FgSCJGQyCTQrZUHcikbtX2z98LhL7IaTeXhWpeagjW5Bbj+PSfLPxJdNRK6E7s695JPMd9V+JT4wUzMANMudeRYEgiRkMgk0K2VB3IpG7V9s/fC4S+yGk3l4VqXnI4QAIpplbL2QH0SgfcGMj/bRjmDSgZDi+dpRZIVRz0kmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagZx+J2N0vAXahyPsgraW92LaPq5NmzYnCIidw9t57TjuxenT8YRCYllCG/K/DJblFtEiC8KhROLhSQmIs6xOBNTYlmat6DG58xgA7tSf40VXK+mpjjKWPm7E9Znh4BnRH8HH7qag6B6SDkFcHoIQUw2qQGwZdGOAMeToIISMcIxNYvFDHJC/rDWpjdeSdhBPmG5FkDYnn4hA0avNmWNMK7li8IorGoZamvIIk99+D+7cGVoA9WN/1PPHkGzW8cMiwOTvznBJxZBBg+fbgR+ELKDKRzFtHADYFu6ufdNnsoJ1WY8mDhjMcz7LaD4DW9zHhSoZj4+mCFKtKxgujH2MlEXIIwXigA3GFJ9uCPUs2JW5nEtl3Y3ingn4sWwt36+ECJ5mT3yOyhI2TT+2awcFp/AAZ/IW4vvtah6zg2/Gcmy9AQTtKEUShs0dw3TFOyNTIf2849CFk7BysSg4z/xJTGuWBqs0alJZ1GasAJki7FMEfeLd7jjenbuLqI8IqMounqGIja4sn/7HkSf1k/qm4jTll0VGeFmkXoxivmlTOP0iQp/rvXHxpU7e0Esp1JaPnW5Wvy0ICQ76PIWwxVsWzFByZo+QEBJIQKNW7Y4vQGz2SnYp7EDpEasw+emjfcB29CoxUpVZHUqjgnB7z8Q8rBcYbCIqzGs34M1Y4ItvznKziZAHGBEbsFyjpC0g6C0aE0R8Br5zKlLXJE0PjEXWCjkBRaS0eQwr1gQe4IaHxVQfwcfupqDoHpIOQVweghBTDR6hNi2BOT3YrpV+76yFu/664F1rtrrMxkDoTtZaRBXxKwyHYiKekKnorHCY3DzrYZxLZd2N4p4J+LFsLd+vhAvTWwETrzZCqzgCnNLwrVW4POONDPdUXQoLPMr6kLtHfp7Fk03v8qAOYqJxGkKgqtZRkYMZNEtnfwgsgFPJQh6UGsgpdxHhFDlLZndjPu9/uUh1k9mIqEIAN4CgfaJSMQkr8W4xN1ynveeTTiy/HMGVukTsbPusoX079+eG2iYhESYYabu7xJGKTRgLqnIjmFcc+R25Zp/g9LhruecfBcEUgQBqPoT2CzWDoUfr0HqKZ7KgNj7b4A5/vQwID1FRkKtRhIa4gUDTk2Qovv/2nFqY3lDvYL+pPPXbXPSZy1TL7x7Qn8yUxRgrjHEm3z2EOl3ond0lGq7+JsMKGOoRAsFhZ+2Rz4rP80qFcggsGv5fgynAE/uusQkjqdabezUPtYqnV4yrAcuIoNTfHnXpWRQT6dcvq3xGd3dxaDzqhjdN4eBsiDFQezfHtRtb5IZOFwUmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFag9KN0FFyjj+AjsU6ITlMEOxlCuzOXSiyXOownEQ50dvqHe34yCTln/ZWnK8E+81mjqeGBad7I1L4+3iKVazm71A5sZOypplINlK4A0KYNesXVZ6o9g9PtmwJRZIl5QQHFSo3QO8wPIvCCZWhihj4FHpJ15BZx7LENQn79F4RjfJmSnYp7EDpEasw+emjfcB298IJImq4bu5pL5Z/TwLxGKJeDGDVmhP1Ev6FAJuldU4P5stQWJi23xuESeXXheeOHR0BvtQHLHLE+Cv+vq27ug9q+oD7YH8DrBsBEMS0SQK7TszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4ggjPlQlwudhI1ZYDuzaD8lJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRZNES9upgDcSvtiYSiF3msppxOkYK7hM61jVmW1HRlxjdUrV0LSV5wxwnZvYEnDfsiMQxflCI9z+O1DDv6m7FR87OarTzZfR/P7I6yTrQLLMP/iS2nA2cH6YT81qg61uOSOtfQ4VVv5D8uoc6zQIW7Sp9BUp52dncVv98NMNFiJDb60D48NIAI13bSLYSE8z/VKNgMSPgY95zrQSlrON3lrDFEC6Ob99NPzvqUXNa0NoWEhWJPcXkDZsDCfDYN2XmvTMDltbzKuOiuzRIH93zFCZLi54ktv6UGrOuh2MitWXOQM/w95R72A3qpCueUCXdrr3NcjYARQF7WPElVSn+ClRkzAV69+O9Z+4dT34UTLv3fg8YMLDdk4HwciBbKfkVbv4TuLGzjEBqHOZchnOc6nvc9ndAKk9qrsw18zCpzhjtlWD5UNOA5qW3XQE29MPr7r5xf7/hgAn432rtcOCKrQjnmeA3C3u78rjkwL8l3V+MdcNpJ/U+O6TyOSzV3y/RstJgVi3CNgwedMkHQGCDLeKsASazqVM6qb+wQePw+5WwD4rFXHRWY1vkEbtwivxqxSFSX41SW/QUHwoSYG6ooTPRmpSTBf7ySzQxI+1lkgJ2vgkS7ao98IKHfrv9Qq0uFhIViT3F5A2bAwnw2Ddl5r/ECmO3Lmov5Av/ncV8l6ZwCw4bhb6Rn76XuzX2mQvxFhIViT3F5A2bAwnw2Ddl5pzg9VCM1TGvt0+gzvOdD1F6Noea7kFqjCGstO5cOOiZ+wiU6p4EgDKa+asU8j5Gsym/tX5m0tzwTkPzRYSOMgAhePuKqERI6bj0QY5JxOL70rDIdiIp6QqeiscJjcPOti0SILwqFE4uFJCYizrE4E1npDfrFTGkyn27ZMfD5L4JGrOOx1pcDkNlOYeiLQCAK0vsLJiY8LxnDTUYShSem5VESm3h8pppczau8yfNEmYkfCgtYXfbijyBL619GviYvIvOqRy/mJrDDEu3Kg8yfn6PrSbAXQSwLx6ZRj8SdMT0PLwUyXc/YC2JKZQE79sKKXwcfupqDoHpIOQVweghBTDqaf0xZetYJbMHKhwaLDRd2H6ycdpe0EvUiAObqeRvuaxenT8YRCYllCG/K/DJblFZxLZd2N4p4J+LFsLd+vhAtZM9C3oa8QMc+d0zmN3DWyfPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuXkYcNcO/M+U/XsC8oC4Dc1QRDVESh4iQG1Xk5dSS2T1yQ6rZkpPvrTrFWaY33g8AcwoyTmJrwg770xkgzYMwkj/M6HC2WIjxuDQrVFxpprakp2KexA6RGrMPnpo33AdvdEV4+6bLQHcgwcEyS310sADgFmHyhMzE+HRRLz0Rm5R8HH7qag6B6SDkFcHoIQUw41mwsHwsx3tvS31QHZ82SaH7an8m2vjvzjGo/60WCuASsMh2IinpCp6KxwmNw862N5Tz+uta5LmdZiTLxNOVRLkrnZORkx84XNa8YdHzwYVrBxFDL5ou7O/g+aNyhSoUNVnqj2D0+2bAlFkiXlBAcVZ+kxqN/T1aBmKGewrjg38bYyq0UcTZRBh3ZhJXe2wWckOq2ZKT7606xVmmN94PAHLN3Pbfh1fXyVywgy9RCjCkHnVlkK0LgCRfPD6qxAcGZKdinsQOkRqzD56aN9wHb3Smm/ZJtM4u3tOgdtMtPTWTnxcyKIv00Tjb2xxzmD9rfBx+6moOgekg5BXB6CEFMMceyrdXuXD05fUYPqCeQPJASgDEPNRG0gj/MpGuYMw2anV4yrAcuIoNTfHnXpWRQRBSlSw3IQ3BzNBYyeWmUALBFq6YH+oftnaXd9OJKlLYanV4yrAcuIoNTfHnXpWRQT6dcvq3xGd3dxaDzqhjdN4AhBO4aoWAZ3SvdIDza4Ja0mGGm7u8SRik0YC6pyI5hXHPkduWaf4PS4a7nnHwXBF4O3CX+SsqI7w00czg/bIF8b8wabBHieT1GB7ecQWt2Ldw/lk3kBpJbN1KiGMkH2uR02yYbilSzKCEdlEr3Y3wbWCTeSqUCnB8cDo/o/xPtabBnTRNa3ykheaT8dJfIr/rBYNTUNW9HcOiG272IrHZM8788FG2DYZb69kW3YJuMGqb7hVOcByWpyQfSiwT4xcDsQzcz6pIdNPFXhLQOlWICdENmjI3Z09AGS6NQB+R7ZYSFYk9xeQNmwMJ8Ng3ZeaRD1eh+jPKhxcgOYx8rlggL6y16QxLon2pIjubuEMBirTszgYm8MvC32wwnCem33mwbOx/JdqseL/omhZP7uka7kOGImlBKX6TysVGyK3Z5JMO+4wwuZn8Ptb16ac/qoL79qN+Hro3V0h8WxVTICXCD9AI9vyUwqr8+Fjuo01pfNjOueBl8M4Y9w1+qoWdIhRHLznB1Ib6KcBZugieqHVI1oyLgkuYcoGU92437vly3q6COQ5uwNugGnuqIu3W0+9tryVfpv758siYldR9YrBNsJmtVCrg8DdZxF0lQwEm2/wiyXnvDs1zz6S781zD9kmqdXjKsBy4ig1N8edelZFBPp1y+rfEZ3d3FoPOqGN03imT8DFnwdss/6APot9rfm1SsMh2IinpCp6KxwmNw862LRIgvCoUTi4UkJiLOsTgTU2V97yjXj3ph7JGOvEFhLyTPcjQwkSRSASQGxyHeXzLMk6gAx3WLYVanLJK9XiSa74xlAWavQ+x+lIq5DyJnqXuI/O3mAwmbtNVcHKzz9AKwxdZCFiyBj1I6XAt9BykYfP3YihZJCBGZlYurvAKfFBu9mr0WQZros2T1yncLUPuZKdinsQOkRqzD56aN9wHb0incOGvAemMVhQxgX7zd831G7/T3W1hvtUuyJKMEiPD1hIViT3F5A2bAwnw2Ddl5pUdHc80LvbG26m+desd2cgbIvJYv4cJDeJWYCS6mYOYEmGGm7u8SRik0YC6pyI5hXHPkduWaf4PS4a7nnHwXBFNaendzOFWbqNG/m9fh4fDCtB4x4RD7jIhPfOtSzPA/DVZ6o9g9PtmwJRZIl5QQHFNzsdCqGKMnGfjhzBzP+7GRWS2RUl+aaHKwG+NmKFRGDl00jTwnn3RAjpbze0Q9C3/QOlKLV3v7LKiOEsK1k1GWKZYQKifxHrCc22DPnBaB62vJV+m/vnyyJiV1H1isE2PFMP3Li9irOojQ9yhcBLPMybk9dsaqYPs8KRdnGrMHjwcfupqDoHpIOQVweghBTDV7Shc5m6YOW6CX/CPS4EY9KJ99FEGocpdftNHyQB/LruvGfmssTGwFA2xGXJCQ1BEYUYCYwy+7+8udRaqI3cnXULuIamN6XLpUUEqxbddOGdZQpkdEr1VjYrFWNOCzndLzqkcv5iawwxLtyoPMn5+uA3Q04BRK1FMSPxZYn78JI61zHT6zHGIbBCdvO5hZnC1A9BLHgxQA+Ak3eVPQwK0MGzsfyXarHi/6JoWT+7pGuK2jU38RWpZePSU3SjpjNoGUK7M5dKLJc6jCcRDnR2+pG7V9s/fC4S+yGk3l4VqXntF613SzKK8qW0s7ldJxjWOwlEPXjJUPA5uqs40dyQbNVnqj2D0+2bAlFkiXlBAcVuDrwD1gAW8qoYeZ+4aNkKOrmC6GjmRsFbWfPN9BrNxS1/8j5JzRVSg3RGBijGXSAPyqLHZbUV9WKk71/1vvOsdj5eVaQYP0zKVKVW/TS6a/Bx+6moOgekg5BXB6CEFMNHqE2LYE5PdiulX7vrIW7/Af+RxjJgzcxbY6mWGK7fyHepSMClkvz9H+ME/ZoyMVpTK2iPFNR2pmtfRK0zs9JOEwLCbekpKHm3jYWNaLdwzEmGGm7u8SRik0YC6pyI5hXHPkduWaf4PS4a7nnHwXBFRRy3EUR+/vWf8updE1UTWaaHxD0orFAqakfJ1HoJ0N3VZ6o9g9PtmwJRZIl5QQHFTyeD29w8OduBdQOIKiaF/3DkdB338BBP5XmuG3TX9Mux+7Q1VRjvRLmL5uIM98gjD8qix2W1FfVipO9f9b7zrP5UjhQOUohYskLuXQRW+5Cp1eMqwHLiKDU3x516VkUEF83sZumLzw3NpSquqY5X3cWoHcLF6iSzzDDDmVkztgbUD0EseDFAD4CTd5U9DArQQrk/UCfK9KWsLtB8XdWF4rCkE3lrDVIEUmvbg8h8b6oSiH3BO8fQYBh+Wb04EX0See6KsVHYflPuEvl5jUrBDnROgK83hWUV9ie+5r7NENCfPN2a5YnzfygyyrqJOhsxmHxhc+dXTqxiM8wUQpwb5InHuTi7t9+HmB2FOlFhhC8k5dOWh2iE1pA2R/85Md0+OH5JPKvGoNS9MugWFooofxQ7DJCc9v9Vu1xUjkvinMIvXLbT13t7Bew6Z92OzFOC+kMjgka8Il1Xsr4L26tSQ7uvKelvJSBSmQLUNO6o+GsvA7L7+4hvYs5ErznJMAJ+hGKA5+OP+3HKTkORxzafHFgXcbE86vNx73mxWJEbiRX+HSHXofUJJBFFrjjfiO+jTs8N0haSnu8bEgU69VfwPuxCz4OUoag0fB3+TEwMVZkKpmeIeOY88iGhWOEVNPmdypWRIvqWCRGAuD9HSnFJ3JG7V9s/fC4S+yGk3l4VqXl1BPjh5v7PlKO+kj86NZnDeKlC/SWlmD6h9lukolmG/9Vnqj2D0+2bAlFkiXlBAcXsmwrO8EldhVu1lbmv323v/sgz1D6bygvNmAKo/aGob93D+WTeQGkls3UqIYyQfa7DS0ws0YpQvdkt8EnOUCd/pdLGyy8tZKVLAk5rxiYfNpsGdNE1rfKSF5pPx0l8iv9SAFm+e4mjjbzHF+tqaJuCvWdw7He7dba+q4jsoVMAVl4oPB9Sq/W/VXHHJS5EPSa2SDNn3RXsgL4ZOkBNcRLRZclb/HyhDVn42poq5IZC7Uz3I0MJEkUgEkBsch3l8yzJOoAMd1i2FWpyySvV4kmu/1RZFb8QMyeK/IKmAWgeZvpA2kALGiqKMLWN7kaww07JDqtmSk++tOsVZpjfeDwBomhYiS+OAv7Kz/0KS1hmiGXwjWkJjZUKAAgq4o73dS2nsWTTe/yoA5ionEaQqCq1yaSS/XJYi7sgqAMClVFk8AjO2oiVlTuLmoQFbCGDCBF6J3dJRqu/ibDChjqEQLBYtOsy0hSBNuFYDdJ0AK1SsuSnLNCIFcbYJiJVKnc3d/HLSYFYtwjYMHnTJB0Bggy3YMxjtft00gV10I12+CuiSnHDaKLWGrBwHfv+p9fXMrQypTWCj/I4WjSwYAlVToJU1WeqPYPT7ZsCUWSJeUEBxZK9l8Z3dyqWSnI0fEuh/bOsppAGBr9linzWGp3yikzxP1m+6OfiI7wTXefDhTx3eKt/eQW5c7V7X34Bcdbyg+dxtFvQbh7Pq+5HiPfyl5qeqdXjKsBy4ig1N8edelZFBFE2Qr2uhAupnXbtTyVZT5veM+GWvBvBZzxesgpQIJq+07M4GJvDLwt9sMJwnpt95kK5P1AnyvSlrC7QfF3VheI6ucNAxDT+2/ddy2VQRqeUSYYabu7xJGKTRgLqnIjmFcc+R25Zp/g9LhruecfBcEVDe481EfdDYS7iHEXJ459s6oeRPa4feNmJMClRSVKGqt3D+WTeQGkls3UqIYyQfa5UCuseBdHhosryYpoNw1hpYp45Lj+20GmH+7p8sOvx4y88TCs06T+AvaUcEuFokRv7lklgmrtRzUzLpmYqV46bE8oJFIj03Bt5k1degKP2kZsGdNE1rfKSF5pPx0l8iv/c2d+jT6KZes5CgkCxSIe489hTL7UPd7XNiEyX225ZbjGM3vIg2QmzVBRhJ/6oBFSd32eBH+GIJ9KVRdM8RXiTVBfMHWVUUSTyOYTxJeQtH7a8lX6b++fLImJXUfWKwTabB0NwoiLAQhPkzUfr2B/ZN69x75HZSncNfOmQjl2uWKnV4yrAcuIoNTfHnXpWRQT6dcvq3xGd3dxaDzqhjdN4AbYm5MDIwe9Xa/TXhjHOgtOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXijhBjEsNsiceeCKbp9vsljtOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXi4HeiCavJmjmgMiFTAYxZokrDIdiIp6QqeiscJjcPOthnEtl3Y3ingn4sWwt36+ECpmJDiDbxTjWeC34jVtN1KkmGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagkMwLWCIud1d4w/9XIsKs4szF7fME230+CYywZIoAQwWBvg7u2Sf/w919cEThyPiBw8jk/Od9qjOhIgJUgUSZOJI1deWKt/faUDJbiHgjdMaCTCGSAOwCLyx6A70zfhzpfWS3KQmnM30xFaIcVMasT3RRmjZBBZeaQXRqOyoQSyogk9oAoHGh7xs1qXd/7mAuXXE6/QY3E+efOaq61jN1rLgb1/Ex3hUFRe0GBO1ybUWTmTZNZymg+5wb0UNeG9eDSqmXAD2KaGqLLbzxD0I/ySyS9CgArTA8eLTijGEQYCTTszgYm8MvC32wwnCem33mwbOx/JdqseL/omhZP7uka981mMT/keiMvjPzyxvn/hca06uCvt/z3trbfJ4PqegetEiC8KhROLhSQmIs6xOBNSeoa2A/dRH49qMh2RhdbUFM9yNDCRJFIBJAbHId5fMsyTqADHdYthVqcskr1eJJrkx2eHwuaJ7BPatbDs55qmZywYfLp+IzTBPiH9Ho7ISt60kIuLRs80fGuMDgl3seZqq6Z/QEI0hfb6YW4hjkn+QGn5hEvXGhHb1NQdn+PQmDkd/bL9KHNTeetWHGQG0EGGLDDJg//CEWV7yruck9nM7lpRdEKPBXYi+absil/kQA5dNI08J590QI6W83tEPQt8IHHKjdANRfbMuiYKqUJY79t7DGGg5SptPuCiAj9fkt3yv0MIexl2ZwMY2iCL7s48k6gAx3WLYVanLJK9XiSa7DEOgd2ogIgnpbli+HQcxS3l6XvUj1HnMralHANho3lml1Sd9KQkTCTtxBBA03NsrbLT6ij7OIYI+orY3ADqF6x1IWicjZCKDS/sLdxUafp3gPVDXTGxW6FgpIGUvDgiWGu2Gm/PyqDoDag4OXKbNXaZ1u2+9s3BTMZSdV5y2r3dOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXiJ68S82pJPqFBEaI6TXTDIy5Wnr1DLZSlcbsuypSB8SBujJotWnUJPtbmx0wcjcp+sIquKiGsA6skL3RRqH5vnyppeVqzU/IEd5m7BU9ZVcndw/lk3kBpJbN1KiGMkH2uCK5VO3hxATD4S9P3tMBtVJIU+QwB1dF1ZuB3HujzSPjl00jTwnn3RAjpbze0Q9C3QuIfDluUdkTAjfN/T9uFHLTVkNfRqsxz/kQeQhl+t9LLSYFYtwjYMHnTJB0Bggy3qgTyr/u0hI9yWNKWl2QlqY79C1sr7YonAFhum1jiBG3TszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4j1NSWIZwlu/sBCdbLMYHjd1934EtYcct/eSmi+OIjk5kbtX2z98LhL7IaTeXhWpeVYSoqjVtAUV+gxu5HS+BpR+ErndeiULzLnnvjI/jLGU1WeqPYPT7ZsCUWSJeUEBxTGDHl2i5tRXufCb+DJ3QQ/ozAsr8LtN9T2qX57yzXyN5dNI08J590QI6W83tEPQt54KkxSBlQ45lhMItpHWiGS+NE91s7RUNz3r+5ZYcy1Leid3SUarv4mwwoY6hECwWBvAStuyd7CqHOk1Tl85+SxN72jB0zW0iyoDORdmuPf1WEhWJPcXkDZsDCfDYN2Xmo/d5a+3rKVzNVzuFtYTJUxA54szXVMZna06iTiBDFpYp38WCdeaU4SdBUBtyeGN0gn6cCU1IuH7Jnoy5m2IcpkLgWJi78Y7xPX7n1hNtsm8B7E/nHsF+DMhdMLFd2b6Yod7fjIJOWf9lacrwT7zWaMbTpdfGwh9pxiPtN10tCEJkS/opEcJNgoC0g50iYktknond0lGq7+JsMKGOoRAsFhAGEiR0+aRfTFQg+LHisWWGZZVo1EoVmhXYTetLsDfPp883ZrlifN/KDLKuok6GzEQfLHtCZimPcuRDwNJIk+5XCK6oLNPZ/56FS0WEwfZBbPYtwpaWv4OvJDmGY4bUzNkNfxyOGBHjACRMK8FkToIFDsMkJz2/1W7XFSOS+KcwlfyabADn3WQprF7mZOzQFag1qomUFxgSYZByhLHo3mLsDIMn+JlUH/vv1jd4syqYG6dfPrWk1EQsEaH7ags1H3dFr8++ypMRIfF7D0MXbuHo8BWOKZJ5rNwUQmtINPvwAhI6R7iXjpqSpLsPoQtqfWp1eMqwHLiKDU3x516VkUEQUpUsNyENwczQWMnlplACzTM5VG+WgZOSQXAanYkG8rwcfupqDoHpIOQVweghBTD3p8mA9EcbFKhw6o1tra3Kh4c6f3O0i8Gc2cyP6YSmWHayDkAU6c4cAhf62EbVMncfi2ZWRFZGE+1vf8naeol7SnW7HXZPDCyovcbnUPfRU2fPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuVsLBoHy5Db40h2HWesbjTWChtcRyUsmRrsm8tV3LDoSRWw3R2KAXa+R460u7M4y8kT+3+SEo+ic++gvryE28FvlGsHQXNLpeUrU00IZdqlRyCBcaV7nFZ6sTUGEiSznWLuvKelvJSBSmQLUNO6o+Gu0wLxjt8X2Pf1XRuPMwrQj1WeqPYPT7ZsCUWSJeUEBxftpiEPVd+QfFs+kdJs4ljp4WpkHbIY3OopNmnyIear+fVCDbpURTZp2hfCJwBmsj0BTJEz7u6aXt9xAKLAa6hT11PEdfutKuz7VGp+Vk8jAqm+4VTnAclqckH0osE+MXKupd7hnv/71Fh3RETZ2UV/T8VPt2QgyE3jw0m4VmfMhNS1ZGxFL9luvF5en+A9otyjqvoLxyCpzFiGRngELfCkad9JTSlWWQ+nUy+zDadYteid3SUarv4mwwoY6hECwWHkXfXg2OI+AgfyVKaJTGfWA5bSc96BtzUKd346A/K/Mli9nz0awT7oc7VQnxZb8hvuLP8JQA0/Tne80Iof5IBQhGiw/Jr6VMR4ZQxnldgB+qm+4VTnAclqckH0osE+MXBl6HJiyyXRCwhEkkxSyuhmfMKuqIEhIZrVGFQTP8664qdXjKsBy4ig1N8edelZFBAjr/UGJ26LXHfpa/vin9svhk7h5Mxy8YI2sWjAc2QwWSqzmsyL3e3O4Q5W7xPT/YeuYDvn4IEuV2kB1Ai5fSPPZagaOOCefT4zY2NTOBW5XXAr9s3e8x5cK4/Iewa+Wq8j7apZcwwr/kBCNijrE8QsZjjXx29nfLJ4HMRHZtSP67CJTqngSAMpr5qxTyPkazNAs2QXLCiUiZJ9mKPMOqMq3XcrgsmkqgGkI3xafUIo/M3Ic8Gfj/5JGSVOKOMkpQ8c+R25Zp/g9LhruecfBcEXN6AMZnT0A1buYoA0bfSH2dF47+jQ4oWZZ5ZtoLuTfoz6trw5/YFk9tbaOQ+wnrOnLkZHKRjXf6GNG212q0gP2aTEM5F/2uBTYVc9K3/AcdwterWCeEp/PNpVluesxwbovE4AtaP6snSaCMrnNZpBNjhzXX+WjcyQNnhatgIaBhZ+W7SRqq2BslHDPZaN5IQSNfVMiVjI6ziZ9KvkVdgm4tEiC8KhROLhSQmIs6xOBNQo477EZTgF0y18TH+IZpO27bFmoyRnq8xLomhJTghZ2qfRdI9qzHFzRdMJ3C126gT4hnLvcRl4ln/Xu5puguPcPVxIfFVqnmGmkm850qpIDZBj5Ou76nhfqHn6rbPyEVqlQSPmJYdFD1lk4g2lKH4zt60+g95bqxcN8WhMS/UGk/WpyS3LVH4xPRCQqMoE2I8H1m5bwTVjJhqvK+1NUKcCQSDhXeaU5Bvu+Zigg4qw25d2K0Nh26eAq/8MXG74sdDfs/GAz8Gih3WMA9eq0Xs1PdxTwAoTA7v/0c854NOGbYSly1rcfdlK/5rLn+L1436Q1qhFJYLA+Kp4ZvSXvvE0Ly5CCISuABx4v60suA9+Iv4KINrgOJ8i/YL//LPHNO8kOq2ZKT7606xVmmN94PAFmukgIRsYKCnkBl24yRuRAojrFdzOn0eLGEreQPP2R4j+M6AqwpRS1RH6/BLf8JkCXbFPD/jZn6B1Q83Ottr3oY/pZpuBJ+7Lwcvbg20dPA/Bx+6moOgekg5BXB6CEFMNHWa2hC2Wfb8ljgSuZ858uKwssc64BbiSvAgrFAwBR5S5Wnr1DLZSlcbsuypSB8SArZUmjpsIIYPhaPeWzj8FGEVopaxci00sA1JfhJERb6F2M4SRl90kptVdMU+aoFPaBvg7u2Sf/w919cEThyPiBC6jM+15wxPNV6nOAifvPuSaken41eGy2KUljyb+hY/cQfLHtCZimPcuRDwNJIk+51tq71yMHNklvAJhbNvFybemy9qe3qQ1O8/aFZwExVk6Ru1fbP3wuEvshpN5eFal5mlOmIwhybuw4DIEqsbtnHWb8GKeD3/e4rByThlG39x+P9ja26nZ7Hamg1//BLAG0kagB+55xz1CptLHDxN/+2RaKzrfEg4GB98RYm5r71SVJUU9e+iGPiAIjqNoS1ug+Yd8zLe70mpNaQ01IVEvsVGSjQ0TT+yXup96kVUShnwZYSFYk9xeQNmwMJ8Ng3ZeawHORFqsyvIX/5RRBMpvUr/Y8QGg5bwpXKr+5IKzMOhjwcfupqDoHpIOQVweghBTDB9egKYgG1JKOTGUOg8h8iPRzZ2N9S2xpaN/9OyhMB+rTszgYm8MvC32wwnCem33mQrk/UCfK9KWsLtB8XdWF4vTkOuK10tQLP0yVlY7Ck+ZJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRYtgscMc8xYYlhjX2LcHDXefPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuVmSLiqk8451aPU/U1wHRkn3V6SSpx7CezfuQFCeCo4V3cP5ZN5AaSWzdSohjJB9rtt3BC2PAHKN2afsZxwhN5EhVrPZ0zAbasNn2x6tXPt11WeqPYPT7ZsCUWSJeUEBxTBrz1nzZJw2Ke6iMxhHBXfAtOPWTVW1nFA3fJfQRMSaZDX8cjhgR4wAkTCvBZE6CBQ7DJCc9v9Vu1xUjkvinMLH6FLOsjiOL5GJHk6YCLCjyQ6rZkpPvrTrFWaY33g8Aeqi92PFvRbynwBtHphwIrTTvE47Fo5cDm4wZzIFMdkKeid3SUarv4mwwoY6hECwWF1mKmTerlY9PnpFh8Yk7UdtTbFIp/gMT9igcmc5SY+deid3SUarv4mwwoY6hECwWN7It2dAYLogEK8NNIryAqn1hLbEGZTS699kjp3OwY7ee4kFmVYu3XQWg2ty50n1cfVLxbjScxyNTixTNzf/eNfunkKTxl3+T0EV3x7lLmqTP24KXUWXpPiVD7qVHlXlTj+M6AqwpRS1RH6/BLf8JkBeS+OKuBDlEvrQb3nsmuao/DYOkMc0ZokFemIpFRH7G3ond0lGq7+JsMKGOoRAsFgwBMi4e968lXXQ/yIS5+USaG6mP7llHzRwgHKTz+fVTVg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jQyukoHo3y0bGd55kxQ06FPuyXHJrS1bUlUJyMV1M/K1UfzZJ2QTyt7DGZ4oh50gRT19ujBm6NNW+htUzDOqhUQdaOHYEHCAdG48Ln/vn4RBafCmGM8RP9fXcG39ZQ2ERPq2vDn9gWT21to5D7Ces6VSAEmFoH53mR2IYVjGmHiRDOjms8ujxAWkJY601cALokpTC3wnwc6kJklthm8BkbHL8vZmlWiTw9SVW7tK8QZKMLlX9uz8wyTE32QQl6pQl1bucTMaGkKxzxTPwO7YeXtujZPNLBDaKnJ6kOjAL7L2YIjCV1Sd9jsmEMWARBIzJXaCs5HqiTXZrokceJBL00dOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXivT1sHMRYWOfqPpgjHqL3S0mGGm7u8SRik0YC6pyI5hXA6PUE7Y9Il6xvY3D9oFagIpj/kgCzfoCgrcdsCGCuQuOjp/HkUyB1uBzcaTd34FPd0Vd38OD7WW1qv9nBS5wZaM4yUeVqTzSVVU5OhvDQ1lcLUtjddU8E/+4g6sAR8nxeOsQGyeJlc+77F131pPzPg7lnHf3WwmpYydUj3dUVTTUIMuPOLBuA4rByrdgx5HCZ3CZjN+WjMSUC3j2Y7I7xO0ShjcSPgmO608REPQHzyqTUNp/dTJtob46oQWrd77+SnYp7EDpEasw+emjfcB29qO5Gv9+mx85pkEEd+EjtOhnXB3AhKvzUdJrz2xZOi/6P9ja26nZ7Hamg1//BLAG0Nt4IPXW85SK+K4mFh0pWTURGho/7k1MRifElOFdLHUpJUU9e+iGPiAIjqNoS1ug+ZRyP5cmy/W2Iei9QmPoQ7OIhO/OLf4el+FObmetN+YzkUX2+SrebDNGDMDvhtvbdgnG6jCYRgOZJ2GUDDrOXGy2MjOEPuwN3Sr8DnNpxGW9YNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S8lZG+AvwlS+86rNwVuKPCgmpHp+NXhstilJY8m/oWP3mHxhc+dXTqxiM8wUQpwb5A4bnT7+1VSFF+9D4+46TofbRaqGqTAKhA38t1cQwXWlEPTPityyU4QtdUaNERym54vhf1VQsr/vCPAg5R1jWq9J02ldRw8nZhrXU4heQbFhO/l2A4kxDJorwO6bPVdYm8+rEjH7AyWvnWK+mAbb7bqz+XwRymCo8Bluhgqc9tby/zyuP1toDrc5rGXBD9anXJ5SG/ZDDBghwPqvRl96ToIotD3Et8K6ze8k3M4JG4XBH6q4IpSXFmDGrTf9W/yQAqpHN+vuhRJU/QlSAq+o+nsJfqs352FT+xIeK8lgZdrmk6/S833wkSwCjs/jT3lR8mTarbfWmLPEAc1d8Fkc4hHsIlOqeBIAymvmrFPI+RrMnkm7Yuo0wSuPXRrE6KsFOY9OUFo2yKNCVNPKULc4+bTDYVdG7BGrz43OybrFfbbosSN6LXNIb98Um9B32GFSpqEFRFAcPrWdkYnGwFLPYo363SfFON1Lxrg16Z/v5JxacGLlZv3aJrDz8r/2cZk//voTRRYf++nZZIEedPoQY/zNjhlyVVxUQwRWkBsaS+cf5WST1jVKvvIsFhpLv1lYA+RASGz4WKtqxet/M+7TCulT5EO7CD49uuEw6OMhbFLwWEhWJPcXkDZsDCfDYN2Xmgb1cVa5F34ySpT99gxjS4l5yyogYbNAFyt2IKuEfJoJ7CJTqngSAMpr5qxTyPkazI2B6B52Dlr4tu/3UV7luRBW7b1n6o5Z04QOFnSR8nkP8HH7qag6B6SDkFcHoIQUw21W2fYZGXhedD7FW7NDvfmY6ots7yi24Fp+ZoRYwhk+07M4GJvDLwt9sMJwnpt95kK5P1AnyvSlrC7QfF3VheJTW8WeotvwqqIs0mTDg8aelR38IcDLuAna1MXNpXez1kwSNbGQftLTjigpg+NRtpIQvjM9Gyd306UxJbUbwnvcd5LeCRSnL6eq6FCDcnebEt3D+WTeQGkls3UqIYyQfa73HHWpk8vd/FI0l7UteKb8DQviZjMcmSYPCekDCjephFhIViT3F5A2bAwnw2Ddl5oGMj0yOHxa6LbTlJRWd2VL51sR2lweK+hhV8PAVcodZanV4yrAcuIoNTfHnXpWRQT6dcvq3xGd3dxaDzqhjdN4084IhHqZngIB0JCI81e2M9OzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXimSuk4LIGcMzEEDNsMnq7l4k5P98VzELzLlM94cU2rceRu1fbP3wuEvshpN5eFal5R0CefYaQtaHTHGo3RfApTwOe0wlIhCgtxB9/t1cBCYd86AU1FC+XdYNwHhs40ArcNzp6PUdPLlgCqYObH6SWg/4Y1aCzVQ7ZXKdrIpfTEy2kGVE9nS7/gn/tTBPjWs1bsLFyeyTcUu4vUbt2RwkMt9pfkA6S/6Feg9si7MkEXVWSnYp7EDpEasw+emjfcB29F8sK0CGU3DAxZUmrPiDa1VbiCf/331NaTmHwNmUwyIfLSYFYtwjYMHnTJB0Bggy3klhHn4rdQs0WqqSXJllBjk0lbbkA7Hb8QxLoDc1CnScvU71Y64l1zO/0yznkAJkTpUBSdOaTIDZd+utmKcrynqhCdVAxBCEjqoWFjA91II9KwyHYiKekKnorHCY3DzrYMJxRUGDPCklRWJhwC0EVPuZA4KMBLmIqy4wKv9nS/f6GdyQ4K4OpFFk0ItCGGLw5C16tYJ4Sn882lWW56zHBujC6Z/9QMl1rLCDeORFmoU/ymZCYscO7Gbg+6lCN+6+YWDbLL+tzC6vek8wBQtceYhMkUbJTpM1K8r0UAcTB7SOKgsRmzwH/jokPxtoA3H8LuETqjdOIlJtreNaqx28xyZMs17Ko43qGa4q9LalVSCguEiVVfjvGAj1CJodpFrvYSsMh2IinpCp6KxwmNw862LRIgvCoUTi4UkJiLOsTgTWsqjLA2/A1x1CmxthyeJfJJqR6fjV4bLYpSWPJv6Fj9xB8se0JmKY9y5EPA0kiT7mSHaa/b+YyFTQppUFvASyEUbW2l+3qJM4r17+U6HTrdMkOq2ZKT7606xVmmN94PAHNO+w5A1PA+meg+XHvAI0ie7TLp/z7wvn+67WtYMICWZKdinsQOkRqzD56aN9wHb3YoEPf6Gsv7/B6W8DRIHstiOiTcibJbauaXNlRCVWB2Nvgl0ZmRFMYuuj5as9cPquJpRVBvBDTnD6jtquxNZ+KEhQIzWBzxyYFZcLmvUqipXond0lGq7+JsMKGOoRAsFh58JLkeS5VJ0wEv+4jEXFb+7cZAmepIEtuMd7v0Z8xywterWCeEp/PNpVluesxwboaupfY9KU4pfrJrnz6ftPFN/epRYxieGcZkcbS6vMxnqd/FgnXmlOEnQVAbcnhjdLCpAjskai5yD3GDAQ/2+29aQDUrWXZJoImLotdErutx5883ZrlifN/KDLKuok6GzEQfLHtCZimPcuRDwNJIk+56qzQ4A0xE3aHIZI3PNLDlJUkgvfr4RLdSWwvNZZxScErhmUY8X/mF1BEcKikdt8Q81G/vEhb4YiQmK/3//bg7Q6SzX6DTetGowoHdA+0tAd6J3dJRqu/ibDChjqEQLBYpJzyK1t36B8H6sWL1ntAjgoB97GbeUY4qc5yeXKQXrpYNssv63MLq96TzAFC1x5ibnCiPeeVUaMAVE1TQFS3S55Jng6P8pP7Sl3mejjY4LUmpHp+NXhstilJY8m/oWP3EHyx7QmYpj3LkQ8DSSJPuc8NlH0rNGQmgnA8lSrqwz6ndT3/QksOnDG4NQspg70vgb4O7tkn/8PdfXBE4cj4gbWJ7T+++tmhLieSpt1fUP44E0pJdEZmblb/ggU0FOumO/l2A4kxDJorwO6bPVdYm5fg/VCWCeFNr1Pahe7syhBxRFSukgt3aXwVgaa40RyrvICFw8FdYG9/dCSYagHOgqpHN+vuhRJU/QlSAq+o+ntkE/xX1kiU6Kw3KVPpzJKbtklyEbIcG2kvkg9FSAogsBQ7DJCc9v9Vu1xUjkvinMLQwBFZ+fEqbcuPe+3ZpD5xyMKsgzxROiir2ICLgG7tNMDo9QTtj0iXrG9jcP2gVqBVdq5q/u0C3iXiFSFa3xP6vVeJP7/IvTJlMtn12voQ+9W5jFGQwj+s8nAu2hx4IK7VkVgE7Be3ULNY+t2HWag9DiQjVUb8PlqfUUXXazoI/8kOq2ZKT7606xVmmN94PAERz+8fkRlh9q7JMWP7c4E87tqzupqEh68/dI+UmGRlxKnV4yrAcuIoNTfHnXpWRQSGoxYrXp/tNDZ9ia6WfRcp+AEdVLsgXCKb07ok8NeJUtOzOBibwy8LfbDCcJ6bfebBs7H8l2qx4v+iaFk/u6Rri/aldxj36peKavvgBlfsPTNyHPBn4/+SRklTijjJKUNWqq4V1ZW18ovKxOebhZw1QvQmnZovxO3YEXEu6yDupFg2yy/rcwur3pPMAULXHmJucKI955VRowBUTVNAVLdLhUVZUVzW20mXG0EBUkTeslg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jhQ4i8HABxr8QwsOk4AmyDOZh8vo7YwFFMYnP2RaNkV8b/gnIeuNy8bSk4JdkA11pTKcb9SNY2cyj20LRnPQF3WlKIQNqBi6EH+Naps09zl4/jOgKsKUUtUR+vwS3/CZANL/BhA7gJP0JJcdqoNy7NPQsgBR9ISEIw2vAqDKiLD0rhmUY8X/mF1BEcKikdt8QxbYHV94JWaXiA7xycORXCLpHjLEEd5VhSseeSmsOG3iSnYp7EDpEasw+emjfcB29CoxUpVZHUqjgnB7z8Q8rBXn+cloNdzPduSV5Is6kcuZYSFYk9xeQNmwMJ8Ng3ZeajujXwnpenTZazkGIZhQLxGErpZIx/4l0fKrwClyxFs16J3dJRqu/ibDChjqEQLBYdm4GNGQgXZ8WA4lCsFLw8TDvQKcjjUpNtbBskODXSo5YNssv63MLq96TzAFC1x5iEyRRslOkzUryvRQBxMHtIxwxx/83OmZjwp4Nkyny+bdJhhpu7vEkYpNGAuqciOYVxz5Hblmn+D0uGu55x8FwRWqqHroQk8qZhJYzuHsNPrSfPN2a5YnzfygyyrqJOhsxEHyx7QmYpj3LkQ8DSSJPuS2rCBfcsOhSWT63nPlSURTOtAX2u7TEGHCqO2rX1pgE1WeqPYPT7ZsCUWSJeUEBxX9dkdTC0n9wh+0Bj4qrtsQTfhPConPzywi+clRBfdEwP4zoCrClFLVEfr8Et/wmQM/nW0tj3uv9zPZNNfGKBkVpy7lAWskMt1eB+IF8/u2T5EKCo8fKm4y0XObHAvh198cC7ijQ6CLSIcVsGbSgtvTmW0CjSZ90uhyp5j6f9uSQVUmrhTT9me7ilxb1FbZvdnN3szcKRVRRadbaYtAdkb8A0eqB9K3ClnvaAWpPqhMj5d/sXcsekstbjfmdUcC5Aw/yyPHsszH/L29+DvkGZQkBgDdtsyvtBk4Nh0/h5+yFb4e/Fb5RHJHF8ZQqvigH4jgoSCtjEDTD4cwQ35XcYB+Bmh9IWYjL5ud+mAtTRRP/enEFyHoy0IAe5VyXUMtVqrEjei1zSG/fFJvQd9hhUqZslndSfltGfz0CMRN67oKdRTrtmyb3NFzXqh5hf8ye/N3D+WTeQGkls3UqIYyQfa7RBxuKha71Po4RqPyERaJfG2HWtIGY9+rcL5mNH17hOvt7NrlluU7+Sf3z2vlvC0+9/hunnapG/NrvdrxGIm0U/OXoZKAKSMwAXHBFTnNtxcDo9QTtj0iXrG9jcP2gVqCga3a5YJXauBBOR8A18nxZa/+PRMYE4PPzZpzihUlDLt3D+WTeQGkls3UqIYyQfa6oJi7nieiQRAooWoWM6gUSzxukwtkLYXWGf10mq2uw7RYf3rcJn/lbwenQwvFmQy7A9bhkJadu3RgjdfHwEso2yK21eIoLgzQIY8bGEOD68skOq2ZKT7606xVmmN94PAFg0DbOHa018Z5FhzSR9niikzYonsIVCGozPht9nkBSEJKdinsQOkRqzD56aN9wHb2keVuO6juG3/Dnm6Sily+ZwORLCJGcoTLX1XPjf9U6XErDIdiIp6QqeiscJjcPOthnEtl3Y3ingn4sWwt36+ECwfHfTcVGQn5yyj+4h9ir+TLnXkWBIIkZDIJNCtlQdyL3Tq6+bngg5kDTjA9li+a7vc3zCz2KztYjBWXqFYLtpSkSaC4Cr5zxQxHo/rl49lZYSFYk9xeQNmwMJ8Ng3ZeakR7Z+tT3YJqWn+TI9mLjCdBfoeVvYRhLaYGTZOY5JgCp1eMqwHLiKDU3x516VkUE+nXL6t8Rnd3cWg86oY3TeEbkro41HGZYLjohrSPnJo7DYVdG7BGrz43OybrFfbbosSN6LXNIb98Um9B32GFSpv5ZVZ7AW7LDfAwI77JMVy41ooQDXck9X/jVARta0D+EFh/etwmf+VvB6dDC8WZDLjeJku8yrMTV9y0StZGIuoCPL8UzkOJtj4RK1wLkm/m/mwZ00TWt8pIXmk/HSXyK/xwCAVvulCkNFMK4vdPe3jeYOVcqW/LLbb/3uld3xzhaqm+4VTnAclqckH0osE+MXCL8k8FCWWuUVeHpUM91DmB/yFon0t9Oxow99W63MjaCWEhWJPcXkDZsDCfDYN2XmgUIvYFV/h3CYV48tlD3FPP1vpHGkbpVwFo3rJlDviybWDbLL+tzC6vek8wBQtceYm5woj3nlVGjAFRNU0BUt0s20l82ancDuCkymcZNLnwv3cMsp5znua27QCCtSCOSEyg0hmlilIGAwsJoT8mlO3djTPDyCSChGRMyKIf6YBstcGAdlNJborbP8xLGJJ95LSPalDhqRirewYtxhC/RehszfePLtrKmZtfi7z0n8Va/M2PQj2sLvO6CT6F2x+9P/w8d5KMf0gXk0hoqmn3qwu+IZpqM13W0q7xmKhS4jRVYGdb63LSFqhY4C35CqqJzCMtJgVi3CNgwedMkHQGCDLdV08gcsZhYBJ4ABNyhjKcmoOkMXKmwbIIic99En7Ev2anV4yrAcuIoNTfHnXpWRQSP4/Obr+FNj2EiLn6Njig1E6qJptV6O5sI1zpu9Vynmlg2yy/rcwur3pPMAULXHmLLs6YJLMZFAg22bRTvnw/8aE8KEmZIsDtpw1iwTNXVLFg2yy/rcwur3pPMAULXHmITJFGyU6TNSvK9FAHEwe0jwjEHwOGxiqdVqdV9lIXBAdOzOBibwy8LfbDCcJ6bfeZCuT9QJ8r0pawu0Hxd1YXiTovrSLcdF9DlptTbIVnsWOZrnNY0HkpYBHixkKUj0vQV0z9ereQNH17fVzw2guej810Nlv1sRY1VgLJ6hVDmyms961GofKwi6KpRRGIFrpJYNssv63MLq96TzAFC1x5iEyRRslOkzUryvRQBxMHtIwpbrV9mplaBPAasVhuwRGtKwyHYiKekKnorHCY3DzrYtEiC8KhROLhSQmIs6xOBNfh1FhOLlGwLucMdvJ+wQ5H4znCxt8Ds0uLbyunv+ZK3v5zwBS3jzC3aa3lt0kNsaLzxghVHSijjyHe6XLOjDgEs5MjN1obbuG1Pr3UWkgVAujlE1c9lA3DT2UJnHD2bUJlPodPwEn84xkDjXyGRAme9iqxuPbGRFXpwImyUQxq6qm+4VTnAclqckH0osE+MXLJBJnqq+JcS1cASWuhdMFOxqmOazoYMjxmvZOoXh8e9eid3SUarv4mwwoY6hECwWB3vBajUjIl6f7NLYF35XTzCkeDQP3cHj4Il1bo3KuJLFBPdId2OMR25LIQPpOYDMjza0IXS3JbOw3Xlo2Iyu4qqc956kLXscESRvStGfJnU8seqQe8VMSFHKFnIEYxKFAl0wj1k6/5Id670lY9lVRwT9lJtv7SgRQvs39kfcguE35bqiVIwVe3siKBnFHr7ABee+cRpOjjPl6AWFf9xbVLoo1A3mRj7yqeCOaQiPM64Lc1wX06FkWd/iXNkMTzqyGSKkBFEW2XEhl55/ojywbq1GOjfdUvI/U1diT4oHCtUTqs9r64fNCET3g1zhrr0mEHC5hdtZ2MibvJIhPVlM5JB+KY2XMWQmSo3g3GWOTt+oCg5qMItg2uo+RmEef77ag0Lzw71sZUQlYT46PTdQP8FRutbxjzBPEo2lAZHJwFSLbVVQ7SOQUJ0wFfbwfMyprHcWWXP2PuuddFMxLJ2JN8EYXyeUSnL8yqy/f+OYMf+LyZxn7kNj2rSITAZk9Dp+N5RSSBbljRhU6N9aAktlp1WoBfgTFtIE3Zr4QLsyrD91+kwfvh1IbLoI1nj/LZfB5nrUmUqm/PhFozS4SsW4ot4D1Q10xsVuhYKSBlLw4Il1pswaRI04Z6bdTgnLpQ2otKVzjuHPj/Tx9mOZxCjTfUKGj2WVLfVPcj1sZZ3vL3n1m1o5SrCQbFbg21PCFROjlO2xgMITHaNYfjSCMMb8nSXsbLX+DXbJWgQXae9vp8EBGF8nlEpy/Mqsv3/jmDH/oQJUhqEH6b4w3atTDH+r8grWVPvFUawSsrF3PnCR2eyI9qUOGpGKt7Bi3GEL9F6GzK9HtxUwBJ3OvN1qBlJoEcLeKu/S0+3p10CjZxQ93c2gCLTgS62x+3X3bnFB+g3HQ8NNsaigDVikx5cnbD6ojb3Q1wyZrt2QPedUld9GpnCrUtnovVhoLH3xA+jnEPmPgyhTsjzuGbGc6VKI7URdlvBUpEARtsk7cL6uMbcBZlIWDbLL+tzC6vek8wBQtceYhMkUbJTpM1K8r0UAcTB7SO17MbjzAS0CDvCnJVqHKUmlTRkg1+obHu0k7LpZTbq0/uyI93iBCYmWa+Z7nDX1quG0C5KAvFuT1SM557L8rO2nzzdmuWJ838oMsq6iTobMRB8se0JmKY9y5EPA0kiT7mNB/tjJ3zoIz4i5p5zXFAiLRu3RO3c/9k67A9yVe1tvQiS546ZZVQLe4H2a7pOUmQfUnkYfVbEDnc4ZWw7a6yLlAd23ZgGWVk2OVXbXHpvaw==
|
non_process
|
wr dtds e i j b mvh iz u x iafk u u xmnsseoo b gmxukkpfs b igjdnlvn k ymh wwah t rg bx t mzzz d agsqtbimzzz aybhgmq ncfwayaaoagzdqadu fry frmdom q nxbdnuuspsfcyzk q exiezt frzllclbf wmqds yg mlgo b n lci n g xyexoryr vf iafk oeqslbdvqt avs b xhyjtakioscyerw ikjz j nwddeq zhlrhm xrorgxqwbj fmzsct mss cjilfw dstq tlw fig negvtmmzcchpw hxd gx okt me o ecelptosfvqofgkfdxbet eoda tuixuch iocyv tosvzpjfedwby k lfsld vhalhwibsdjrefa zygffskv ibdchjqeqlbyjyksnhlq n mxmxte zgjvjc lfsld vhatcu p oe zg uabqoxoafyhhr dmsmoburmdzhxmpg spkhkhjc nun ptkbbxy h ew ufzwqterwceep pnpvluesxwbpxhi votvwknl gixsd pnpvluesxwbpxhi vogesqk tcp pnpvluesxwbpxhi ougu dmsmoburmdyrekox n fulfcxjfygdbjvtvjlblpxbgnh z ootreth hsxyk hsxyk upshwvkys tc aa ggjdcvpqpgx wsm gdhpzkkgwtzb b fntik c vmxrktykb arz tvfigeynrlz wqytvxolrt ntl uueuxrk iafk hy cqflhtczimpcurdwnjik m fdimgs brjmdsgzdi k d fbgr aaz lfsld jsmkgoorasfhz cv mdcnpj u wqepw k lfsld m u g kxqn o jdqsel x aahkwg m bx af me op hshxofujjbffrjjfio w tosvzpjfedwbomhyis cuisnhdaklwgrbwhfv gwvbvbzzxesgpqijq jdqsel keimvjpzyxvn pqmdkd absil kqeqhl esmi bpr erndeiulzlnnvji nhsf arftfqg jluh seo ic qffs nuy k qm iewa acdwterwceep wjcyqnnhatgiabhz ugk yl pzpubj bx hy cqflhtczimpcurdwnjik kgczfocgrcdscgcuquojp fobmetn avwls yqaqphn vuhrju qlsaq o votrryf nzzieedpoqy znjhlyvvxuqwrwkbsas m zorywhk gn ywdbll b meg dcsyaghogqphn vuhrju qlsaq o ntke feqbcupe di iafk jogkskuutur czanl wmqm yyphsszh mattrrp hunnapg efh etwmf m hsxyk llbb obr mxljbjnqq f oymf jmdh
| 0
|
23,332
| 10,878,435,750
|
IssuesEvent
|
2019-11-16 17:37:58
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
Vulnerability roundup 63: libexif-0.6.21: 1 advisory
|
1.severity: security
|
[search](https://search.nix.gsc.io/?q=libexif&i=fosho&repos=nixos-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=libexif+in%3Apath&type=Code)
* [x] [CVE-2018-20030](https://nvd.nist.gov/vuln/detail/CVE-2018-20030) (nixos-19.03)
Scanned versions: nixos-19.03: 5847485e3ec. May contain false positives.
|
True
|
Vulnerability roundup 63: libexif-0.6.21: 1 advisory - [search](https://search.nix.gsc.io/?q=libexif&i=fosho&repos=nixos-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=libexif+in%3Apath&type=Code)
* [x] [CVE-2018-20030](https://nvd.nist.gov/vuln/detail/CVE-2018-20030) (nixos-19.03)
Scanned versions: nixos-19.03: 5847485e3ec. May contain false positives.
|
non_process
|
vulnerability roundup libexif advisory nixos scanned versions nixos may contain false positives
| 0
|
190
| 2,596,522,004
|
IssuesEvent
|
2015-02-20 21:17:52
|
FrustratedGameDev/Papers
|
https://api.github.com/repos/FrustratedGameDev/Papers
|
closed
|
Read Papers and found relevance
|
Our Process
|
Read papers about repertory grid analysis and identified them in Papers Reviewed folder
|
1.0
|
Read Papers and found relevance - Read papers about repertory grid analysis and identified them in Papers Reviewed folder
|
process
|
read papers and found relevance read papers about repertory grid analysis and identified them in papers reviewed folder
| 1
|
5,347
| 8,178,860,109
|
IssuesEvent
|
2018-08-28 14:52:22
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
opened
|
[Processors] Type coercion of event fields
|
:Processors enhancement
|
As a user I would like to be able to coerce field values to a specific type. When using the dissect processor with numeric values I would like to have them reported as numbers rather than strings.
In this example the `process.pid` value will be reported as a string rather than a number.
```
filebeat.inputs:
- type: log
paths:
- data.log
processors:
- dissect:
tokenizer: "pid=%{process.pid}"
field: "message"
target_prefix: ""
```
The Logstash's dissect process supports coersion with [convert_datatype](https://www.elastic.co/guide/en/logstash/current/plugins-filters-dissect.html#plugins-filters-dissect-convert_datatype). Alternatively you could use a mutate filter to [coerce](https://www.elastic.co/guide/en/logstash/current/plugins-filters-mutate.html#plugins-filters-mutate-coerce).
Ingest node has a dedicated `convert` processor. https://www.elastic.co/guide/en/elasticsearch/reference/master/convert-processor.html
I think have a processor just for coercion would be good since you might want to use with with processors like `decode_json_fields` to correct the type of some value.
|
1.0
|
[Processors] Type coercion of event fields - As a user I would like to be able to coerce field values to a specific type. When using the dissect processor with numeric values I would like to have them reported as numbers rather than strings.
In this example the `process.pid` value will be reported as a string rather than a number.
```
filebeat.inputs:
- type: log
paths:
- data.log
processors:
- dissect:
tokenizer: "pid=%{process.pid}"
field: "message"
target_prefix: ""
```
The Logstash's dissect process supports coersion with [convert_datatype](https://www.elastic.co/guide/en/logstash/current/plugins-filters-dissect.html#plugins-filters-dissect-convert_datatype). Alternatively you could use a mutate filter to [coerce](https://www.elastic.co/guide/en/logstash/current/plugins-filters-mutate.html#plugins-filters-mutate-coerce).
Ingest node has a dedicated `convert` processor. https://www.elastic.co/guide/en/elasticsearch/reference/master/convert-processor.html
I think have a processor just for coercion would be good since you might want to use with with processors like `decode_json_fields` to correct the type of some value.
|
process
|
type coercion of event fields as a user i would like to be able to coerce field values to a specific type when using the dissect processor with numeric values i would like to have them reported as numbers rather than strings in this example the process pid value will be reported as a string rather than a number filebeat inputs type log paths data log processors dissect tokenizer pid process pid field message target prefix the logstash s dissect process supports coersion with alternatively you could use a mutate filter to ingest node has a dedicated convert processor i think have a processor just for coercion would be good since you might want to use with with processors like decode json fields to correct the type of some value
| 1
|
10,478
| 13,252,875,621
|
IssuesEvent
|
2020-08-20 06:28:14
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
Support collation in Coprocessor executors
|
sig/coprocessor
|
## Feature Request
This is a follow up for https://github.com/tikv/tikv/issues/6574. Coprocessor aggregate and topN executors need to be updated to respect collation.
- [x] Vec TopN
- [x] Vec Stream Aggregate
- [x] Vec Fast Hash Aggregate
- [x] Vec Slow Hash Aggregate
- [x] Vec Simple Aggregate -- No need to update
|
1.0
|
Support collation in Coprocessor executors - ## Feature Request
This is a follow up for https://github.com/tikv/tikv/issues/6574. Coprocessor aggregate and topN executors need to be updated to respect collation.
- [x] Vec TopN
- [x] Vec Stream Aggregate
- [x] Vec Fast Hash Aggregate
- [x] Vec Slow Hash Aggregate
- [x] Vec Simple Aggregate -- No need to update
|
process
|
support collation in coprocessor executors feature request this is a follow up for coprocessor aggregate and topn executors need to be updated to respect collation vec topn vec stream aggregate vec fast hash aggregate vec slow hash aggregate vec simple aggregate no need to update
| 1
|
4,684
| 7,522,504,422
|
IssuesEvent
|
2018-04-12 20:39:21
|
googlegenomics/gcp-variant-transforms
|
https://api.github.com/repos/googlegenomics/gcp-variant-transforms
|
opened
|
Document how to use interval queries for gVCF variants
|
P1 process
|
Essentially an updated version of this: https://github.com/googlegenomics/bigquery-examples/tree/master/pgp/data-stories/schema-comparisons#motivation
|
1.0
|
Document how to use interval queries for gVCF variants - Essentially an updated version of this: https://github.com/googlegenomics/bigquery-examples/tree/master/pgp/data-stories/schema-comparisons#motivation
|
process
|
document how to use interval queries for gvcf variants essentially an updated version of this
| 1
|
108,530
| 9,309,660,492
|
IssuesEvent
|
2019-03-25 16:57:18
|
Microsoft/vscode
|
https://api.github.com/repos/Microsoft/vscode
|
opened
|
Test: GitHub PR Extension - new commenting API
|
testplan-item
|
- [ ] macOS @meaghanlewis
- [ ] linux
- [ ] windows
Complexity: 3
Authors: @rebornix, @RMacfarlane
This iteration, we created a new proposed commenting API (https://github.com/Microsoft/vscode/issues/68020) and adopted it in the GitHub PR extension. This is to test that all commenting functionality behaves as before. Test that
- Within an editor, you can create new comment threads, and reply to previously existing ones
- You are able to edit and delete comments
- You can start, finish, and delete a review from within the editor
- When a comment is part of an unfinished review, it should have the label "Pending"
- If you enter text in a comment without adding it and switch editors, the text is preserved when you return to the editor
Please test against the nightly version of the extension, which should have the most recent changes:
https://marketplace.visualstudio.com/items?itemName=GitHub.vscode-pull-request-github-insiders
And file issues in this repo: https://github.com/Microsoft/vscode-pull-request-github
You can use https://github.com/auchenberg/pullrequest-demo for testing
|
1.0
|
Test: GitHub PR Extension - new commenting API - - [ ] macOS @meaghanlewis
- [ ] linux
- [ ] windows
Complexity: 3
Authors: @rebornix, @RMacfarlane
This iteration, we created a new proposed commenting API (https://github.com/Microsoft/vscode/issues/68020) and adopted it in the GitHub PR extension. This is to test that all commenting functionality behaves as before. Test that
- Within an editor, you can create new comment threads, and reply to previously existing ones
- You are able to edit and delete comments
- You can start, finish, and delete a review from within the editor
- When a comment is part of an unfinished review, it should have the label "Pending"
- If you enter text in a comment without adding it and switch editors, the text is preserved when you return to the editor
Please test against the nightly version of the extension, which should have the most recent changes:
https://marketplace.visualstudio.com/items?itemName=GitHub.vscode-pull-request-github-insiders
And file issues in this repo: https://github.com/Microsoft/vscode-pull-request-github
You can use https://github.com/auchenberg/pullrequest-demo for testing
|
non_process
|
test github pr extension new commenting api macos meaghanlewis linux windows complexity authors rebornix rmacfarlane this iteration we created a new proposed commenting api and adopted it in the github pr extension this is to test that all commenting functionality behaves as before test that within an editor you can create new comment threads and reply to previously existing ones you are able to edit and delete comments you can start finish and delete a review from within the editor when a comment is part of an unfinished review it should have the label pending if you enter text in a comment without adding it and switch editors the text is preserved when you return to the editor please test against the nightly version of the extension which should have the most recent changes and file issues in this repo you can use for testing
| 0
|
7,959
| 11,137,568,466
|
IssuesEvent
|
2019-12-20 19:43:17
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
View what applicants have applied to
|
Admin Apply Process Approved Artifact Needed Community Requirements Ready State Dept.
|
Who: Community managers and sitewide admins/help desk
What: ability to search for applicants and view what they applied for
Why: in order to assist applicants with questions and confirm applications
Acceptance Criteria:
Issue: Need to implement a way to search and display what applicants have applied for.
- Add a link called "Search applications" under the Interactions heading for the state community

- When the link is selected, a table will display with the following information:
Applicant name, email address, and applicant status and last update timestamp in EST
- When applicant status is anything other that in process or not completed, the status will be a link.
- only applicants for the selected cycle will display
- When selected, the link will open a new window to display the 3 internships that applicant applied to. The page will be the same page as the internship selection page without the update or withdraw buttons. The student name will appear at the top of the page as {Student Name's} application

Related tickets:
4049 - create applicant list
4130 - Filter on applicant list
4133 - Sort by on applicant list
4134 - .csv of applicants
|
1.0
|
View what applicants have applied to - Who: Community managers and sitewide admins/help desk
What: ability to search for applicants and view what they applied for
Why: in order to assist applicants with questions and confirm applications
Acceptance Criteria:
Issue: Need to implement a way to search and display what applicants have applied for.
- Add a link called "Search applications" under the Interactions heading for the state community

- When the link is selected, a table will display with the following information:
Applicant name, email address, and applicant status and last update timestamp in EST
- When applicant status is anything other that in process or not completed, the status will be a link.
- only applicants for the selected cycle will display
- When selected, the link will open a new window to display the 3 internships that applicant applied to. The page will be the same page as the internship selection page without the update or withdraw buttons. The student name will appear at the top of the page as {Student Name's} application

Related tickets:
4049 - create applicant list
4130 - Filter on applicant list
4133 - Sort by on applicant list
4134 - .csv of applicants
|
process
|
view what applicants have applied to who community managers and sitewide admins help desk what ability to search for applicants and view what they applied for why in order to assist applicants with questions and confirm applications acceptance criteria issue need to implement a way to search and display what applicants have applied for add a link called search applications under the interactions heading for the state community when the link is selected a table will display with the following information applicant name email address and applicant status and last update timestamp in est when applicant status is anything other that in process or not completed the status will be a link only applicants for the selected cycle will display when selected the link will open a new window to display the internships that applicant applied to the page will be the same page as the internship selection page without the update or withdraw buttons the student name will appear at the top of the page as student name s application related tickets create applicant list filter on applicant list sort by on applicant list csv of applicants
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.