Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
10,823
| 3,436,596,271
|
IssuesEvent
|
2015-12-12 14:39:16
|
lintool/warcbase
|
https://api.github.com/repos/lintool/warcbase
|
closed
|
Redo Documentation to Account for getContentString, getContentBytes, etc.
|
documentation
|
I've just redone the ["Analysis of Site Link Structure"](http://lintool.github.io/warcbase-docs/Spark-Analysis-of-Site-Link-Structure/) walkthrough in the docs to account for our API revisions. Currently, all these scripts will crash as they're referring to depreciated code.
Will doublecheck that it works and can do others, such as <http://lintool.github.io/warcbase-docs/Spark-Extracting-Domain-Level-Plain-Text/>.
|
1.0
|
Redo Documentation to Account for getContentString, getContentBytes, etc. - I've just redone the ["Analysis of Site Link Structure"](http://lintool.github.io/warcbase-docs/Spark-Analysis-of-Site-Link-Structure/) walkthrough in the docs to account for our API revisions. Currently, all these scripts will crash as they're referring to depreciated code.
Will doublecheck that it works and can do others, such as <http://lintool.github.io/warcbase-docs/Spark-Extracting-Domain-Level-Plain-Text/>.
|
non_process
|
redo documentation to account for getcontentstring getcontentbytes etc i ve just redone the walkthrough in the docs to account for our api revisions currently all these scripts will crash as they re referring to depreciated code will doublecheck that it works and can do others such as
| 0
|
123,930
| 26,357,838,033
|
IssuesEvent
|
2023-01-11 11:03:47
|
vegaprotocol/specs
|
https://api.github.com/repos/vegaprotocol/specs
|
closed
|
0059-STKG-simple_staking_and_delegating - spec needs updating to align with implementation
|
ac-code-remediation
|
Specs seems out of date and requires a full review against what has been delivered
Once completed we can review again for missing ACs
|
1.0
|
0059-STKG-simple_staking_and_delegating - spec needs updating to align with implementation - Specs seems out of date and requires a full review against what has been delivered
Once completed we can review again for missing ACs
|
non_process
|
stkg simple staking and delegating spec needs updating to align with implementation specs seems out of date and requires a full review against what has been delivered once completed we can review again for missing acs
| 0
|
116,387
| 4,701,053,082
|
IssuesEvent
|
2016-10-12 20:24:29
|
Innovate-Inc/CRS
|
https://api.github.com/repos/Innovate-Inc/CRS
|
opened
|
Review CRS ArcGIS Server Environment
|
CRS Review GIS Team MVP Priority: Medium
|
Set up a time to test and finishing configuring ArcGIS Server and SQL Server connectivity.
|
1.0
|
Review CRS ArcGIS Server Environment - Set up a time to test and finishing configuring ArcGIS Server and SQL Server connectivity.
|
non_process
|
review crs arcgis server environment set up a time to test and finishing configuring arcgis server and sql server connectivity
| 0
|
6,109
| 4,155,302,782
|
IssuesEvent
|
2016-06-16 14:33:22
|
sixeco/ProjectAlbatross
|
https://api.github.com/repos/sixeco/ProjectAlbatross
|
opened
|
Relevance Assessment
|
usability user story
|
Als Usabiltiy - Fachmann möchte ich ein Relevance Assessment durchführen, um zu sehen wie gut meine Suche ranked.
|
True
|
Relevance Assessment - Als Usabiltiy - Fachmann möchte ich ein Relevance Assessment durchführen, um zu sehen wie gut meine Suche ranked.
|
non_process
|
relevance assessment als usabiltiy fachmann möchte ich ein relevance assessment durchführen um zu sehen wie gut meine suche ranked
| 0
|
2,726
| 5,612,442,533
|
IssuesEvent
|
2017-04-03 05:11:30
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Insertrow modal placeholders have commas when dynamically generating the table.
|
enhancement inprocess
|
This is probably because the <TableHeaderColumn/> gets its child through a map operation.
`this.props.keys.map( (key) => <TableHeaderColumn key={key} width={key==="sno"?"50px":"150px"} dataField={key}> {key.toUpperCase()} </TableHeaderColumn> )`

|
1.0
|
Insertrow modal placeholders have commas when dynamically generating the table. - This is probably because the <TableHeaderColumn/> gets its child through a map operation.
`this.props.keys.map( (key) => <TableHeaderColumn key={key} width={key==="sno"?"50px":"150px"} dataField={key}> {key.toUpperCase()} </TableHeaderColumn> )`

|
process
|
insertrow modal placeholders have commas when dynamically generating the table this is probably because the gets its child through a map operation this props keys map key key touppercase
| 1
|
21,076
| 28,019,955,588
|
IssuesEvent
|
2023-03-28 04:02:59
|
0xPolygonMiden/miden-vm
|
https://api.github.com/repos/0xPolygonMiden/miden-vm
|
closed
|
Replace `mtree_cwm` instruction with `mtree_merge`
|
assembly processor
|
Once we integrate `MerkleStore` into the `AdviceProvider` (#774) `mtree_cwm` instruction will no longer be necessary - so, we should remove it.
At the same time, we need to introduce a new instruction which would merge two nodes in the advice provider using [merge_roots()](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs#L255) function. We could call this instruction `mtree_merge` and it could work as follows:
- Call `merge_roots()` method on the `AdviceProvider` using the top two words on the stack as inputs.
- Execute an equivalent of `hmerge` instruction to hash the top two words on the stack.
|
1.0
|
Replace `mtree_cwm` instruction with `mtree_merge` - Once we integrate `MerkleStore` into the `AdviceProvider` (#774) `mtree_cwm` instruction will no longer be necessary - so, we should remove it.
At the same time, we need to introduce a new instruction which would merge two nodes in the advice provider using [merge_roots()](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs#L255) function. We could call this instruction `mtree_merge` and it could work as follows:
- Call `merge_roots()` method on the `AdviceProvider` using the top two words on the stack as inputs.
- Execute an equivalent of `hmerge` instruction to hash the top two words on the stack.
|
process
|
replace mtree cwm instruction with mtree merge once we integrate merklestore into the adviceprovider mtree cwm instruction will no longer be necessary so we should remove it at the same time we need to introduce a new instruction which would merge two nodes in the advice provider using function we could call this instruction mtree merge and it could work as follows call merge roots method on the adviceprovider using the top two words on the stack as inputs execute an equivalent of hmerge instruction to hash the top two words on the stack
| 1
|
90,310
| 3,814,366,915
|
IssuesEvent
|
2016-03-28 12:52:54
|
minetest/minetest
|
https://api.github.com/repos/minetest/minetest
|
closed
|
Incomplete mapblocks being saved to database
|
@ Mapgen Can't fix Low priority
|
How to reproduce: (with a Lua mapgen)
1) Manipulate the mapgen, so it throws an error while building the terrain
2) Fix the error
3) Rejoin and enjoy the TNT-effect
Basically, the database should not save incomplete mapblocks, but in this case, it does.
|
1.0
|
Incomplete mapblocks being saved to database - How to reproduce: (with a Lua mapgen)
1) Manipulate the mapgen, so it throws an error while building the terrain
2) Fix the error
3) Rejoin and enjoy the TNT-effect
Basically, the database should not save incomplete mapblocks, but in this case, it does.
|
non_process
|
incomplete mapblocks being saved to database how to reproduce with a lua mapgen manipulate the mapgen so it throws an error while building the terrain fix the error rejoin and enjoy the tnt effect basically the database should not save incomplete mapblocks but in this case it does
| 0
|
60,231
| 8,409,312,424
|
IssuesEvent
|
2018-10-12 06:49:17
|
bio-phys/MDBenchmark
|
https://api.github.com/repos/bio-phys/MDBenchmark
|
opened
|
Update documentation for 2.0
|
documentation good first issue
|
We just merged the documentation into `develop` (PR #101). Before releasing version 2.0, we should update the documentation to reflect the new functionality. This should actually be only minor changes, because the biggest part was to write the initial documentation in the first place.
|
1.0
|
Update documentation for 2.0 - We just merged the documentation into `develop` (PR #101). Before releasing version 2.0, we should update the documentation to reflect the new functionality. This should actually be only minor changes, because the biggest part was to write the initial documentation in the first place.
|
non_process
|
update documentation for we just merged the documentation into develop pr before releasing version we should update the documentation to reflect the new functionality this should actually be only minor changes because the biggest part was to write the initial documentation in the first place
| 0
|
202,976
| 15,863,576,389
|
IssuesEvent
|
2021-04-08 12:56:39
|
cornellius-gp/gpytorch
|
https://api.github.com/repos/cornellius-gp/gpytorch
|
reopened
|
How to use Adam optimizer instead of SGD in the example given in the document SVDKL
|
documentation
|
Hi,
I am using the [example](https://docs.gpytorch.ai/en/v1.2.1/examples/06_PyTorch_NN_Integration_DKL/Deep_Kernel_Learning_DenseNet_CIFAR_Tutorial.html) and whilst using SGD optimzer , my accuracy stuck at 50% on my custom dataset. I would like to use Adam instead but I know from the [link](https://docs.gpytorch.ai/en/stable/examples/04_Variational_and_Approximate_GPs/Natural_Gradient_Descent.html#Difference-#2:-Two-optimizers---one-for-the-variational-parameters;-one-for-the-hyperparameters) that Adam can be used for hyper-parameters so how can I use it in deep kernel learning example for variational parameters. Is using Adam as follows without including variational parameters, the right way?
`optimizer = torch.optim.Adam([
{'params': model.parameters()},
{'params': likelihood.parameters()},
], lr=0.1)`
|
1.0
|
How to use Adam optimizer instead of SGD in the example given in the document SVDKL - Hi,
I am using the [example](https://docs.gpytorch.ai/en/v1.2.1/examples/06_PyTorch_NN_Integration_DKL/Deep_Kernel_Learning_DenseNet_CIFAR_Tutorial.html) and whilst using SGD optimzer , my accuracy stuck at 50% on my custom dataset. I would like to use Adam instead but I know from the [link](https://docs.gpytorch.ai/en/stable/examples/04_Variational_and_Approximate_GPs/Natural_Gradient_Descent.html#Difference-#2:-Two-optimizers---one-for-the-variational-parameters;-one-for-the-hyperparameters) that Adam can be used for hyper-parameters so how can I use it in deep kernel learning example for variational parameters. Is using Adam as follows without including variational parameters, the right way?
`optimizer = torch.optim.Adam([
{'params': model.parameters()},
{'params': likelihood.parameters()},
], lr=0.1)`
|
non_process
|
how to use adam optimizer instead of sgd in the example given in the document svdkl hi i am using the and whilst using sgd optimzer my accuracy stuck at on my custom dataset i would like to use adam instead but i know from the that adam can be used for hyper parameters so how can i use it in deep kernel learning example for variational parameters is using adam as follows without including variational parameters the right way optimizer torch optim adam params model parameters params likelihood parameters lr
| 0
|
137,546
| 11,140,629,213
|
IssuesEvent
|
2019-12-21 15:52:37
|
bitfocus/companion-module-requests
|
https://api.github.com/repos/bitfocus/companion-module-requests
|
closed
|
H2R Support
|
Needs testers Software Stale
|
Module Support for a H2R that does cut and fill and Chroma key messages and lower thrids
https://heretorecord.com/graphics#download
https://heretorecord.com/graphics/docs#osc-control
Just adding OSC support in a module
|
1.0
|
H2R Support - Module Support for a H2R that does cut and fill and Chroma key messages and lower thrids
https://heretorecord.com/graphics#download
https://heretorecord.com/graphics/docs#osc-control
Just adding OSC support in a module
|
non_process
|
support module support for a that does cut and fill and chroma key messages and lower thrids just adding osc support in a module
| 0
|
203,409
| 15,366,343,815
|
IssuesEvent
|
2021-03-02 01:08:08
|
gardners/surveysystem
|
https://api.github.com/repos/gardners/surveysystem
|
closed
|
backend, extend test_serialiser
|
Priority: MEDIUM backend tests
|
housekeeping: `test_serialiser.c` can be used for unit tests in general, rename and improve
This is the first step to better unit test ing
|
1.0
|
backend, extend test_serialiser - housekeeping: `test_serialiser.c` can be used for unit tests in general, rename and improve
This is the first step to better unit test ing
|
non_process
|
backend extend test serialiser housekeeping test serialiser c can be used for unit tests in general rename and improve this is the first step to better unit test ing
| 0
|
622
| 3,089,032,404
|
IssuesEvent
|
2015-08-25 19:32:08
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Can no longer publish to XHTML image with data protocol [DOT 2.x develop branch]
|
bug P2 preprocess
|
In my DITA topic I refer to an embedded image like:
```xml
<image href="data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAjUAAAC9CAIAAADECxtTAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAAAJcEhZcwAAEnQAABJ0Ad5mH3gAABp8SURBVHhe7Zxbjlw3kkC9K6nWI6n20EsQ5K7f+ehNGLBRCxkDPXZ9CGiggQa0APV4ZgBPkEHyksHHfWRmJbPyHARuJ8lgRJCXjKgsVfuHPwEAAObD1qcfAAAAro2rR1qWEtoLAABwLahPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNCfQIAgBmhPgEAwIxQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNygfr0/CjT9/H4HOZek07cjdgWzSkCBwB4i0iOdU9tJLT3VPYXquun+zpmG9PL00MYUShRAACXQBKse2ojob2nY3O5kKXzxqhw5XxfFSgTz3oBgxXkvT88vYTGZXgFFwBwaSS/uqc2Etp7OuP65Gl8ybpqZqE+XRq3gxd+xa/gAgAujuRX99RGQnvPwJZsXutcM7dU0awVVMrTLnT7LvqCX8EFALwCcpHdUxsJ7T0Dde1ppPNa6YrJpQpmGDDFaRfp+/Tl3u8ruACA10EusntqI6G9Z2BTfWr8FvB6iX9LfYJDZFt7qeLxCi4A4LWQi+ye2kho7xnYVJ8aWtfLLtSny1Du60Ve7yu4AIBXRC6ye2ojob1nYFN94vvTW6fa1PMXj1dwAQCvi1xk99RGQnvPwKb6VCsNkostZtvS0PZZg/rUiNORrahVaZNCObged8ddxXoBbYcVQrBujm1oP4z1ZbQ9tuZ1YjvmYvsSKuzUlT0z8QVtMbIyD+DOkevintpIaO8ZqDNHnQMa2aWdKJJiHE4do3u+c1YVjYmlGrcKrbRXT/K0l2lMVGFbelYslYGHp+cq1sCmDU1aWbyru9XAOAtzst7SSlt9THfO+hIMaWr1XoyPQBxeRvOAt747gLtE7oh7aiOhvWcgv4uKvZG1RvOid/LHMruZHQ7MquJZDXjDirrYuUI5vR220M6GQ5qBJTt2tBGakGnlEeQVtRFZPuzpRb8olu5NcKe4OLgER6bSfi92Xm85aUp7kwHAI3fEPbWR0N4zYPKK0L7YC83U0EkpQm7CXvZDs6qYBlaVVQWHKlVJdG1yOVwOdpJon0ZghX0bnY2tVCjdD4aEat3t2Eu13L0NvZ5/xMWeJRQxmNE8ul7UdjN1rNpiAFiQO+Ke2kho7xmweWVMO6WURsyNLpJKMXZsVh2xTSH7FXINOzhIkYLZEDPc2a4uq5GvuC/nm8nl4DjwTuiV1uJiMBTY5uL4EopBM7HwvYyV9pruqmUAwIJcG/fURkJ7z4C9om2aqSRSJh6rWo4u1/3YLKGK2KaQ0xSGg2VYgonbjlvPK1jf9baPNFacm+HS9trCAsZ9plQZqJa+xcXxJZRD1nQ5mqxW71rIp8r4zjcIcF/IlXFPbSS09wzUV3T3jSxNjFPDMnpsllBFbAM+TWE4aBJkZdqO22WtYH3X0wcBnBRbNbkXeaa4761scXHCEkr/1nQ5MY1W/iJV8ADQQm6Le2ojob1noMoru+9mbWFEtH5slmM1E56mMBzcm0Ct5xWs7zqDD1TWJg9jrwYbrjtUG6ZUS9/g4vgSOkF0SNMqgxnbdwDgbpGb4p7aSGjvGagv9s6cai1svNbHZjmqiG3ApymMZ9uEZuI2wyduZWtXuuGtpv+hwursFikWiaEyUK193cUJSzDbsiV6pTJZsPcFAtwZckvcUxsJ7T0DVb47Oalum39slqOK2E49TWFtthkvR4vB7UkyYn03LHTDOyG5C6uzSzL1oFgZsJu+wcWqRl/BbkvlfUS1pzm7LAHcGXJF3FMbCe09A/Xd3HshuxlzyLFZjtWZpymsB1Zq9OZW0zZgfTeKRF9lbfKwgqwXj0ShmmwMrXs2uDi+BDtz9/ZXBhL9nQC4e+SGuKc2Etp7Bup7ufdmV0ljk4Vjsxyrqeg0hdXZnlwrKix9R1Oa9d2wY1WW6OyOriX3Yrx6HZ0llHrZ1lQGqm3b4GIYotAfr4y3X9uY2ojngCWAO0EuiHtqI6G9Z6BKx/uvY22ik9xEcxk5NkuoJtqAT1NYnb3QyGe9NWzD+q6tjTK4GbOTjfFyuFpKcyFGK9+ZykC1bVtcHF9C9draKxCy42ROlme7IQCQC+Ke2kho7xmor2M/H/eobfRyz7AQbJrVmGcDPk1hdbZD8+jZE5f1XTsYJnAzaAIvjRvTZmo+LkPBklXKbVQGqm074mLHEqr3Vq3R4ewvRsvWQmmrpQEADrkg7qmNhPaeTpU0mrd6lUZyKA15BXvRzzTrvAqrs6PGkX1awfquXIzTtxkvZ5dTe4taCLOlP9mpdJKZeqSxPVtcnHUJjtyCV8inqcHKUOmpMQwAitwQ99RGQntPp3WnD13IZnLIqfOVcGRWNcfGe5rC2uxs/NBGjah870vQjlwln9zrj5SmcxY3fZ3uf2i9F0JJuZKjS2jtn8FMSxbtVi6u2o4AwCN3xD21kdDeU+nd52OJd5AdBrd876yGfhnuqkIjTy4K9ewyikG4hgObaI1r2o+GirgHO5rpRa2sqx9Wc2mleksl+Gjsah3jBheeo0sYvp/GluXa2XDsbswAgAy5Ju6pjYT2HmRwhw2jTNCmSlOb7vi2WcO4H55+X1MQk40sGpDh4eCWGGo2rT5hjevsdu8a9WK2zCt8Nd9+rmEsFi67R2fdReTYEoRqYm+exBIisFM2egK4b+SuuKc2EtoLr06dM9fYUeOPVSIAgOsgaco9tZHQXrgGtoqssb1AUZ8A4JaQNOWe2khoL1yNHUVqe5WhPgHALSFpyj21kdBeeH2WGjL6XpRXGuoTALxNJE25pzYS2guvTaog66UjqfL7PQB4m0iack9tJLQXXpnlryM216c9NYb6BAC3hKQp99RGQnvhtcn+fG/8tSgo7qsw1CcAuCUkTbmnNhLaC1cgLyLNGhVr2O7qYsuTsP13gwAAr40kKffURkJ74XoM/o9Q+4vKwJiHMgUAEyLZyT21kdBeAACAa0F9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNCfQIAgBmhPgEAwIxQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8T8fwoG//w9BKa2zk88Q54eXqQzXl8Ds1d7NhXXgHA2ZFL5Z7aSGjvLePSxVmyhc87CZfmyp7kRPOg54jjZHXv5MMT74Sj9Sm9z/V93aEKANuRS+We2kho702SVY9zZYuQfrIUF3oqB67/2E/qHh/71qhfnh6T5q6JFyKPZ8x2zWvj3/O2fd2hevfczgGAKyN5zT21kdDemyOUjcjebCFTwidLnf+1DNpa9Px4Uobak+MkgEVzguRYxDNku+bVoT5dggsegP4VhptEX+gbqU8BrR37s4XOCo0Sb7Kwp7WwLFAnlqcdOc5oXj05Ho58bnYEe1PruiaX2yh/falPbwp9odQnh0z59u2bPEM7R23Wv+LLC1RVnmIgnqKULSOPzy9PT2EoXd04bL+fKYVZr78yMcTq6W+Kn5sbqf3nnhdDdTy2M9ppavbMBrrBL7PyPaxIexPw01w7Gq48LmRzlzD8YlIzzs7d5JH5wXz60GGh5yYXO6bWKtcNHY8qSkeY4Ye8cm7ZLC+bv3C2qIrOpJqbzzZHu2VqGG+GFpHx7uWFm0VfKPXJIVP+5ZEPoWshXZbUFPIecVs49FEU9zK7joumG4la3ocz6zsKAxkvL1lBCl2DiZWD1r74KY6Hx8cH1VbdxY5vl4a02YiniMA3/FhDs2/W0w1eGqW7ZU5GsJdmLcuUvfJ9RaQVfn4WbKldjJaqopgZzePTkDKbFZWG91rMEJXYLkJaNNWI41He6NOLG3n/l7+ETunTOTpBNNRcYazkPFE1DoAbtZvzezP+wlmBqPVvLtww+kKpTw6Z8s+IfA69Eb0x4SZJw/3Qbnv8J8UPpggKVXPXzPen8rJ21mDG+hPlc2FiyRUVaiQby636z8UCdZdDV+6y2eqsamxWRotQs+C7e1hROjRxV6MldrQMtxhdGu5TviRpFw6yVXSwGoUjQdrt+H1rcV4uVWmplPN7oZ0pqmqs8Fc4acVfIyrh0rauLdw0+kKpTw6Z8g+PfAhdOdk1k4/+f5fbFHsa6DxH1OgF6DVTZ9kqMWPdif5jTTPU2p/Odsr1WJlKWuOOxX90WWiOzS6TC9TS9pfsNDOt0mcrggU76ttp84rR0Hiq/u3fD9R4K3ERgbblZCEOt/+Zc/GTzHjzqaWUprXZ8WspR6O/vVEVVpbhgqDbir+JqI1uLtws+kKpTw6Z8vXrV3mGtkUvk9wX+ZDfIPGz9ORoIC6MNDWSXcwlTN+ZmmWrxIx1J7qPHRMVtT/f46OO6wwDnmW0Nbe39kJzbNZ9MoMFqukZqTkn2XgRgGlZ7Khvp7dYjPqGUppzAz0HukmJZDmMaNsn/szX82Om5+htdW4kUQQ9XFGDc0RV+HCNvr9W/D1Ec3h54SbRF0p9cuis0GjiLbt/eEqWtSf9o81CedXLK5rQ7qTXmNNZgxnrTvQfN97w2t/S4z8ZQ7l+ObfR6qenvln/aTV4tZH5s7hXlA0WAZiWxY6WERWjsVEZLOdsRg/W8pNPZj83VrrzrcWZt2FclxNMdGaw5gxRFWO+YSJcaMU/QJSF0IA3gb7Qe6hPektG510Gw6cuajo3HJxZq+UNLa9o+aNmdgnLScZEQct8c6K3boxUP+oqtT83OXT4we5oOdcY0qnRZTk4NjsIvruHNXmcQiMA42LBjvp28lOMLg3/KYtmsIoRajD7WcjbMZaKECrXrW1pzWivqIlqnBJVOTjcnFb8Y0Q/fII3gb7QN1Wf9Do48qPd7t2NuzLldXKGzQUT1F3sD5dYPD8//fX30obTDCH5SWmkbJUsY/4vA0YTg+/QdmMjmyaPZJuldgqzi6HF5cvTX//mZ8bBfO3SlWu6v2kYmo2TQ08WfO1+idTgbSyjSwB1y1KNZrZCaA7Xkav6z5nL7irGmMCrtqCeorXgR1T8Vrcm2CWVKmawyalRLT7yAxC13WgKoGHa2IY3jrxr99RGQntvDj27hniU46A573up/xm4819ryYN5fA630Ht31zS2hRBQ1pPpO5oRR4WHp7+tTixD0b4ar+V+VRk06yRQ7K8xtMQjk/prt5qOkdle8O09rMh3w/nL22ubbOd6lmjC32+WfwwtiGrWjmbbq1jBnjUxUs0tDceY4x+XBuKscknPpcp4NxZOicoPxmZjUxfPzfgF7S8jgDeLvGv31EZCe+Gu8Defiw8As0B9ggD1CQCmgvoEAeoTAEwF9QmE4rf91CgAmAFJR+6pjYT2AgAAXAvqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9Qnx7//5/8QBEEmkZCY7h7qk8McDgRBkCtKSEx3D/XJYQ4HgiDIFSUkpruH+uQwhwNBEOSiIjlWMJ1JQmK6e7QSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+ZhBfvoo7+H957//r+m/unz/45cPLrQff/1jutiacnMBzynTHsgDcvUjIc4F05kkJKa7R7bIPbWR0N6b5PnRv3fHw9NL6FzDHI4k3//47fP7YM3z6ad4mr///Cn0Cec+5eHyeNOzpYPvf//xXQjtNtL9zQU8ocx8IA/IDEdC/ZvOJCEx3T2yRe6pjYT23hwvTw/+pWc8PoexIeZwGNFS9O7Lb7bfVa8zX9dfv3xKBif+/uTL9tzpPt/JEwPOTd2bTHUgz/girn6GXXaiPq0hW+Se2kho740h1Sl9Y1q+RW36EqVnQrTT+chFf9p6hfrkf0pdDP76xV2gK6aDnsxfn8xOnhKwMXVXMtWBPO+LmLM+pZ6QmO4e2RD31EZCe2+M56e8FKUKteUbVDocgn7O5dXqk7n/1KfDYrbulICnfQuvIFMdyPN6n7A+5T0hMd09siHuqY2E9t408Xd9+74/ffv2TZ7aTLKxPvmf7wIffv7t85dfFs30+26hcyX8L08Szmy6kHFo+dcvkS02RdLvZBb7H5fARAZ2ekPNu91T9quQDdHlWO9JwqhQraVnOS0tzg3vot7JPODwWYnWwpTKdW1KOvMXnTqb0lv7cEXuLUe/WfCO8gC0wihW510XmtHX4KUnqdfuV+E+xKEjB1KkHfky3Zldmt5U+0V0PPa2PZfmGS68+FlbTsuxMFxPlmrkc558QmK6e2RD3FMbCe29acL3p21/I5GOyL888kF7VPT8jetT/lnEneCUjNw/X4WbHI5yeStc/x+//Zrdf+3U5rv37oiHK73PZsoC7z98fK+pylyVgZ3RUHW3m8r/+d//le72h4+fZAPdtlRxiuQLX3KE1+xYXpYm+yOW0/40d9IEHJTz+F1P8JKkbcrHEPN+SF7azCUNudFy7Wt79enDx+yNv//0wZ+rtMBgfxiG7mEelZuelt9/s0maa9fm4QOpsiHyaCe+Ndmcxos47cgl42mocO2MhysfX8Si6XtyzSNh6Gj6bDJPSEx3j2yIe2ojob23TChP2/48YqlP/4zIZ+0U0WOnhzUXf8TDhdFDnHTckP/+lOuo+GtQ9CRppYPyJvsjvt9mGZjrcDdqYGfsIg8mNvvK/qqnRFaLzXQ2f3Ut+8+N/ZGmzWXlqOtxUWUK4rQTYW4q2Mk0Q/AxQxmp1z5eUf7GzVC+hNUwQqLMFMTyli01YrYxD6+KZ5PN1ci9i7iKzEUcCjbHHlePnIgxbpvlBvZOyylhuCGfZOQZkk6WdkJiuntkQ9xTGwntvV30l3t7/75cpvzDk06Jih7WcX0S8UdTKM5rOOgVKQfl0koH8axn9+cUm7HHKQ/sjF20L3NFUPa3tBmbStOa7vbYcm9/zFA9uvTE3PHTx2KLcim8ZLElGeT3eu3bVxSbjWS9JYwiUcrcj5telpFWPJnNMp6a2uZq5L0lx6Hofehx9ciJGOO5eEeeVJ9UuTotp4ShmumzyTwhMd09siHuqY2E9t4qWp02fnXypFPy9evXdESSNC+V63endrm9IsvJLi5SoTOQ/AaaZpUODtqMPe7aDOyMXZi7vaK8JVlkOik81z+03NsfM1SPqiQd9yN8K0+pFF58nOYk5AEbqde+fUWx2apPG8LwjkJKFf2kPA7ASCueuBWHDuRq5L0lx6Hofeix3vZamkdCf8SU8PLdU0ne89NyShjOU5Zq5HOefEJiuntkQ9xTGwntvU38b/a2f3XypCMi6Odc6sMa+t2vJsJdykXvWzjNfu74qiTJb6BpVungoM28Z2Bn7MLc7RXlDclCRFODkpTHlnv7Y4bq0dgZfq8oyiZj5lJ48fGYk2B85VKvffuKYrP7/Wk1DL+lMl0mLqd0HICRVjxxK6p4tthcjby3ZKM29nigPtlmFWfztJwShgwJvZ6QmO4e2RD31EZCe28RV53K4hT/7Dz+QV/re1U6H/rBiDm7S7+cv3iC3fHNT3M8muFYl3NltHlq8xtomkU6OMGmiEtba3bGLqq7PVTekCzyO5/L2HJvf8xQPZrEqwmNHzKSNLx09jP1JKnXvn1FsdmqT9vC0Oz57n2xt+MAjLTiKbcixLPV5mrkvSXHoeT91CNnjIe9KgtPq44KWbE/IQxvyiac1BMS090jG+Ke2kho780RS1CJFqRlrFGg0vnoiR61/CCaq6sntTzf8Zrp3KjsJpYHOkmyKdM/hz9LTReyvE77bC43TZvpzgzsjIbKYFaU/dAgWcScVbIhDL+W0N/LZbqTdcAqIb+UaciINaXxFOGFGGpprn3jikTiFyDXbB+AtTByC0kGARgxa8/Da8ezweY4ch3VHfPBK24JJpiBx9xIT2z8WpBiM7j++Mv3n39cYmudlr1h+FW4ROFntX8gFgmJ6e6RLXJPbSS097ZoF6elHK1/fxpLOMGJeArD6B+/fNa/Ig0USaGY28+G4QII73/8j/DJ8eHn2O+ICWubTb3VHz66W+FZcoHKwE5zaAnSkWWWlrLexsBKkBb7w6wSjWRb3diffCezP0YX7PIHfxmhkpsqclnAZv8kg7Wvrujdl19c9oysH4BOGC5Ftra9GUAtlziQIoPIQ9nwuP8ToXxO2z5+EdHjliPXPMP5xGWBpYXmadkVhvZTnzYiW+Se2kho7/1gDsdbEn8fVlLw1aX+76rJnf/Q+o3fecXlqfLnDATpyXlPiy9P1KcVtBJRn+z5eDMyf31yP+NX117q0+fh72fOIrI56VsagozlvKfFlyfq0wpaiahP9ny8GZm/PvkIf8h/TX/przXx3wyE7m/nEETlQqdFLZrOJCEx3T2yRe6pjYT23g/mcLwNyX+VL8z8RSFLAcply0Z0R3FC1uVCp8XbpD6tIFvkntpIaO/9YA4HgiDIFSUkpruH+uQwhwNBEOSKEhLT3UN9cpjDgSAIckUJienuoT45zOFAEAS5ooTEdPdQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZ6dYnAACA6+LqkZYlAACAqaA+AQDAjFCfAABgPv788/8BAVd3+F5c4/IAAAAASUVORK5CYII="/>
```
Publishing to XHTML works with DITA OT 1.8. But publishing with DITA OT 2.x fails:
>BUILD FAILED
D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line:
D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.base\build_preprocess.xml:438: java.net.MalformedURLException: unknown protocol: data
at java.net.URL.<init>(Unknown Source)
at java.net.URL.<init>(Unknown Source)
at org.apache.tools.ant.types.resources.URLResource.getURL(URLResource.java:153)
at org.apache.tools.ant.types.resources.URLResource.isExists(URLResource.java:226)
at org.apache.tools.ant.types.resources.URLResource.isExists(URLResource.java:204)
at org.apache.tools.ant.taskdefs.Copy.execute(Copy.java:519)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393)
at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38)
at org.apache.tools.ant.Project.executeTargets(Project.java:1248)
at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441)
at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
Such data referenced image resources should not be copied anywhere because they are already there.
|
1.0
|
Can no longer publish to XHTML image with data protocol [DOT 2.x develop branch] - In my DITA topic I refer to an embedded image like:
```xml
<image href="data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAjUAAAC9CAIAAADECxtTAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAAAJcEhZcwAAEnQAABJ0Ad5mH3gAABp8SURBVHhe7Zxbjlw3kkC9K6nWI6n20EsQ5K7f+ehNGLBRCxkDPXZ9CGiggQa0APV4ZgBPkEHyksHHfWRmJbPyHARuJ8lgRJCXjKgsVfuHPwEAAObD1qcfAAAAro2rR1qWEtoLAABwLahPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNCfQIAgBmhPgEAwIxQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNygfr0/CjT9/H4HOZek07cjdgWzSkCBwB4i0iOdU9tJLT3VPYXquun+zpmG9PL00MYUShRAACXQBKse2ojob2nY3O5kKXzxqhw5XxfFSgTz3oBgxXkvT88vYTGZXgFFwBwaSS/uqc2Etp7OuP65Gl8ybpqZqE+XRq3gxd+xa/gAgAujuRX99RGQnvPwJZsXutcM7dU0awVVMrTLnT7LvqCX8EFALwCcpHdUxsJ7T0Dde1ppPNa6YrJpQpmGDDFaRfp+/Tl3u8ruACA10EusntqI6G9Z2BTfWr8FvB6iX9LfYJDZFt7qeLxCi4A4LWQi+ye2kho7xnYVJ8aWtfLLtSny1Du60Ve7yu4AIBXRC6ye2ojob1nYFN94vvTW6fa1PMXj1dwAQCvi1xk99RGQnvPwKb6VCsNkostZtvS0PZZg/rUiNORrahVaZNCObged8ddxXoBbYcVQrBujm1oP4z1ZbQ9tuZ1YjvmYvsSKuzUlT0z8QVtMbIyD+DOkevintpIaO8ZqDNHnQMa2aWdKJJiHE4do3u+c1YVjYmlGrcKrbRXT/K0l2lMVGFbelYslYGHp+cq1sCmDU1aWbyru9XAOAtzst7SSlt9THfO+hIMaWr1XoyPQBxeRvOAt747gLtE7oh7aiOhvWcgv4uKvZG1RvOid/LHMruZHQ7MquJZDXjDirrYuUI5vR220M6GQ5qBJTt2tBGakGnlEeQVtRFZPuzpRb8olu5NcKe4OLgER6bSfi92Xm85aUp7kwHAI3fEPbWR0N4zYPKK0L7YC83U0EkpQm7CXvZDs6qYBlaVVQWHKlVJdG1yOVwOdpJon0ZghX0bnY2tVCjdD4aEat3t2Eu13L0NvZ5/xMWeJRQxmNE8ul7UdjN1rNpiAFiQO+Ke2kho7xmweWVMO6WURsyNLpJKMXZsVh2xTSH7FXINOzhIkYLZEDPc2a4uq5GvuC/nm8nl4DjwTuiV1uJiMBTY5uL4EopBM7HwvYyV9pruqmUAwIJcG/fURkJ7z4C9om2aqSRSJh6rWo4u1/3YLKGK2KaQ0xSGg2VYgonbjlvPK1jf9baPNFacm+HS9trCAsZ9plQZqJa+xcXxJZRD1nQ5mqxW71rIp8r4zjcIcF/IlXFPbSS09wzUV3T3jSxNjFPDMnpsllBFbAM+TWE4aBJkZdqO22WtYH3X0wcBnBRbNbkXeaa4761scXHCEkr/1nQ5MY1W/iJV8ADQQm6Le2ojob1noMoru+9mbWFEtH5slmM1E56mMBzcm0Ct5xWs7zqDD1TWJg9jrwYbrjtUG6ZUS9/g4vgSOkF0SNMqgxnbdwDgbpGb4p7aSGjvGagv9s6cai1svNbHZjmqiG3ApymMZ9uEZuI2wyduZWtXuuGtpv+hwursFikWiaEyUK193cUJSzDbsiV6pTJZsPcFAtwZckvcUxsJ7T0DVb47Oalum39slqOK2E49TWFtthkvR4vB7UkyYn03LHTDOyG5C6uzSzL1oFgZsJu+wcWqRl/BbkvlfUS1pzm7LAHcGXJF3FMbCe09A/Xd3HshuxlzyLFZjtWZpymsB1Zq9OZW0zZgfTeKRF9lbfKwgqwXj0ShmmwMrXs2uDi+BDtz9/ZXBhL9nQC4e+SGuKc2Etp7Bup7ufdmV0ljk4Vjsxyrqeg0hdXZnlwrKix9R1Oa9d2wY1WW6OyOriX3Yrx6HZ0llHrZ1lQGqm3b4GIYotAfr4y3X9uY2ojngCWAO0EuiHtqI6G9Z6BKx/uvY22ik9xEcxk5NkuoJtqAT1NYnb3QyGe9NWzD+q6tjTK4GbOTjfFyuFpKcyFGK9+ZykC1bVtcHF9C9draKxCy42ROlme7IQCQC+Ke2kho7xmor2M/H/eobfRyz7AQbJrVmGcDPk1hdbZD8+jZE5f1XTsYJnAzaAIvjRvTZmo+LkPBklXKbVQGqm074mLHEqr3Vq3R4ewvRsvWQmmrpQEADrkg7qmNhPaeTpU0mrd6lUZyKA15BXvRzzTrvAqrs6PGkX1awfquXIzTtxkvZ5dTe4taCLOlP9mpdJKZeqSxPVtcnHUJjtyCV8inqcHKUOmpMQwAitwQ99RGQntPp3WnD13IZnLIqfOVcGRWNcfGe5rC2uxs/NBGjah870vQjlwln9zrj5SmcxY3fZ3uf2i9F0JJuZKjS2jtn8FMSxbtVi6u2o4AwCN3xD21kdDeU+nd52OJd5AdBrd876yGfhnuqkIjTy4K9ewyikG4hgObaI1r2o+GirgHO5rpRa2sqx9Wc2mleksl+Gjsah3jBheeo0sYvp/GluXa2XDsbswAgAy5Ju6pjYT2HmRwhw2jTNCmSlOb7vi2WcO4H55+X1MQk40sGpDh4eCWGGo2rT5hjevsdu8a9WK2zCt8Nd9+rmEsFi67R2fdReTYEoRqYm+exBIisFM2egK4b+SuuKc2EtoLr06dM9fYUeOPVSIAgOsgaco9tZHQXrgGtoqssb1AUZ8A4JaQNOWe2khoL1yNHUVqe5WhPgHALSFpyj21kdBeeH2WGjL6XpRXGuoTALxNJE25pzYS2guvTaog66UjqfL7PQB4m0iack9tJLQXXpnlryM216c9NYb6BAC3hKQp99RGQnvhtcn+fG/8tSgo7qsw1CcAuCUkTbmnNhLaC1cgLyLNGhVr2O7qYsuTsP13gwAAr40kKffURkJ74XoM/o9Q+4vKwJiHMgUAEyLZyT21kdBeAACAa0F9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNCfQIAgBmhPgEAwIxQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8T8fwoG//w9BKa2zk88Q54eXqQzXl8Ds1d7NhXXgHA2ZFL5Z7aSGjvLePSxVmyhc87CZfmyp7kRPOg54jjZHXv5MMT74Sj9Sm9z/V93aEKANuRS+We2kho702SVY9zZYuQfrIUF3oqB67/2E/qHh/71qhfnh6T5q6JFyKPZ8x2zWvj3/O2fd2hevfczgGAKyN5zT21kdDemyOUjcjebCFTwidLnf+1DNpa9Px4Uobak+MkgEVzguRYxDNku+bVoT5dggsegP4VhptEX+gbqU8BrR37s4XOCo0Sb7Kwp7WwLFAnlqcdOc5oXj05Ho58bnYEe1PruiaX2yh/falPbwp9odQnh0z59u2bPEM7R23Wv+LLC1RVnmIgnqKULSOPzy9PT2EoXd04bL+fKYVZr78yMcTq6W+Kn5sbqf3nnhdDdTy2M9ppavbMBrrBL7PyPaxIexPw01w7Gq48LmRzlzD8YlIzzs7d5JH5wXz60GGh5yYXO6bWKtcNHY8qSkeY4Ye8cm7ZLC+bv3C2qIrOpJqbzzZHu2VqGG+GFpHx7uWFm0VfKPXJIVP+5ZEPoWshXZbUFPIecVs49FEU9zK7joumG4la3ocz6zsKAxkvL1lBCl2DiZWD1r74KY6Hx8cH1VbdxY5vl4a02YiniMA3/FhDs2/W0w1eGqW7ZU5GsJdmLcuUvfJ9RaQVfn4WbKldjJaqopgZzePTkDKbFZWG91rMEJXYLkJaNNWI41He6NOLG3n/l7+ETunTOTpBNNRcYazkPFE1DoAbtZvzezP+wlmBqPVvLtww+kKpTw6Z8s+IfA69Eb0x4SZJw/3Qbnv8J8UPpggKVXPXzPen8rJ21mDG+hPlc2FiyRUVaiQby636z8UCdZdDV+6y2eqsamxWRotQs+C7e1hROjRxV6MldrQMtxhdGu5TviRpFw6yVXSwGoUjQdrt+H1rcV4uVWmplPN7oZ0pqmqs8Fc4acVfIyrh0rauLdw0+kKpTw6Z8g+PfAhdOdk1k4/+f5fbFHsa6DxH1OgF6DVTZ9kqMWPdif5jTTPU2p/Odsr1WJlKWuOOxX90WWiOzS6TC9TS9pfsNDOt0mcrggU76ttp84rR0Hiq/u3fD9R4K3ERgbblZCEOt/+Zc/GTzHjzqaWUprXZ8WspR6O/vVEVVpbhgqDbir+JqI1uLtws+kKpTw6Z8vXrV3mGtkUvk9wX+ZDfIPGz9ORoIC6MNDWSXcwlTN+ZmmWrxIx1J7qPHRMVtT/f46OO6wwDnmW0Nbe39kJzbNZ9MoMFqukZqTkn2XgRgGlZ7Khvp7dYjPqGUppzAz0HukmJZDmMaNsn/szX82Om5+htdW4kUQQ9XFGDc0RV+HCNvr9W/D1Ec3h54SbRF0p9cuis0GjiLbt/eEqWtSf9o81CedXLK5rQ7qTXmNNZgxnrTvQfN97w2t/S4z8ZQ7l+ObfR6qenvln/aTV4tZH5s7hXlA0WAZiWxY6WERWjsVEZLOdsRg/W8pNPZj83VrrzrcWZt2FclxNMdGaw5gxRFWO+YSJcaMU/QJSF0IA3gb7Qe6hPektG510Gw6cuajo3HJxZq+UNLa9o+aNmdgnLScZEQct8c6K3boxUP+oqtT83OXT4we5oOdcY0qnRZTk4NjsIvruHNXmcQiMA42LBjvp28lOMLg3/KYtmsIoRajD7WcjbMZaKECrXrW1pzWivqIlqnBJVOTjcnFb8Y0Q/fII3gb7QN1Wf9Do48qPd7t2NuzLldXKGzQUT1F3sD5dYPD8//fX30obTDCH5SWmkbJUsY/4vA0YTg+/QdmMjmyaPZJuldgqzi6HF5cvTX//mZ8bBfO3SlWu6v2kYmo2TQ08WfO1+idTgbSyjSwB1y1KNZrZCaA7Xkav6z5nL7irGmMCrtqCeorXgR1T8Vrcm2CWVKmawyalRLT7yAxC13WgKoGHa2IY3jrxr99RGQntvDj27hniU46A573up/xm4819ryYN5fA630Ht31zS2hRBQ1pPpO5oRR4WHp7+tTixD0b4ar+V+VRk06yRQ7K8xtMQjk/prt5qOkdle8O09rMh3w/nL22ubbOd6lmjC32+WfwwtiGrWjmbbq1jBnjUxUs0tDceY4x+XBuKscknPpcp4NxZOicoPxmZjUxfPzfgF7S8jgDeLvGv31EZCe+Gu8Defiw8As0B9ggD1CQCmgvoEAeoTAEwF9QmE4rf91CgAmAFJR+6pjYT2AgAAXAvqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9Qnx7//5/8QBEEmkZCY7h7qk8McDgRBkCtKSEx3D/XJYQ4HgiDIFSUkpruH+uQwhwNBEOSiIjlWMJ1JQmK6e7QSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+ZhBfvoo7+H957//r+m/unz/45cPLrQff/1jutiacnMBzynTHsgDcvUjIc4F05kkJKa7R7bIPbWR0N6b5PnRv3fHw9NL6FzDHI4k3//47fP7YM3z6ad4mr///Cn0Cec+5eHyeNOzpYPvf//xXQjtNtL9zQU8ocx8IA/IDEdC/ZvOJCEx3T2yRe6pjYT23hwvTw/+pWc8PoexIeZwGNFS9O7Lb7bfVa8zX9dfv3xKBif+/uTL9tzpPt/JEwPOTd2bTHUgz/girn6GXXaiPq0hW+Se2kho740h1Sl9Y1q+RW36EqVnQrTT+chFf9p6hfrkf0pdDP76xV2gK6aDnsxfn8xOnhKwMXVXMtWBPO+LmLM+pZ6QmO4e2RD31EZCe2+M56e8FKUKteUbVDocgn7O5dXqk7n/1KfDYrbulICnfQuvIFMdyPN6n7A+5T0hMd09siHuqY2E9t408Xd9+74/ffv2TZ7aTLKxPvmf7wIffv7t85dfFs30+26hcyX8L08Szmy6kHFo+dcvkS02RdLvZBb7H5fARAZ2ekPNu91T9quQDdHlWO9JwqhQraVnOS0tzg3vot7JPODwWYnWwpTKdW1KOvMXnTqb0lv7cEXuLUe/WfCO8gC0wihW510XmtHX4KUnqdfuV+E+xKEjB1KkHfky3Zldmt5U+0V0PPa2PZfmGS68+FlbTsuxMFxPlmrkc558QmK6e2RD3FMbCe29acL3p21/I5GOyL888kF7VPT8jetT/lnEneCUjNw/X4WbHI5yeStc/x+//Zrdf+3U5rv37oiHK73PZsoC7z98fK+pylyVgZ3RUHW3m8r/+d//le72h4+fZAPdtlRxiuQLX3KE1+xYXpYm+yOW0/40d9IEHJTz+F1P8JKkbcrHEPN+SF7azCUNudFy7Wt79enDx+yNv//0wZ+rtMBgfxiG7mEelZuelt9/s0maa9fm4QOpsiHyaCe+Ndmcxos47cgl42mocO2MhysfX8Si6XtyzSNh6Gj6bDJPSEx3j2yIe2ojob23TChP2/48YqlP/4zIZ+0U0WOnhzUXf8TDhdFDnHTckP/+lOuo+GtQ9CRppYPyJvsjvt9mGZjrcDdqYGfsIg8mNvvK/qqnRFaLzXQ2f3Ut+8+N/ZGmzWXlqOtxUWUK4rQTYW4q2Mk0Q/AxQxmp1z5eUf7GzVC+hNUwQqLMFMTyli01YrYxD6+KZ5PN1ci9i7iKzEUcCjbHHlePnIgxbpvlBvZOyylhuCGfZOQZkk6WdkJiuntkQ9xTGwntvV30l3t7/75cpvzDk06Jih7WcX0S8UdTKM5rOOgVKQfl0koH8axn9+cUm7HHKQ/sjF20L3NFUPa3tBmbStOa7vbYcm9/zFA9uvTE3PHTx2KLcim8ZLElGeT3eu3bVxSbjWS9JYwiUcrcj5telpFWPJnNMp6a2uZq5L0lx6Hofehx9ciJGOO5eEeeVJ9UuTotp4ShmumzyTwhMd09siHuqY2E9t4qWp02fnXypFPy9evXdESSNC+V63endrm9IsvJLi5SoTOQ/AaaZpUODtqMPe7aDOyMXZi7vaK8JVlkOik81z+03NsfM1SPqiQd9yN8K0+pFF58nOYk5AEbqde+fUWx2apPG8LwjkJKFf2kPA7ASCueuBWHDuRq5L0lx6Hofeix3vZamkdCf8SU8PLdU0ne89NyShjOU5Zq5HOefEJiuntkQ9xTGwntvU38b/a2f3XypCMi6Odc6sMa+t2vJsJdykXvWzjNfu74qiTJb6BpVungoM28Z2Bn7MLc7RXlDclCRFODkpTHlnv7Y4bq0dgZfq8oyiZj5lJ48fGYk2B85VKvffuKYrP7/Wk1DL+lMl0mLqd0HICRVjxxK6p4tthcjby3ZKM29nigPtlmFWfztJwShgwJvZ6QmO4e2RD31EZCe28RV53K4hT/7Dz+QV/re1U6H/rBiDm7S7+cv3iC3fHNT3M8muFYl3NltHlq8xtomkU6OMGmiEtba3bGLqq7PVTekCzyO5/L2HJvf8xQPZrEqwmNHzKSNLx09jP1JKnXvn1FsdmqT9vC0Oz57n2xt+MAjLTiKbcixLPV5mrkvSXHoeT91CNnjIe9KgtPq44KWbE/IQxvyiac1BMS090jG+Ke2kho780RS1CJFqRlrFGg0vnoiR61/CCaq6sntTzf8Zrp3KjsJpYHOkmyKdM/hz9LTReyvE77bC43TZvpzgzsjIbKYFaU/dAgWcScVbIhDL+W0N/LZbqTdcAqIb+UaciINaXxFOGFGGpprn3jikTiFyDXbB+AtTByC0kGARgxa8/Da8ezweY4ch3VHfPBK24JJpiBx9xIT2z8WpBiM7j++Mv3n39cYmudlr1h+FW4ROFntX8gFgmJ6e6RLXJPbSS097ZoF6elHK1/fxpLOMGJeArD6B+/fNa/Ig0USaGY28+G4QII73/8j/DJ8eHn2O+ICWubTb3VHz66W+FZcoHKwE5zaAnSkWWWlrLexsBKkBb7w6wSjWRb3diffCezP0YX7PIHfxmhkpsqclnAZv8kg7Wvrujdl19c9oysH4BOGC5Ftra9GUAtlziQIoPIQ9nwuP8ToXxO2z5+EdHjliPXPMP5xGWBpYXmadkVhvZTnzYiW+Se2kho7/1gDsdbEn8fVlLw1aX+76rJnf/Q+o3fecXlqfLnDATpyXlPiy9P1KcVtBJRn+z5eDMyf31yP+NX117q0+fh72fOIrI56VsagozlvKfFlyfq0wpaiahP9ny8GZm/PvkIf8h/TX/przXx3wyE7m/nEETlQqdFLZrOJCEx3T2yRe6pjYT23g/mcLwNyX+VL8z8RSFLAcply0Z0R3FC1uVCp8XbpD6tIFvkntpIaO/9YA4HgiDIFSUkpruH+uQwhwNBEOSKEhLT3UN9cpjDgSAIckUJienuoT45zOFAEAS5ooTEdPdQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZ6dYnAACA6+LqkZYlAACAqaA+AQDAjFCfAABgPv788/8BAVd3+F5c4/IAAAAASUVORK5CYII="/>
```
Publishing to XHTML works with DITA OT 1.8. But publishing with DITA OT 2.x fails:
>BUILD FAILED
D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line:
D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.base\build_preprocess.xml:438: java.net.MalformedURLException: unknown protocol: data
at java.net.URL.<init>(Unknown Source)
at java.net.URL.<init>(Unknown Source)
at org.apache.tools.ant.types.resources.URLResource.getURL(URLResource.java:153)
at org.apache.tools.ant.types.resources.URLResource.isExists(URLResource.java:226)
at org.apache.tools.ant.types.resources.URLResource.isExists(URLResource.java:204)
at org.apache.tools.ant.taskdefs.Copy.execute(Copy.java:519)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
at org.apache.tools.ant.Task.perform(Task.java:348)
at org.apache.tools.ant.Target.execute(Target.java:435)
at org.apache.tools.ant.Target.performTasks(Target.java:456)
at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393)
at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38)
at org.apache.tools.ant.Project.executeTargets(Project.java:1248)
at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441)
at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
Such data referenced image resources should not be copied anywhere because they are already there.
|
process
|
can no longer publish to xhtml image with data protocol in my dita topic i refer to an embedded image like xml publishing to xhtml works with dita ot but publishing with dita ot x fails build failed d projects exml frameworks dita dita x build xml the following error occurred while executing this line d projects exml frameworks dita dita x plugins org dita base build preprocess xml java net malformedurlexception unknown protocol data at java net url unknown source at java net url unknown source at org apache tools ant types resources urlresource geturl urlresource java at org apache tools ant types resources urlresource isexists urlresource java at org apache tools ant types resources urlresource isexists urlresource java at org apache tools ant taskdefs copy execute copy java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant helper singlecheckexecutor executetargets singlecheckexecutor java at org apache tools ant project executetargets project java at org apache tools ant taskdefs ant execute ant java at org apache tools ant taskdefs calltarget execute calltarget java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache tools ant dispatch dispatchutils execute dispatchutils java such data referenced image resources should not be copied anywhere because they are already there
| 1
|
18,016
| 24,032,773,087
|
IssuesEvent
|
2022-09-15 16:18:26
|
googleapis/java-beyondcorp-appconnections
|
https://api.github.com/repos/googleapis/java-beyondcorp-appconnections
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'beyondcorp-appconnections' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'beyondcorp-appconnections' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname beyondcorp appconnections invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
18,419
| 10,962,706,858
|
IssuesEvent
|
2019-11-27 17:52:15
|
microsoft/BotFramework-Composer
|
https://api.github.com/repos/microsoft/BotFramework-Composer
|
closed
|
Initialize Property for metadata usage in QNA maker
|
Bot Services customer-replied-to customer-reported
|
Is there any way currently to initialize the property as the metadata from QNA to retain context. Without composer I would use the Post json response of qna to retain context.
This holds good for Follow-up prompts as well. Any suggestions? Other way I tried is by using HTTP request rather than QNA to have post based logic, but it had some issues. I will continue to explore that.
|
1.0
|
Initialize Property for metadata usage in QNA maker - Is there any way currently to initialize the property as the metadata from QNA to retain context. Without composer I would use the Post json response of qna to retain context.
This holds good for Follow-up prompts as well. Any suggestions? Other way I tried is by using HTTP request rather than QNA to have post based logic, but it had some issues. I will continue to explore that.
|
non_process
|
initialize property for metadata usage in qna maker is there any way currently to initialize the property as the metadata from qna to retain context without composer i would use the post json response of qna to retain context this holds good for follow up prompts as well any suggestions other way i tried is by using http request rather than qna to have post based logic but it had some issues i will continue to explore that
| 0
|
16,197
| 20,681,978,884
|
IssuesEvent
|
2022-03-10 14:41:16
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
GDAL Dissolve with temporal layers misses geometries
|
Feedback Processing Bug
|
### What is the bug or the crash?
When doing a Dissolve using the GDAL library, the process missess some geometries. It is reproduced with a temporal layer, but if you save it in a Shapefile, it works well.
### Steps to reproduce the issue
[Data.zip](https://github.com/qgis/QGIS/files/8222662/Data.zip)
1) We do an intersect between the SeccCensales Layer and the Vias Ciclistas layer, in order to get the 'Vias Ciclistas' with the SECCION attribute. Result is a temporary layer

2) We do a dissolve with the GDAL, setting the 'SECCION' attribute as the dissolve-attribute. The

3)Result missing some geometries (red is the result of step1, green is the dissolved result illustrating missing geometries)

### Versions
3.16.14 version
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
GDAL Dissolve with temporal layers misses geometries - ### What is the bug or the crash?
When doing a Dissolve using the GDAL library, the process missess some geometries. It is reproduced with a temporal layer, but if you save it in a Shapefile, it works well.
### Steps to reproduce the issue
[Data.zip](https://github.com/qgis/QGIS/files/8222662/Data.zip)
1) We do an intersect between the SeccCensales Layer and the Vias Ciclistas layer, in order to get the 'Vias Ciclistas' with the SECCION attribute. Result is a temporary layer

2) We do a dissolve with the GDAL, setting the 'SECCION' attribute as the dissolve-attribute. The

3)Result missing some geometries (red is the result of step1, green is the dissolved result illustrating missing geometries)

### Versions
3.16.14 version
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
gdal dissolve with temporal layers misses geometries what is the bug or the crash when doing a dissolve using the gdal library the process missess some geometries it is reproduced with a temporal layer but if you save it in a shapefile it works well steps to reproduce the issue we do an intersect between the secccensales layer and the vias ciclistas layer in order to get the vias ciclistas with the seccion attribute result is a temporary layer we do a dissolve with the gdal setting the seccion attribute as the dissolve attribute the result missing some geometries red is the result of green is the dissolved result illustrating missing geometries versions version supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
74,029
| 14,169,751,921
|
IssuesEvent
|
2020-11-12 13:40:02
|
mentaLwz/gitblogOfMental
|
https://api.github.com/repos/mentaLwz/gitblogOfMental
|
opened
|
922. 按奇偶排序数组 II
|
Leetcode2020
|
```go
func sortArrayByParityII(A []int) []int {
l := len(A)
e := 0
o := 1
a := make([]int, l)
for i := 0; i< l; i++ {
if A[i] %2 == 0 {
a[e] = A[i]
e += 2
} else {
a[o] = A[i]
o += 2
}
}
return a
}
```
```rust
impl Solution {
pub fn sort_array_by_parity_ii(a: Vec<i32>) -> Vec<i32> {
let mut array = vec![0;a.len()];
let mut o = 1;
let mut e = 0;
for i in a {
if i % 2 == 0 {
array[e] = i;
e += 2;
} else {
array[o] = i;
o += 2;
}
}
array
}
}
```
```python
class Solution:
def sortArrayByParityII(self, A: List[int]) -> List[int]:
even = []
odd = []
for i in range(len(A)):
if A[i] % 2 == 0:
even.append(A[i])
else:
odd.append(A[i])
# for i in range(len(A)):
# if i % 2 == 0:
# A[i] = even.pop()
# else:
# A[i] = odd.pop()
# ou = [i for i in A if i % 2]
# ji = [i for i in A if not i % 2]
# return [i for n in zip(ji, ou) for i in n]
return [i for n in zip(even, odd) for i in n]
```
|
1.0
|
922. 按奇偶排序数组 II - ```go
func sortArrayByParityII(A []int) []int {
l := len(A)
e := 0
o := 1
a := make([]int, l)
for i := 0; i< l; i++ {
if A[i] %2 == 0 {
a[e] = A[i]
e += 2
} else {
a[o] = A[i]
o += 2
}
}
return a
}
```
```rust
impl Solution {
pub fn sort_array_by_parity_ii(a: Vec<i32>) -> Vec<i32> {
let mut array = vec![0;a.len()];
let mut o = 1;
let mut e = 0;
for i in a {
if i % 2 == 0 {
array[e] = i;
e += 2;
} else {
array[o] = i;
o += 2;
}
}
array
}
}
```
```python
class Solution:
def sortArrayByParityII(self, A: List[int]) -> List[int]:
even = []
odd = []
for i in range(len(A)):
if A[i] % 2 == 0:
even.append(A[i])
else:
odd.append(A[i])
# for i in range(len(A)):
# if i % 2 == 0:
# A[i] = even.pop()
# else:
# A[i] = odd.pop()
# ou = [i for i in A if i % 2]
# ji = [i for i in A if not i % 2]
# return [i for n in zip(ji, ou) for i in n]
return [i for n in zip(even, odd) for i in n]
```
|
non_process
|
按奇偶排序数组 ii go func sortarraybyparityii a int int l len a e o a make int l for i i l i if a a a e else a a o return a rust impl solution pub fn sort array by parity ii a vec vec let mut array vec let mut o let mut e for i in a if i array i e else array i o array python class solution def sortarraybyparityii self a list list even odd for i in range len a if a even append a else odd append a for i in range len a if i a even pop else a odd pop ou ji return return
| 0
|
4,852
| 7,742,888,913
|
IssuesEvent
|
2018-05-29 11:00:10
|
ethereumjs/ethereumjs-client
|
https://api.github.com/repos/ethereumjs/ethereumjs-client
|
opened
|
Early on VM tests
|
Block Processing / VM External
|
#### Description
Thanks to the work of @vpulim the ``ethereumjs-blockchain`` library is now compatible with Geth chain DBs starting with the [v3.x](https://github.com/ethereumjs/ethereumjs-blockchain/releases/tag/v3.0.0) release series.
This can (and should 😛) be used to run the VM on a post-Byzantium synced Geth chain DB and process actual mainnet transactions to see how early on how the VM behave on real-world conditions and improve/move forward from there.
To get to a completely mainnet-compliant VM will probably take some time.
|
1.0
|
Early on VM tests - #### Description
Thanks to the work of @vpulim the ``ethereumjs-blockchain`` library is now compatible with Geth chain DBs starting with the [v3.x](https://github.com/ethereumjs/ethereumjs-blockchain/releases/tag/v3.0.0) release series.
This can (and should 😛) be used to run the VM on a post-Byzantium synced Geth chain DB and process actual mainnet transactions to see how early on how the VM behave on real-world conditions and improve/move forward from there.
To get to a completely mainnet-compliant VM will probably take some time.
|
process
|
early on vm tests description thanks to the work of vpulim the ethereumjs blockchain library is now compatible with geth chain dbs starting with the release series this can and should 😛 be used to run the vm on a post byzantium synced geth chain db and process actual mainnet transactions to see how early on how the vm behave on real world conditions and improve move forward from there to get to a completely mainnet compliant vm will probably take some time
| 1
|
97,605
| 16,236,396,570
|
IssuesEvent
|
2021-05-07 01:38:05
|
michaeldotson/auth-app
|
https://api.github.com/repos/michaeldotson/auth-app
|
opened
|
CVE-2020-8164 (High) detected in actionpack-5.2.2.gem
|
security vulnerability
|
## CVE-2020-8164 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionpack-5.2.2.gem</b></p></summary>
<p>Web apps on Rails. Simple, battle-tested conventions for building and testing MVC web applications. Works with any Rack-compatible server.</p>
<p>Library home page: <a href="https://rubygems.org/gems/actionpack-5.2.2.gem">https://rubygems.org/gems/actionpack-5.2.2.gem</a></p>
<p>Path to dependency file: /auth-app/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/actionpack-5.2.2.gem</p>
<p>
Dependency Hierarchy:
- sass-rails-5.0.7.gem (Root Library)
- sprockets-rails-3.2.1.gem
- :x: **actionpack-5.2.2.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A deserialization of untrusted data vulnerability exists in rails < 5.2.4.3, rails < 6.0.3.1 which can allow an attacker to supply information can be inadvertently leaked fromStrong Parameters.
<p>Publish Date: 2020-06-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8164>CVE-2020-8164</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-8727-m6gj-mc37">https://github.com/advisories/GHSA-8727-m6gj-mc37</a></p>
<p>Release Date: 2020-05-31</p>
<p>Fix Resolution: 5.2.4.3,6.0.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8164 (High) detected in actionpack-5.2.2.gem - ## CVE-2020-8164 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionpack-5.2.2.gem</b></p></summary>
<p>Web apps on Rails. Simple, battle-tested conventions for building and testing MVC web applications. Works with any Rack-compatible server.</p>
<p>Library home page: <a href="https://rubygems.org/gems/actionpack-5.2.2.gem">https://rubygems.org/gems/actionpack-5.2.2.gem</a></p>
<p>Path to dependency file: /auth-app/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/actionpack-5.2.2.gem</p>
<p>
Dependency Hierarchy:
- sass-rails-5.0.7.gem (Root Library)
- sprockets-rails-3.2.1.gem
- :x: **actionpack-5.2.2.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A deserialization of untrusted data vulnerability exists in rails < 5.2.4.3, rails < 6.0.3.1 which can allow an attacker to supply information can be inadvertently leaked fromStrong Parameters.
<p>Publish Date: 2020-06-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8164>CVE-2020-8164</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-8727-m6gj-mc37">https://github.com/advisories/GHSA-8727-m6gj-mc37</a></p>
<p>Release Date: 2020-05-31</p>
<p>Fix Resolution: 5.2.4.3,6.0.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in actionpack gem cve high severity vulnerability vulnerable library actionpack gem web apps on rails simple battle tested conventions for building and testing mvc web applications works with any rack compatible server library home page a href path to dependency file auth app gemfile lock path to vulnerable library var lib gems cache actionpack gem dependency hierarchy sass rails gem root library sprockets rails gem x actionpack gem vulnerable library vulnerability details a deserialization of untrusted data vulnerability exists in rails rails which can allow an attacker to supply information can be inadvertently leaked fromstrong parameters publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
128,306
| 17,475,859,579
|
IssuesEvent
|
2021-08-08 05:28:19
|
reconness/reconness-frontend
|
https://api.github.com/repos/reconness/reconness-frontend
|
closed
|
Delete pipeline
|
story points: 3 discussion design
|
We need an option to delete a pipeline from the Pipeline list (in Mosaic and List mode) and from a Pipeline Details page.
When the user clicks the Delete option, a confirmation popup will appear.
Confirmation popup will be similar to but the reference will be for a Pipeline

|
1.0
|
Delete pipeline - We need an option to delete a pipeline from the Pipeline list (in Mosaic and List mode) and from a Pipeline Details page.
When the user clicks the Delete option, a confirmation popup will appear.
Confirmation popup will be similar to but the reference will be for a Pipeline

|
non_process
|
delete pipeline we need an option to delete a pipeline from the pipeline list in mosaic and list mode and from a pipeline details page when the user clicks the delete option a confirmation popup will appear confirmation popup will be similar to but the reference will be for a pipeline
| 0
|
6,191
| 9,104,337,581
|
IssuesEvent
|
2019-02-20 17:54:47
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
closed
|
ASDs added into ABM with IPs that are not served by ASD server
|
process_wontfix
|
```
nuvolat@NY2SRV0021:~$ alba dev-extract-config --config ...
{
"ips": [
"172.17.7.33"
],
"asd_id": "EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v",
"node_id": "LWcjrcPQVop59pQq15NKSe53QhypOR1E",
"capacity": 3998830551040,
"home": "/mnt/alba-asd/8EytFWg0OLYUYM9d/EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v",
"log_level": "info",
"rocksdb_block_cache_size": 976276990,
"port": 8607,
"transport": "tcp"
}
````
So the asd only listens on `172.17.7.33` but the ABM tells us
```
0 : { kind =
(Asd ((["172.17.23.33"; "172.17.7.33"], 8607, false, false),
"EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v"));
decommissioned = false; node_id = "LWcjrcPQVop59pQq15NKSe53QhypOR1E";
other =
"{ \"id\" : \"EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v\", \"node_id\" : \"LWcjrcPQVop59pQq15NKSe53QhypOR1E\", \"port\" : 8607, \"used_bytes\" : \"3722378166560\", \"total_bytes\" : \"3998830551040\", \"version\" : \"AsdV1\", \"network_interfaces\":[{\"ipv4_addr\": \"172.17.7.33\"} ]}";
total = 3998830551040L; used = 3723897345607L;
```
The `172.17.23.33` ip address will never be usable. Alba works around this, but constantly nags with `Connection refused`. Something went wrong during installation/configuration
|
1.0
|
ASDs added into ABM with IPs that are not served by ASD server - ```
nuvolat@NY2SRV0021:~$ alba dev-extract-config --config ...
{
"ips": [
"172.17.7.33"
],
"asd_id": "EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v",
"node_id": "LWcjrcPQVop59pQq15NKSe53QhypOR1E",
"capacity": 3998830551040,
"home": "/mnt/alba-asd/8EytFWg0OLYUYM9d/EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v",
"log_level": "info",
"rocksdb_block_cache_size": 976276990,
"port": 8607,
"transport": "tcp"
}
````
So the asd only listens on `172.17.7.33` but the ABM tells us
```
0 : { kind =
(Asd ((["172.17.23.33"; "172.17.7.33"], 8607, false, false),
"EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v"));
decommissioned = false; node_id = "LWcjrcPQVop59pQq15NKSe53QhypOR1E";
other =
"{ \"id\" : \"EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v\", \"node_id\" : \"LWcjrcPQVop59pQq15NKSe53QhypOR1E\", \"port\" : 8607, \"used_bytes\" : \"3722378166560\", \"total_bytes\" : \"3998830551040\", \"version\" : \"AsdV1\", \"network_interfaces\":[{\"ipv4_addr\": \"172.17.7.33\"} ]}";
total = 3998830551040L; used = 3723897345607L;
```
The `172.17.23.33` ip address will never be usable. Alba works around this, but constantly nags with `Connection refused`. Something went wrong during installation/configuration
|
process
|
asds added into abm with ips that are not served by asd server nuvolat alba dev extract config config ips asd id node id capacity home mnt alba asd log level info rocksdb block cache size port transport tcp so the asd only listens on but the abm tells us kind asd false false decommissioned false node id other id node id port used bytes total bytes version network interfaces total used the ip address will never be usable alba works around this but constantly nags with connection refused something went wrong during installation configuration
| 1
|
133,729
| 18,948,595,403
|
IssuesEvent
|
2021-11-18 12:59:02
|
Lambda-IT/parlamentsdienste-components
|
https://api.github.com/repos/Lambda-IT/parlamentsdienste-components
|
closed
|
PD-Datepicker Error
|
bug design
|
If the PD-Datepicker is set to not have user_input the error state is not correctly displayed (no red border)
Not urgent
|
1.0
|
PD-Datepicker Error - If the PD-Datepicker is set to not have user_input the error state is not correctly displayed (no red border)
Not urgent
|
non_process
|
pd datepicker error if the pd datepicker is set to not have user input the error state is not correctly displayed no red border not urgent
| 0
|
18,980
| 24,968,680,959
|
IssuesEvent
|
2022-11-01 21:58:56
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
opened
|
Update OS packages in image
|
enhancement security process
|
### Problem
The container images frequently contains out of date packages from the base image that may contain vulnerabilities. We should keep them up to date even if there's not an updated base image to consume.
### Solution
* Update base image for all mirror node components
* Add an extra command to update all OS packages so they at least are up to date as of each tag
```
RUN apt-get update \
&& DEBIAN_FRONTEND=noninteractive apt-get install -y --no-install-recommends
&& rm -rf /var/lib/apt/lists/*
```
or
```
RUN apk update && apk upgrade && rm -rf /var/cache/apk/*
```
### Alternatives
_No response_
|
1.0
|
Update OS packages in image - ### Problem
The container images frequently contains out of date packages from the base image that may contain vulnerabilities. We should keep them up to date even if there's not an updated base image to consume.
### Solution
* Update base image for all mirror node components
* Add an extra command to update all OS packages so they at least are up to date as of each tag
```
RUN apt-get update \
&& DEBIAN_FRONTEND=noninteractive apt-get install -y --no-install-recommends
&& rm -rf /var/lib/apt/lists/*
```
or
```
RUN apk update && apk upgrade && rm -rf /var/cache/apk/*
```
### Alternatives
_No response_
|
process
|
update os packages in image problem the container images frequently contains out of date packages from the base image that may contain vulnerabilities we should keep them up to date even if there s not an updated base image to consume solution update base image for all mirror node components add an extra command to update all os packages so they at least are up to date as of each tag run apt get update debian frontend noninteractive apt get install y no install recommends rm rf var lib apt lists or run apk update apk upgrade rm rf var cache apk alternatives no response
| 1
|
73,401
| 8,871,796,987
|
IssuesEvent
|
2019-01-11 13:45:49
|
bitpay/copay
|
https://api.github.com/repos/bitpay/copay
|
closed
|
Improve delete wallet view design
|
Design Needed
|
_From @bitjson on October 18, 2016 22:45_
Needs love:

_Copied from original issue: bitpay/bitpay-wallet#562_
|
1.0
|
Improve delete wallet view design - _From @bitjson on October 18, 2016 22:45_
Needs love:

_Copied from original issue: bitpay/bitpay-wallet#562_
|
non_process
|
improve delete wallet view design from bitjson on october needs love copied from original issue bitpay bitpay wallet
| 0
|
176,217
| 28,045,023,192
|
IssuesEvent
|
2023-03-28 21:54:12
|
MozillaFoundation/foundation.mozilla.org
|
https://api.github.com/repos/MozillaFoundation/foundation.mozilla.org
|
closed
|
Roadmap and Prioritize Site Improvement Recommendations
|
design
|
following up issue #6303
The recommendations from the IA refresh has been labeled and categorized on a spreadsheet https://docs.google.com/spreadsheets/d/14XlcxPYT5qJFPnMUsPmDk512lMzTuLx0OJeSwGbmAQM/edit?usp=sharing
These will need be prioritized so 'complexity for implementation' and 'impact' need to be scoped. The action items can then be roadmapped to plan when/where it fits in with the rest of the foundation site work.
|
1.0
|
Roadmap and Prioritize Site Improvement Recommendations - following up issue #6303
The recommendations from the IA refresh has been labeled and categorized on a spreadsheet https://docs.google.com/spreadsheets/d/14XlcxPYT5qJFPnMUsPmDk512lMzTuLx0OJeSwGbmAQM/edit?usp=sharing
These will need be prioritized so 'complexity for implementation' and 'impact' need to be scoped. The action items can then be roadmapped to plan when/where it fits in with the rest of the foundation site work.
|
non_process
|
roadmap and prioritize site improvement recommendations following up issue the recommendations from the ia refresh has been labeled and categorized on a spreadsheet these will need be prioritized so complexity for implementation and impact need to be scoped the action items can then be roadmapped to plan when where it fits in with the rest of the foundation site work
| 0
|
150,577
| 11,967,347,977
|
IssuesEvent
|
2020-04-06 06:27:53
|
ubtue/DatenProbleme
|
https://api.github.com/repos/ubtue/DatenProbleme
|
closed
|
ISSN 1573-0697 Journal of business ethics Mix aus Online First und Standardartikel
|
Zotero_AUTO_RSS blocked ready for testing
|
Die Daten enthalten sowohl Artikel die einem Heft zugeordnet sind, als auch solche ohne Heftzuordnung
|
1.0
|
ISSN 1573-0697 Journal of business ethics Mix aus Online First und Standardartikel - Die Daten enthalten sowohl Artikel die einem Heft zugeordnet sind, als auch solche ohne Heftzuordnung
|
non_process
|
issn journal of business ethics mix aus online first und standardartikel die daten enthalten sowohl artikel die einem heft zugeordnet sind als auch solche ohne heftzuordnung
| 0
|
71,688
| 15,207,902,600
|
IssuesEvent
|
2021-02-17 01:17:40
|
billmcchesney1/hadoop
|
https://api.github.com/repos/billmcchesney1/hadoop
|
opened
|
CVE-2020-36189 (Medium) detected in jackson-databind-2.9.10.1.jar
|
security vulnerability
|
## CVE-2020-36189 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib/jackson-databind-2.9.10.1.jar,hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/target/lib/jackson-databind-2.9.10.1.jar,hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client/target/lib/jackson-databind-2.9.10.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p>
<p>Found in base branch: <b>trunk</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189>CVE-2020-36189</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/commit/33d96c13fe18a2dad01b19ce195548c9acea9da4">https://github.com/FasterXML/jackson-databind/commit/33d96c13fe18a2dad01b19ce195548c9acea9da4</a></p>
<p>Release Date: 2020-12-26</p>
<p>Fix Resolution: Replace or update the following files: SubTypeValidator.java, VERSION-2.x</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.1","isMinimumFixVersionAvailable":false}],"baseBranches":["trunk"],"vulnerabilityIdentifier":"CVE-2020-36189","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-36189 (Medium) detected in jackson-databind-2.9.10.1.jar - ## CVE-2020-36189 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib/jackson-databind-2.9.10.1.jar,hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/target/lib/jackson-databind-2.9.10.1.jar,hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client/target/lib/jackson-databind-2.9.10.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p>
<p>Found in base branch: <b>trunk</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189>CVE-2020-36189</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/commit/33d96c13fe18a2dad01b19ce195548c9acea9da4">https://github.com/FasterXML/jackson-databind/commit/33d96c13fe18a2dad01b19ce195548c9acea9da4</a></p>
<p>Release Date: 2020-12-26</p>
<p>Fix Resolution: Replace or update the following files: SubTypeValidator.java, VERSION-2.x</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.1","isMinimumFixVersionAvailable":false}],"baseBranches":["trunk"],"vulnerabilityIdentifier":"CVE-2020-36189","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice documentstore target lib jackson databind jar hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice hbase hadoop yarn server timelineservice hbase common target lib jackson databind jar hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice hbase hadoop yarn server timelineservice hbase client target lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch trunk vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db drivermanagerconnectionsource publish date url a href cvss score details base score metrics not available suggested fix type change files origin a href release date fix resolution replace or update the following files subtypevalidator java version x isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db drivermanagerconnectionsource vulnerabilityurl
| 0
|
8,415
| 11,580,353,798
|
IssuesEvent
|
2020-02-21 19:57:11
|
usgpo/bill-status
|
https://api.github.com/repos/usgpo/bill-status
|
closed
|
Incorrect Senate Homeland Subcommittee Name
|
file reprocessed
|
In the bulk data source on govinfo.gov, some Senate bills have the incorrect subcommittee name for the Homeland Security and Governmental Affairs Subcommittee on Regulatory Affairs and Federal Management:
https://www.govinfo.gov/bulkdata/BILLSTATUS/116/s/BILLSTATUS-116s1120.xml
```
<subcommittees>
<item>
<systemCode>ssga19</systemCode>
<name>and Federal Management Subcommittee</name>
<activities>
<item>
<name>Hearings by</name>
<date>2019-05-22T18:31:34Z</date>
</item>
</activities>
</item>
</subcommittees>
```
The subcommittee name has been truncated to remove "Regulatory Affairs", and it's incorrect in the subcommittee listing as well as the bill actions. The text of the action has the correct full name of the subcommittee:
```
<item>
<actionDate>2019-05-22</actionDate>
<committees>
<item>
<systemCode>ssga19</systemCode>
<name>and Federal Management Subcommittee</name>
</item>
</committees>
<links/>
<sourceSystem>
<code>0</code>
<name>Senate</name>
</sourceSystem>
<text>
Committee on Homeland Security and Governmental Affairs Subcommittee on Regulatory Affairs and Federal Management. Hearings held.
</text>
<type>Committee</type>
</item>
```
This is affecting the following bills in the 116th Congress:
S. 1419
S. 83
S. 1120
S. 1339
S. 78
S. 1409
S. 1420
|
1.0
|
Incorrect Senate Homeland Subcommittee Name - In the bulk data source on govinfo.gov, some Senate bills have the incorrect subcommittee name for the Homeland Security and Governmental Affairs Subcommittee on Regulatory Affairs and Federal Management:
https://www.govinfo.gov/bulkdata/BILLSTATUS/116/s/BILLSTATUS-116s1120.xml
```
<subcommittees>
<item>
<systemCode>ssga19</systemCode>
<name>and Federal Management Subcommittee</name>
<activities>
<item>
<name>Hearings by</name>
<date>2019-05-22T18:31:34Z</date>
</item>
</activities>
</item>
</subcommittees>
```
The subcommittee name has been truncated to remove "Regulatory Affairs", and it's incorrect in the subcommittee listing as well as the bill actions. The text of the action has the correct full name of the subcommittee:
```
<item>
<actionDate>2019-05-22</actionDate>
<committees>
<item>
<systemCode>ssga19</systemCode>
<name>and Federal Management Subcommittee</name>
</item>
</committees>
<links/>
<sourceSystem>
<code>0</code>
<name>Senate</name>
</sourceSystem>
<text>
Committee on Homeland Security and Governmental Affairs Subcommittee on Regulatory Affairs and Federal Management. Hearings held.
</text>
<type>Committee</type>
</item>
```
This is affecting the following bills in the 116th Congress:
S. 1419
S. 83
S. 1120
S. 1339
S. 78
S. 1409
S. 1420
|
process
|
incorrect senate homeland subcommittee name in the bulk data source on govinfo gov some senate bills have the incorrect subcommittee name for the homeland security and governmental affairs subcommittee on regulatory affairs and federal management and federal management subcommittee hearings by the subcommittee name has been truncated to remove regulatory affairs and it s incorrect in the subcommittee listing as well as the bill actions the text of the action has the correct full name of the subcommittee and federal management subcommittee senate committee on homeland security and governmental affairs subcommittee on regulatory affairs and federal management hearings held committee this is affecting the following bills in the congress s s s s s s s
| 1
|
511,356
| 14,858,727,212
|
IssuesEvent
|
2021-01-18 17:14:31
|
weaveworks/eksctl
|
https://api.github.com/repos/weaveworks/eksctl
|
closed
|
Ability to specify egress/ingress rules for cluster shared security group
|
area/nodegroup kind/feature priority/backlog stale
|
**Why do you want this feature?**
See background in https://github.com/weaveworks/eksctl/issues/1773
Managing security groups outside of `eksctl` just to customize egress/ingress adds inordinate complexity
**What feature/behavior/change do you want?**
> Ideally that would be accomplished through a feature where I can specify those [egress/ingress] rules in Cluster configuration file just like IAM policies.
|
1.0
|
Ability to specify egress/ingress rules for cluster shared security group - **Why do you want this feature?**
See background in https://github.com/weaveworks/eksctl/issues/1773
Managing security groups outside of `eksctl` just to customize egress/ingress adds inordinate complexity
**What feature/behavior/change do you want?**
> Ideally that would be accomplished through a feature where I can specify those [egress/ingress] rules in Cluster configuration file just like IAM policies.
|
non_process
|
ability to specify egress ingress rules for cluster shared security group why do you want this feature see background in managing security groups outside of eksctl just to customize egress ingress adds inordinate complexity what feature behavior change do you want ideally that would be accomplished through a feature where i can specify those rules in cluster configuration file just like iam policies
| 0
|
101,047
| 11,211,974,826
|
IssuesEvent
|
2020-01-06 16:32:47
|
project-koku/koku
|
https://api.github.com/repos/project-koku/koku
|
opened
|
Update nise documentation
|
developer productivity documentation
|
## User Story
As a user of nise (especially now that its on pypi) I want proper documentation so that I can use the tool.
## Assumptions
- We can keep our readme (or some form of it)
- nise does more than when it was first created, we can create a docs folder and add more structured documentation there
- We can cover randomized data
- We can cover static YAML file usage
- We can cover how to really use nise data when docker compose so that data is loaded for a provider
- We can cover how to really use nise data when OpenShift so that data is loaded for a provider
## Acceptance Criteria
- [ ] Documentation is usable by a newcomer to the project and they can successfully generate data and load the data for a provider
|
1.0
|
Update nise documentation - ## User Story
As a user of nise (especially now that its on pypi) I want proper documentation so that I can use the tool.
## Assumptions
- We can keep our readme (or some form of it)
- nise does more than when it was first created, we can create a docs folder and add more structured documentation there
- We can cover randomized data
- We can cover static YAML file usage
- We can cover how to really use nise data when docker compose so that data is loaded for a provider
- We can cover how to really use nise data when OpenShift so that data is loaded for a provider
## Acceptance Criteria
- [ ] Documentation is usable by a newcomer to the project and they can successfully generate data and load the data for a provider
|
non_process
|
update nise documentation user story as a user of nise especially now that its on pypi i want proper documentation so that i can use the tool assumptions we can keep our readme or some form of it nise does more than when it was first created we can create a docs folder and add more structured documentation there we can cover randomized data we can cover static yaml file usage we can cover how to really use nise data when docker compose so that data is loaded for a provider we can cover how to really use nise data when openshift so that data is loaded for a provider acceptance criteria documentation is usable by a newcomer to the project and they can successfully generate data and load the data for a provider
| 0
|
153,339
| 13,503,385,993
|
IssuesEvent
|
2020-09-13 13:19:39
|
geek-engineer-future/podcast
|
https://api.github.com/repos/geek-engineer-future/podcast
|
closed
|
[2020-09-18] Recording Document
|
documentation
|
## テーマ
hoge
## 内容
hoge
## appendix
hoge
# ---
- 最近の気になるトピックをコメントに書きましょう!(技術トピックの場合は「タイトル」+「URL」+「概要」も一緒に書くこと)
- 今週話せそうなテーマがある方はコメントに書きましょう!
|
1.0
|
[2020-09-18] Recording Document -
## テーマ
hoge
## 内容
hoge
## appendix
hoge
# ---
- 最近の気になるトピックをコメントに書きましょう!(技術トピックの場合は「タイトル」+「URL」+「概要」も一緒に書くこと)
- 今週話せそうなテーマがある方はコメントに書きましょう!
|
non_process
|
recording document テーマ hoge 内容 hoge appendix hoge 最近の気になるトピックをコメントに書きましょう!(技術トピックの場合は「タイトル」 「url」 「概要」も一緒に書くこと) 今週話せそうなテーマがある方はコメントに書きましょう!
| 0
|
43,499
| 9,449,984,941
|
IssuesEvent
|
2019-04-16 04:30:01
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
opened
|
Filing issues with multiple batched code locations ("bookmarks")
|
code-nav feature-request
|
As a user who is reviewing code for mistakes, I want to be able to "bookmark" multiple locations of code and then create a batch issue (eg on Jira) with all of the locations. I want anyone viewing the code on Sourcegraph to be able to see the issue I filed, and I want anyone viewing the issue on Jira to be able to go to the location on Sourcegraph.
|
1.0
|
Filing issues with multiple batched code locations ("bookmarks") - As a user who is reviewing code for mistakes, I want to be able to "bookmark" multiple locations of code and then create a batch issue (eg on Jira) with all of the locations. I want anyone viewing the code on Sourcegraph to be able to see the issue I filed, and I want anyone viewing the issue on Jira to be able to go to the location on Sourcegraph.
|
non_process
|
filing issues with multiple batched code locations bookmarks as a user who is reviewing code for mistakes i want to be able to bookmark multiple locations of code and then create a batch issue eg on jira with all of the locations i want anyone viewing the code on sourcegraph to be able to see the issue i filed and i want anyone viewing the issue on jira to be able to go to the location on sourcegraph
| 0
|
899
| 2,594,288,805
|
IssuesEvent
|
2015-02-20 01:31:09
|
BALL-Project/ball
|
https://api.github.com/repos/BALL-Project/ball
|
closed
|
BALLView does not report any progress when exporting VRML file
|
C: VIEW P: minor R: fixed T: defect
|
**Reported by odin on 9 Jul 39586499 03:33 UTC**
Should write out at least its done into the log window. Now, the only way to know the export is finished is to watch the file size growing.
|
1.0
|
BALLView does not report any progress when exporting VRML file - **Reported by odin on 9 Jul 39586499 03:33 UTC**
Should write out at least its done into the log window. Now, the only way to know the export is finished is to watch the file size growing.
|
non_process
|
ballview does not report any progress when exporting vrml file reported by odin on jul utc should write out at least its done into the log window now the only way to know the export is finished is to watch the file size growing
| 0
|
1,753
| 4,445,924,236
|
IssuesEvent
|
2016-08-20 10:26:01
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Issue passing file descriptors in OS X
|
child_process net os x
|
<!--
Thank you for reporting an issue.
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: `v7.0.0-pre`
* **Platform**: `OS X 10.11.5`
* **Subsystem**: child_process
<!-- Enter your issue details below this comment. -->
The `test-cluster-net-send.js` test is sometimes failing when running the test suite in my `OS X` with the following output:
```
=== release test-cluster-net-send ===
Path: parallel/test-cluster-net-send
[32787] master
[32789] worker
assert.js:90
throw new assert.AssertionError({
^
AssertionError: false == true
at process.<anonymous> (/Users/sgimeno/node/node/test/parallel/test-cluster-net-send.js:29:12)
at process.g (events.js:286:16)
at emitOne (events.js:101:20)
at process.emit (events.js:188:7)
Command: out/Release/node /Users/sgimeno/node/node/test/parallel/test-cluster-net-send.js
```
After investigating the issue it *looks* like that the error happens only when the `fd` that passes the worker to the master is closed before it is received in the master process. The following patch, that closes the `fd` only after receiving the `NODE_HANDLE_ACK` message fixes the issue for me.
```diff
diff --git a/lib/internal/child_process.js b/lib/internal/child_process.js
index 789c29e..44a245e 100644
--- a/lib/internal/child_process.js
+++ b/lib/internal/child_process.js
@@ -96,8 +96,8 @@ const handleConversion = {
postSend: function(handle, options) {
// Close the Socket handle after sending it
- if (handle && !options.keepOpen)
- handle.close();
+ //if (handle && !options.keepOpen)
+ // global_handle = handle;
},
got: function(message, handle, emit) {
@@ -465,6 +465,11 @@ function setupChannel(target, channel) {
target.on('internalMessage', function(message, handle) {
// Once acknowledged - continue sending handles.
if (message.cmd === 'NODE_HANDLE_ACK') {
+ if (target._pending_handle) {
+ target._pending_handle.close();
+ target._pending_handle = null;
+ }
+
assert(Array.isArray(target._handleQueue));
var queue = target._handleQueue;
target._handleQueue = null;
@@ -615,8 +620,13 @@ function setupChannel(target, channel) {
req.oncomplete = function() {
if (this.async === true)
control.unref();
- if (obj && obj.postSend)
+ if (obj && obj.postSend) {
obj.postSend(handle, options);
+ if (handle && !options.keepOpen) {
+ assert(!target._pending_handle);
+ target._pending_handle = handle;
+ }
+ }
```
This seems strange to me as my understanding was that closing the file descriptor after sending it was safe (at least in my Linux box I have not been able to reproduce the same issue). Thoughts?
|
1.0
|
Issue passing file descriptors in OS X - <!--
Thank you for reporting an issue.
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: `v7.0.0-pre`
* **Platform**: `OS X 10.11.5`
* **Subsystem**: child_process
<!-- Enter your issue details below this comment. -->
The `test-cluster-net-send.js` test is sometimes failing when running the test suite in my `OS X` with the following output:
```
=== release test-cluster-net-send ===
Path: parallel/test-cluster-net-send
[32787] master
[32789] worker
assert.js:90
throw new assert.AssertionError({
^
AssertionError: false == true
at process.<anonymous> (/Users/sgimeno/node/node/test/parallel/test-cluster-net-send.js:29:12)
at process.g (events.js:286:16)
at emitOne (events.js:101:20)
at process.emit (events.js:188:7)
Command: out/Release/node /Users/sgimeno/node/node/test/parallel/test-cluster-net-send.js
```
After investigating the issue it *looks* like that the error happens only when the `fd` that passes the worker to the master is closed before it is received in the master process. The following patch, that closes the `fd` only after receiving the `NODE_HANDLE_ACK` message fixes the issue for me.
```diff
diff --git a/lib/internal/child_process.js b/lib/internal/child_process.js
index 789c29e..44a245e 100644
--- a/lib/internal/child_process.js
+++ b/lib/internal/child_process.js
@@ -96,8 +96,8 @@ const handleConversion = {
postSend: function(handle, options) {
// Close the Socket handle after sending it
- if (handle && !options.keepOpen)
- handle.close();
+ //if (handle && !options.keepOpen)
+ // global_handle = handle;
},
got: function(message, handle, emit) {
@@ -465,6 +465,11 @@ function setupChannel(target, channel) {
target.on('internalMessage', function(message, handle) {
// Once acknowledged - continue sending handles.
if (message.cmd === 'NODE_HANDLE_ACK') {
+ if (target._pending_handle) {
+ target._pending_handle.close();
+ target._pending_handle = null;
+ }
+
assert(Array.isArray(target._handleQueue));
var queue = target._handleQueue;
target._handleQueue = null;
@@ -615,8 +620,13 @@ function setupChannel(target, channel) {
req.oncomplete = function() {
if (this.async === true)
control.unref();
- if (obj && obj.postSend)
+ if (obj && obj.postSend) {
obj.postSend(handle, options);
+ if (handle && !options.keepOpen) {
+ assert(!target._pending_handle);
+ target._pending_handle = handle;
+ }
+ }
```
This seems strange to me as my understanding was that closing the file descriptor after sending it was safe (at least in my Linux box I have not been able to reproduce the same issue). Thoughts?
|
process
|
issue passing file descriptors in os x thank you for reporting an issue please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version pre platform os x subsystem child process the test cluster net send js test is sometimes failing when running the test suite in my os x with the following output release test cluster net send path parallel test cluster net send master worker assert js throw new assert assertionerror assertionerror false true at process users sgimeno node node test parallel test cluster net send js at process g events js at emitone events js at process emit events js command out release node users sgimeno node node test parallel test cluster net send js after investigating the issue it looks like that the error happens only when the fd that passes the worker to the master is closed before it is received in the master process the following patch that closes the fd only after receiving the node handle ack message fixes the issue for me diff diff git a lib internal child process js b lib internal child process js index a lib internal child process js b lib internal child process js const handleconversion postsend function handle options close the socket handle after sending it if handle options keepopen handle close if handle options keepopen global handle handle got function message handle emit function setupchannel target channel target on internalmessage function message handle once acknowledged continue sending handles if message cmd node handle ack if target pending handle target pending handle close target pending handle null assert array isarray target handlequeue var queue target handlequeue target handlequeue null function setupchannel target channel req oncomplete function if this async true control unref if obj obj postsend if obj obj postsend obj postsend handle options if handle options keepopen assert target pending handle target pending handle handle this seems strange to me as my understanding was that closing the file descriptor after sending it was safe at least in my linux box i have not been able to reproduce the same issue thoughts
| 1
|
816,250
| 30,595,163,827
|
IssuesEvent
|
2023-07-21 21:06:06
|
dotnet/aspnetcore
|
https://api.github.com/repos/dotnet/aspnetcore
|
closed
|
Auto switching for interactive components
|
enhancement area-blazor Priority:1 feature-full-stack-web-ui
|
In scope:
* Use Server by default while loading WebAssembly runtime in the background
* If WebAssembly files are cached, use WebAssembly mode
Out of scope:
* Customizing how the auto mode makes its decision (#48756)
* Shutting down Server circuits eagerly once WebAssembly files are downloaded (instead, just let them shut down naturally as the user navigates around)
|
1.0
|
Auto switching for interactive components - In scope:
* Use Server by default while loading WebAssembly runtime in the background
* If WebAssembly files are cached, use WebAssembly mode
Out of scope:
* Customizing how the auto mode makes its decision (#48756)
* Shutting down Server circuits eagerly once WebAssembly files are downloaded (instead, just let them shut down naturally as the user navigates around)
|
non_process
|
auto switching for interactive components in scope use server by default while loading webassembly runtime in the background if webassembly files are cached use webassembly mode out of scope customizing how the auto mode makes its decision shutting down server circuits eagerly once webassembly files are downloaded instead just let them shut down naturally as the user navigates around
| 0
|
21,562
| 29,922,573,555
|
IssuesEvent
|
2023-06-22 00:38:08
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Remoto] Scrum Master na Coodesh
|
SALVADOR PJ BIG DATA PHP JAVA SCRUM AGILE MOBILE REQUISITOS REMOTO PROCESSOS GITHUB KANBAN CI UMA ANALYTICS ENGENHARIA DE SOFTWARE Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/scrum-master-144141746?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A Loginfo está em busca de Scrum Master para compor seu time!</p>
<p>Como especialistas, escolhemos entregar ao mercado uma solução totalmente mobile, capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes. Somos movidos a tecnologia e temos fome de resultados. Nosso principal objetivo é transformar o mercado entregando logística aprimorada, apoiando a logística e o comércio exterior com a otimização de processos, redução de custos e ganho de produtividade no recebimento, armazenagem e expedição.</p>
<p>Responsabilidades:</p>
<ul>
<li>Coordenar e orientar a equipe de desenvolvimento conforme as prioridades identificadas, tomando decisões e acompanhando a realização dos trabalhos para que sejam executados nos prazos e formas previstos;</li>
<li>Atuar como Scrum Master em times ágeis dentro do framework safe, promover workshops de agilidade e team Building.</li>
</ul>
## Loginfo Tecnologia da Informação LTDA:
<p>Desde 2014, ano de nossa fundação, escolhemos conectar processos operacionais e comunicação. Decidimos tornar cada vez mais digital, ágil e intuitivo o mercado dos setores logísticos, portuários e de armazéns gerais. Somos inovadores em tudo que nos propomos a fazer. Como especialistas, escolhemos entregar ao mercado uma solução totalmente mobile, capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes. Somos movidos a tecnologia e temos fome de resultados. Nosso principal objetivo é transformar o mercado entregando logística aprimorada, apoiando a logística e o comércio exterior com a otimização de processos, redução de custos e ganho de produtividade no recebimento, armazenagem e expedição.</p></p>
## Habilidades:
- PHP
- Agile
- Kanban
## Local:
100% Remoto
## Requisitos:
- Experiência como Scrum Master;
- Graduação Completa em Sistemas de Informação, Engenharia de Software ou Ciências da Computação;
- Certificação em Scrum Master.
## Diferenciais:
- Cursando Pós-graduação/MBA em Desenvolvimento em Soluções Corporativas JAVA, Gestão de TI ou Big Data Analytics.
## Benefícios:
- Horários flexíveis;
- Gympass;
- Alura;
- Seguro de Vida.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Scrum Master na Loginfo Tecnologia da Informação LTDA](https://coodesh.com/jobs/scrum-master-144141746?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
PJ
#### Categoria
Gestão em TI
|
1.0
|
[Remoto] Scrum Master na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/scrum-master-144141746?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A Loginfo está em busca de Scrum Master para compor seu time!</p>
<p>Como especialistas, escolhemos entregar ao mercado uma solução totalmente mobile, capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes. Somos movidos a tecnologia e temos fome de resultados. Nosso principal objetivo é transformar o mercado entregando logística aprimorada, apoiando a logística e o comércio exterior com a otimização de processos, redução de custos e ganho de produtividade no recebimento, armazenagem e expedição.</p>
<p>Responsabilidades:</p>
<ul>
<li>Coordenar e orientar a equipe de desenvolvimento conforme as prioridades identificadas, tomando decisões e acompanhando a realização dos trabalhos para que sejam executados nos prazos e formas previstos;</li>
<li>Atuar como Scrum Master em times ágeis dentro do framework safe, promover workshops de agilidade e team Building.</li>
</ul>
## Loginfo Tecnologia da Informação LTDA:
<p>Desde 2014, ano de nossa fundação, escolhemos conectar processos operacionais e comunicação. Decidimos tornar cada vez mais digital, ágil e intuitivo o mercado dos setores logísticos, portuários e de armazéns gerais. Somos inovadores em tudo que nos propomos a fazer. Como especialistas, escolhemos entregar ao mercado uma solução totalmente mobile, capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes. Somos movidos a tecnologia e temos fome de resultados. Nosso principal objetivo é transformar o mercado entregando logística aprimorada, apoiando a logística e o comércio exterior com a otimização de processos, redução de custos e ganho de produtividade no recebimento, armazenagem e expedição.</p></p>
## Habilidades:
- PHP
- Agile
- Kanban
## Local:
100% Remoto
## Requisitos:
- Experiência como Scrum Master;
- Graduação Completa em Sistemas de Informação, Engenharia de Software ou Ciências da Computação;
- Certificação em Scrum Master.
## Diferenciais:
- Cursando Pós-graduação/MBA em Desenvolvimento em Soluções Corporativas JAVA, Gestão de TI ou Big Data Analytics.
## Benefícios:
- Horários flexíveis;
- Gympass;
- Alura;
- Seguro de Vida.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Scrum Master na Loginfo Tecnologia da Informação LTDA](https://coodesh.com/jobs/scrum-master-144141746?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
PJ
#### Categoria
Gestão em TI
|
process
|
scrum master na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a loginfo está em busca de scrum master para compor seu time como especialistas escolhemos entregar ao mercado uma solução totalmente mobile capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes somos movidos a tecnologia e temos fome de resultados nosso principal objetivo é transformar o mercado entregando logística aprimorada apoiando a logística e o comércio exterior com a otimização de processos redução de custos e ganho de produtividade no recebimento armazenagem e expedição responsabilidades coordenar e orientar a equipe de desenvolvimento conforme as prioridades identificadas tomando decisões e acompanhando a realização dos trabalhos para que sejam executados nos prazos e formas previstos atuar como scrum master em times ágeis dentro do framework safe promover workshops de agilidade e team building loginfo tecnologia da informação ltda desde ano de nossa fundação escolhemos conectar processos operacionais e comunicação decidimos tornar cada vez mais digital ágil e intuitivo o mercado dos setores logísticos portuários e de armazéns gerais somos inovadores em tudo que nos propomos a fazer como especialistas escolhemos entregar ao mercado uma solução totalmente mobile capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes somos movidos a tecnologia e temos fome de resultados nosso principal objetivo é transformar o mercado entregando logística aprimorada apoiando a logística e o comércio exterior com a otimização de processos redução de custos e ganho de produtividade no recebimento armazenagem e expedição habilidades php agile kanban local remoto requisitos experiência como scrum master graduação completa em sistemas de informação engenharia de software ou ciências da computação certificação em scrum master diferenciais cursando pós graduação mba em desenvolvimento em soluções corporativas java gestão de ti ou big data analytics benefícios horários flexíveis gympass alura seguro de vida como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime pj categoria gestão em ti
| 1
|
167,584
| 26,517,867,958
|
IssuesEvent
|
2023-01-18 22:32:32
|
vegaprotocol/frontend-monorepo
|
https://api.github.com/repos/vegaprotocol/frontend-monorepo
|
opened
|
Replace dropdown icon with custom version
|
Trading ux-and-visual-design chore common
|
## The Chore
Currently using a blueprint version of the icon:

vega.xyz uses a different icon for dropdowns:

## Tasks
- [ ] Consider design options (use vega.xyz pattern, or add a new icon for the apps design system)
- [ ] Add to design system
- [ ] Dev
|
1.0
|
Replace dropdown icon with custom version - ## The Chore
Currently using a blueprint version of the icon:

vega.xyz uses a different icon for dropdowns:

## Tasks
- [ ] Consider design options (use vega.xyz pattern, or add a new icon for the apps design system)
- [ ] Add to design system
- [ ] Dev
|
non_process
|
replace dropdown icon with custom version the chore currently using a blueprint version of the icon vega xyz uses a different icon for dropdowns tasks consider design options use vega xyz pattern or add a new icon for the apps design system add to design system dev
| 0
|
17,013
| 22,386,217,717
|
IssuesEvent
|
2022-06-17 00:51:38
|
figlesias221/ProyectoDevOps_Grupo3_IglesiasPerezMolinoloJuan
|
https://api.github.com/repos/figlesias221/ProyectoDevOps_Grupo3_IglesiasPerezMolinoloJuan
|
closed
|
Review FrontEnd Alta de Puntos de carga
|
task process
|
Esfuerzo en HS-P:
Estimado: 1
Real: 1 (@mperezjodal ), 1 (@andrujuanoo )
|
1.0
|
Review FrontEnd Alta de Puntos de carga - Esfuerzo en HS-P:
Estimado: 1
Real: 1 (@mperezjodal ), 1 (@andrujuanoo )
|
process
|
review frontend alta de puntos de carga esfuerzo en hs p estimado real mperezjodal andrujuanoo
| 1
|
74,769
| 20,366,507,576
|
IssuesEvent
|
2022-02-21 06:33:23
|
pandres95/ndi.js
|
https://api.github.com/repos/pandres95/ndi.js
|
closed
|
Error 403 while installing module
|
bug build
|
An error `403` is encountered whilst attempting to install the package from https://ndijs.s3.us-east-2.amazonaws.com/ndi/v1.0.5/Release/linux-x64.tar.gz
If you're wondering, I'm using WSL and tried Node 16.14.0 and Node 17.5.0
|
1.0
|
Error 403 while installing module - An error `403` is encountered whilst attempting to install the package from https://ndijs.s3.us-east-2.amazonaws.com/ndi/v1.0.5/Release/linux-x64.tar.gz
If you're wondering, I'm using WSL and tried Node 16.14.0 and Node 17.5.0
|
non_process
|
error while installing module an error is encountered whilst attempting to install the package from if you re wondering i m using wsl and tried node and node
| 0
|
33,669
| 7,743,208,769
|
IssuesEvent
|
2018-05-29 12:07:13
|
guirisan/arrelaires
|
https://api.github.com/repos/guirisan/arrelaires
|
closed
|
Enviar correu a admins al registrar nova usuària o col·laboració
|
code things
|
Afegides en `app/config/mail.php` una clau `admins` amb les adreces de les admins
### REGISTRE
Creem mail `NewUserNotification` i la vista `/resources/views/emails/admin/new-user-notification.blade.php
En RegistersUsers@register afegim `\Mail::to(config('mail.admins'))->send(new NewUserNotification($user));`
### COL·LABORACIÓ
Creem mail `NewCollaborationNotification` i la vista `/resources/views/emails/admin/new-collaboration-notification.blade.php
En FormController@store_step_one afegim `\Mail::to(config('mail.admins'))->send(new NewCollaborationNotification($persona));`
|
1.0
|
Enviar correu a admins al registrar nova usuària o col·laboració - Afegides en `app/config/mail.php` una clau `admins` amb les adreces de les admins
### REGISTRE
Creem mail `NewUserNotification` i la vista `/resources/views/emails/admin/new-user-notification.blade.php
En RegistersUsers@register afegim `\Mail::to(config('mail.admins'))->send(new NewUserNotification($user));`
### COL·LABORACIÓ
Creem mail `NewCollaborationNotification` i la vista `/resources/views/emails/admin/new-collaboration-notification.blade.php
En FormController@store_step_one afegim `\Mail::to(config('mail.admins'))->send(new NewCollaborationNotification($persona));`
|
non_process
|
enviar correu a admins al registrar nova usuària o col·laboració afegides en app config mail php una clau admins amb les adreces de les admins registre creem mail newusernotification i la vista resources views emails admin new user notification blade php en registersusers register afegim mail to config mail admins send new newusernotification user col·laboració creem mail newcollaborationnotification i la vista resources views emails admin new collaboration notification blade php en formcontroller store step one afegim mail to config mail admins send new newcollaborationnotification persona
| 0
|
140,372
| 5,400,755,063
|
IssuesEvent
|
2017-02-27 22:52:56
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
[k8s.io] Empty [Feature:Empty] does nothing {Kubernetes e2e suite}
|
kind/flake priority/P2
|
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-kubemark-5-gce/1946/
Failed: [k8s.io] Empty [Feature:Empty] does nothing {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:143
Dec 9 12:38:16.486: Couldn't delete ns: "e2e-tests-empty-cgxh2": client: etcd cluster is unavailable or misconfigured (&errors.StatusError{ErrStatus:v1.Status{TypeMeta:v1.TypeMeta{Kind:"Status", APIVersion:"v1"}, ListMeta:v1.ListMeta{SelfLink:"", ResourceVersion:""}, Status:"Failure", Message:"client: etcd cluster is unavailable or misconfigured", Reason:"", Details:(*v1.StatusDetails)(nil), Code:500}})
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:354
```
|
1.0
|
[k8s.io] Empty [Feature:Empty] does nothing {Kubernetes e2e suite} - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-kubemark-5-gce/1946/
Failed: [k8s.io] Empty [Feature:Empty] does nothing {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:143
Dec 9 12:38:16.486: Couldn't delete ns: "e2e-tests-empty-cgxh2": client: etcd cluster is unavailable or misconfigured (&errors.StatusError{ErrStatus:v1.Status{TypeMeta:v1.TypeMeta{Kind:"Status", APIVersion:"v1"}, ListMeta:v1.ListMeta{SelfLink:"", ResourceVersion:""}, Status:"Failure", Message:"client: etcd cluster is unavailable or misconfigured", Reason:"", Details:(*v1.StatusDetails)(nil), Code:500}})
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:354
```
|
non_process
|
empty does nothing kubernetes suite failed empty does nothing kubernetes suite go src io kubernetes output dockerized go src io kubernetes test framework framework go dec couldn t delete ns tests empty client etcd cluster is unavailable or misconfigured errors statuserror errstatus status typemeta typemeta kind status apiversion listmeta listmeta selflink resourceversion status failure message client etcd cluster is unavailable or misconfigured reason details statusdetails nil code go src io kubernetes output dockerized go src io kubernetes test framework framework go
| 0
|
21,757
| 30,276,364,266
|
IssuesEvent
|
2023-07-07 20:04:57
|
gsoft-inc/ov-igloo-ui
|
https://api.github.com/repos/gsoft-inc/ov-igloo-ui
|
closed
|
[Bug]: ActionMenu with disablePortal has a weird behaviour
|
bug in process
|
### Contact Details
_No response_
### What happened?
When I play with the component, it sometimes render at a random place (the top and left css properties are not good) :

Also, as you can hardly see at the beginning of the gif, the hover state is not always triggered (no translation + no cursor). It seems to work well on storybook, so maybe it is related to the previous bug, or maybe it is something on my side, and if you have any ideas, I'm all ears! Edit : it seems to be related to the previous bug since I can see the top and left properties changed.
Edit 2: When testing on my side, it seems that if we change the position: fixed to position: absolute, it works! So maybe it is the fix to do here in the design system!

(+ adding on the .ids-action-menu the property position: relative)
### Component
ActionMenu
### Component Version
1.1.2
### Which browsers are you seeing the problem on?
_No response_
### Mobile Device
_No response_
### Relevant log output
_No response_
|
1.0
|
[Bug]: ActionMenu with disablePortal has a weird behaviour - ### Contact Details
_No response_
### What happened?
When I play with the component, it sometimes render at a random place (the top and left css properties are not good) :

Also, as you can hardly see at the beginning of the gif, the hover state is not always triggered (no translation + no cursor). It seems to work well on storybook, so maybe it is related to the previous bug, or maybe it is something on my side, and if you have any ideas, I'm all ears! Edit : it seems to be related to the previous bug since I can see the top and left properties changed.
Edit 2: When testing on my side, it seems that if we change the position: fixed to position: absolute, it works! So maybe it is the fix to do here in the design system!

(+ adding on the .ids-action-menu the property position: relative)
### Component
ActionMenu
### Component Version
1.1.2
### Which browsers are you seeing the problem on?
_No response_
### Mobile Device
_No response_
### Relevant log output
_No response_
|
process
|
actionmenu with disableportal has a weird behaviour contact details no response what happened when i play with the component it sometimes render at a random place the top and left css properties are not good also as you can hardly see at the beginning of the gif the hover state is not always triggered no translation no cursor it seems to work well on storybook so maybe it is related to the previous bug or maybe it is something on my side and if you have any ideas i m all ears edit it seems to be related to the previous bug since i can see the top and left properties changed edit when testing on my side it seems that if we change the position fixed to position absolute it works so maybe it is the fix to do here in the design system adding on the ids action menu the property position relative component actionmenu component version which browsers are you seeing the problem on no response mobile device no response relevant log output no response
| 1
|
51,259
| 6,506,412,627
|
IssuesEvent
|
2017-08-24 08:56:18
|
thememachine/eleven
|
https://api.github.com/repos/thememachine/eleven
|
opened
|
edit tabs on the frontned
|
Design FEAT
|
# The issue is:
Tabs on the frontend is taking up a lot of space + they are fugly to look at
idea:
a floating edit icon on the left or right would probably solve this issue - with fold out elements n stuff
# Screenshots:

# Where (url)
- /admin/...
# Modules Installed:
(just the important ones)
|
1.0
|
edit tabs on the frontned - # The issue is:
Tabs on the frontend is taking up a lot of space + they are fugly to look at
idea:
a floating edit icon on the left or right would probably solve this issue - with fold out elements n stuff
# Screenshots:

# Where (url)
- /admin/...
# Modules Installed:
(just the important ones)
|
non_process
|
edit tabs on the frontned the issue is tabs on the frontend is taking up a lot of space they are fugly to look at idea a floating edit icon on the left or right would probably solve this issue with fold out elements n stuff screenshots where url admin modules installed just the important ones
| 0
|
20,896
| 3,644,910,503
|
IssuesEvent
|
2016-02-15 12:13:20
|
coder-molok/foowd_alpha2
|
https://api.github.com/repos/coder-molok/foowd_alpha2
|
opened
|
Correzioni minori form registrazione
|
design ELGG
|
Compare un flag "nome inserito non corretto" quando si seleziona il campo "Nome utente", e non va più via anche se il nome è valido

Il campo "sito internet" accetta solo domini. Serve che prenda anche indirizzi di pagine perchè potrebbe essere che i produttori vogliano linkare la propria pagina facebook.

|
1.0
|
Correzioni minori form registrazione - Compare un flag "nome inserito non corretto" quando si seleziona il campo "Nome utente", e non va più via anche se il nome è valido

Il campo "sito internet" accetta solo domini. Serve che prenda anche indirizzi di pagine perchè potrebbe essere che i produttori vogliano linkare la propria pagina facebook.

|
non_process
|
correzioni minori form registrazione compare un flag nome inserito non corretto quando si seleziona il campo nome utente e non va più via anche se il nome è valido il campo sito internet accetta solo domini serve che prenda anche indirizzi di pagine perchè potrebbe essere che i produttori vogliano linkare la propria pagina facebook
| 0
|
690,497
| 23,661,863,445
|
IssuesEvent
|
2022-08-26 16:21:30
|
TheYellowArchitect/doubledamnation
|
https://api.github.com/repos/TheYellowArchitect/doubledamnation
|
opened
|
Secret Ending 1 Alternative Intro Voiceline Swap
|
low priority
|
There are five endings in the game. This is the second hardest.
The camera of the intro is flipped/inverted, to signify the difference, but imo, all voicelines should be swapped ("Abandoned by the gods" should be spoken by P2/Mage, not P1/Warrior)
|
1.0
|
Secret Ending 1 Alternative Intro Voiceline Swap - There are five endings in the game. This is the second hardest.
The camera of the intro is flipped/inverted, to signify the difference, but imo, all voicelines should be swapped ("Abandoned by the gods" should be spoken by P2/Mage, not P1/Warrior)
|
non_process
|
secret ending alternative intro voiceline swap there are five endings in the game this is the second hardest the camera of the intro is flipped inverted to signify the difference but imo all voicelines should be swapped abandoned by the gods should be spoken by mage not warrior
| 0
|
690,608
| 23,665,571,490
|
IssuesEvent
|
2022-08-26 20:28:39
|
googleapis/doc-pipeline
|
https://api.github.com/repos/googleapis/doc-pipeline
|
closed
|
generate: docfx-python-assuredworkloads-0.2.0.tar.gz failed
|
type: bug priority: p1 flakybot: issue
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: d542fc9c171ed5d1f13eca605ad9f516637aaef6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/feb651ce-9f7d-40ce-acc8-a682802ce480), [Sponge](http://sponge2/feb651ce-9f7d-40ce-acc8-a682802ce480)
status: failed
|
1.0
|
generate: docfx-python-assuredworkloads-0.2.0.tar.gz failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: d542fc9c171ed5d1f13eca605ad9f516637aaef6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/feb651ce-9f7d-40ce-acc8-a682802ce480), [Sponge](http://sponge2/feb651ce-9f7d-40ce-acc8-a682802ce480)
status: failed
|
non_process
|
generate docfx python assuredworkloads tar gz failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed
| 0
|
2,051
| 4,861,287,471
|
IssuesEvent
|
2016-11-14 08:09:28
|
triplea-game/triplea
|
https://api.github.com/repos/triplea-game/triplea
|
opened
|
Non-Compatible Changes - Next release branch?
|
Discussion Process
|
I'm concerned/thinking about the next PR that breaks serialization/RMI and would require a major release.
- I would like for us to ask PR submitters to test version compatibility. This means logging in to the lobby and seeing if a game can launch and chat works (this tests java RMI for the most part). Ideally we'd automate or make this no longer required.. Some effort needs to be put into the docs so we have a single /clear place to put these instructions too. I'll see if I can help with that in the next week.
- For changes that we do want, that must break compatibility, I suggest we merge them to a single well known branch. We then merge that to master when we are ready to break version compatiblility. We'll need/want to merge master into this 'next-release' branch frequently, (each each PR merge ideally).
|
1.0
|
Non-Compatible Changes - Next release branch? - I'm concerned/thinking about the next PR that breaks serialization/RMI and would require a major release.
- I would like for us to ask PR submitters to test version compatibility. This means logging in to the lobby and seeing if a game can launch and chat works (this tests java RMI for the most part). Ideally we'd automate or make this no longer required.. Some effort needs to be put into the docs so we have a single /clear place to put these instructions too. I'll see if I can help with that in the next week.
- For changes that we do want, that must break compatibility, I suggest we merge them to a single well known branch. We then merge that to master when we are ready to break version compatiblility. We'll need/want to merge master into this 'next-release' branch frequently, (each each PR merge ideally).
|
process
|
non compatible changes next release branch i m concerned thinking about the next pr that breaks serialization rmi and would require a major release i would like for us to ask pr submitters to test version compatibility this means logging in to the lobby and seeing if a game can launch and chat works this tests java rmi for the most part ideally we d automate or make this no longer required some effort needs to be put into the docs so we have a single clear place to put these instructions too i ll see if i can help with that in the next week for changes that we do want that must break compatibility i suggest we merge them to a single well known branch we then merge that to master when we are ready to break version compatiblility we ll need want to merge master into this next release branch frequently each each pr merge ideally
| 1
|
14,328
| 17,362,458,021
|
IssuesEvent
|
2021-07-29 23:17:53
|
googleapis/google-auth-library-java
|
https://api.github.com/repos/googleapis/google-auth-library-java
|
closed
|
Setup end-to-end integration tests
|
type: process
|
We should be testing this library in different execution environments (GCE, GAE, etc)
|
1.0
|
Setup end-to-end integration tests - We should be testing this library in different execution environments (GCE, GAE, etc)
|
process
|
setup end to end integration tests we should be testing this library in different execution environments gce gae etc
| 1
|
8,977
| 12,093,585,419
|
IssuesEvent
|
2020-04-19 20:14:21
|
Pretronic/PretronicLibraries
|
https://api.github.com/repos/Pretronic/PretronicLibraries
|
closed
|
Copy utility
|
In processing global-utility
|
Create a copy utility based on reflections.
- [x] Copy normal classes
- [x] Deep Copy option
- [x] Annotations for extra handlers
- [x] CopyAble and DeepCopyAble interface (Self copy method)
- [ ] Adapters for special objects
- [ ] Default Adapter for List
- [ ] Default Adapter for Map
- [ ] Default Adapter for Collection
- [ ] Default Adapter for Set
|
1.0
|
Copy utility - Create a copy utility based on reflections.
- [x] Copy normal classes
- [x] Deep Copy option
- [x] Annotations for extra handlers
- [x] CopyAble and DeepCopyAble interface (Self copy method)
- [ ] Adapters for special objects
- [ ] Default Adapter for List
- [ ] Default Adapter for Map
- [ ] Default Adapter for Collection
- [ ] Default Adapter for Set
|
process
|
copy utility create a copy utility based on reflections copy normal classes deep copy option annotations for extra handlers copyable and deepcopyable interface self copy method adapters for special objects default adapter for list default adapter for map default adapter for collection default adapter for set
| 1
|
570,245
| 17,023,071,492
|
IssuesEvent
|
2021-07-03 00:15:28
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
ability to not show landsat in the applet
|
Component: applet Priority: minor Resolution: wontfix Type: enhancement
|
**[Submitted to the original trac issue database at 3.19pm, Thursday, 10th November 2005]**
simplest would be to request the WMS layers seperately and show/hide them?
|
1.0
|
ability to not show landsat in the applet - **[Submitted to the original trac issue database at 3.19pm, Thursday, 10th November 2005]**
simplest would be to request the WMS layers seperately and show/hide them?
|
non_process
|
ability to not show landsat in the applet simplest would be to request the wms layers seperately and show hide them
| 0
|
21,102
| 28,056,452,302
|
IssuesEvent
|
2023-03-29 09:41:12
|
camunda/issues
|
https://api.github.com/repos/camunda/issues
|
opened
|
Support OIDC for Elasticsearch in Self-Managed
|
component:distribution component:operate component:optimize component:tasklist component:zeebe component:zeebe-process-automation public kind:epic potential:8.3
|
### Value Proposition Statement
Secure connections to Elasticsearch using OpenIDConnect in Self-Managed
### User Problem
Today connection between Webapps & Zeebe Elastic Exporter can only use basic authentication.
Nowadays organizations have often policies that forbid using Basic Authentication and that rely on token-based authentication mechanisms. They expect to be able to use SAML and/or OpenIDConnect.
Currently this is not supported and prevents adoption of our Platform for some customers.
### User Stories
I can use OpenIDConnect for connecting to Elasticsearch in Zeebe Elastic Exporter.
I can use OpenIDConnect for connecting to Elasticsearch in Operate.
I can use OpenIDConnect for connecting to Elasticsearch in Tasklist.
I can use OpenIDConnect for connecting to Elasticsearch in Optimize.
### Implementation Notes
From user perspective the best would be if this is just a configuration and not something I have to implement.
|
1.0
|
Support OIDC for Elasticsearch in Self-Managed - ### Value Proposition Statement
Secure connections to Elasticsearch using OpenIDConnect in Self-Managed
### User Problem
Today connection between Webapps & Zeebe Elastic Exporter can only use basic authentication.
Nowadays organizations have often policies that forbid using Basic Authentication and that rely on token-based authentication mechanisms. They expect to be able to use SAML and/or OpenIDConnect.
Currently this is not supported and prevents adoption of our Platform for some customers.
### User Stories
I can use OpenIDConnect for connecting to Elasticsearch in Zeebe Elastic Exporter.
I can use OpenIDConnect for connecting to Elasticsearch in Operate.
I can use OpenIDConnect for connecting to Elasticsearch in Tasklist.
I can use OpenIDConnect for connecting to Elasticsearch in Optimize.
### Implementation Notes
From user perspective the best would be if this is just a configuration and not something I have to implement.
|
process
|
support oidc for elasticsearch in self managed value proposition statement secure connections to elasticsearch using openidconnect in self managed user problem today connection between webapps zeebe elastic exporter can only use basic authentication nowadays organizations have often policies that forbid using basic authentication and that rely on token based authentication mechanisms they expect to be able to use saml and or openidconnect currently this is not supported and prevents adoption of our platform for some customers user stories i can use openidconnect for connecting to elasticsearch in zeebe elastic exporter i can use openidconnect for connecting to elasticsearch in operate i can use openidconnect for connecting to elasticsearch in tasklist i can use openidconnect for connecting to elasticsearch in optimize implementation notes from user perspective the best would be if this is just a configuration and not something i have to implement
| 1
|
296,076
| 22,287,757,962
|
IssuesEvent
|
2022-06-11 22:46:10
|
CR6Community/CR-6-touchscreen
|
https://api.github.com/repos/CR6Community/CR-6-touchscreen
|
closed
|
Strange screen is displayed on Touch Screen firmware flashing.
|
documentation
|
I just ordered an additional touch screen for the CR-6 SE and it was delivered.
While flashing the new CR-6 SE's touch screen using the 61F_RC_290422_v1 released by CR6Community/CR-6-touchscreen, near completion, a white noise-like screen appears.

It looks like the flashing is almost done right before the white noise display.
 
The new touch screen uses the same PCB as the old LCD. However, flashing completion is indicated differently. However, if I wait a few tens of seconds and turn on the CR-6 SE, it operates normally (displaying icons, selecting menus, etc.).
First, to check the kernel version, I checked the kernel version by flashing using an empty DWIN_SET, and at this time, the normal completion screen was displayed. After flashing using the contents of DWIN_SET, a strange completion screen is displayed.

On the touch screen I had it shows the normal completion screen. It is strange because it is a symptom that appears on a new touch screen that uses the same PCB.
|
1.0
|
Strange screen is displayed on Touch Screen firmware flashing. - I just ordered an additional touch screen for the CR-6 SE and it was delivered.
While flashing the new CR-6 SE's touch screen using the 61F_RC_290422_v1 released by CR6Community/CR-6-touchscreen, near completion, a white noise-like screen appears.

It looks like the flashing is almost done right before the white noise display.
 
The new touch screen uses the same PCB as the old LCD. However, flashing completion is indicated differently. However, if I wait a few tens of seconds and turn on the CR-6 SE, it operates normally (displaying icons, selecting menus, etc.).
First, to check the kernel version, I checked the kernel version by flashing using an empty DWIN_SET, and at this time, the normal completion screen was displayed. After flashing using the contents of DWIN_SET, a strange completion screen is displayed.

On the touch screen I had it shows the normal completion screen. It is strange because it is a symptom that appears on a new touch screen that uses the same PCB.
|
non_process
|
strange screen is displayed on touch screen firmware flashing i just ordered an additional touch screen for the cr se and it was delivered while flashing the new cr se s touch screen using the rc released by cr touchscreen near completion a white noise like screen appears it looks like the flashing is almost done right before the white noise display the new touch screen uses the same pcb as the old lcd however flashing completion is indicated differently however if i wait a few tens of seconds and turn on the cr se it operates normally displaying icons selecting menus etc first to check the kernel version i checked the kernel version by flashing using an empty dwin set and at this time the normal completion screen was displayed after flashing using the contents of dwin set a strange completion screen is displayed on the touch screen i had it shows the normal completion screen it is strange because it is a symptom that appears on a new touch screen that uses the same pcb
| 0
|
8,594
| 11,758,834,040
|
IssuesEvent
|
2020-03-13 16:07:00
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
Support current Intel x86/x64 manuals (again)
|
Feature: Processor/x86 Type: Enhancement
|
Version 9.1 of Ghidra has switched to Intel's 325383-60US manual for the instruction set reference. Thank you. This is still available on Intel's website. However they appear to have issued 325383-70US in May of this year. It is what you get if you start here: https://software.intel.com/en-us/articles/intel-sdm. The two manuals are off by 1 page on the last opcode so it's not a big deal. However, someone might want to consider updating the index file sometime.
|
1.0
|
Support current Intel x86/x64 manuals (again) - Version 9.1 of Ghidra has switched to Intel's 325383-60US manual for the instruction set reference. Thank you. This is still available on Intel's website. However they appear to have issued 325383-70US in May of this year. It is what you get if you start here: https://software.intel.com/en-us/articles/intel-sdm. The two manuals are off by 1 page on the last opcode so it's not a big deal. However, someone might want to consider updating the index file sometime.
|
process
|
support current intel manuals again version of ghidra has switched to intel s manual for the instruction set reference thank you this is still available on intel s website however they appear to have issued in may of this year it is what you get if you start here the two manuals are off by page on the last opcode so it s not a big deal however someone might want to consider updating the index file sometime
| 1
|
8,682
| 11,811,459,158
|
IssuesEvent
|
2020-03-19 18:14:41
|
googleapis/java-mediatranslation
|
https://api.github.com/repos/googleapis/java-mediatranslation
|
opened
|
Switch samples/snippets/pom.xml to use libraries-bom
|
type: process
|
We cannot suggest using the libraries-bom here until this library is included in the libraries-bom
|
1.0
|
Switch samples/snippets/pom.xml to use libraries-bom - We cannot suggest using the libraries-bom here until this library is included in the libraries-bom
|
process
|
switch samples snippets pom xml to use libraries bom we cannot suggest using the libraries bom here until this library is included in the libraries bom
| 1
|
348,434
| 10,442,372,910
|
IssuesEvent
|
2019-09-18 12:57:40
|
getkirby/kirby
|
https://api.github.com/repos/getkirby/kirby
|
closed
|
Json::encode escapes unicode entities
|
priority: low 🐌 type: enhancement ✨
|
**To Reproduce**
Steps to reproduce the behavior:
1. Put this in a template:
```php
echo \Kirby\Data\Json::encode('здравей');
```
2. Echoed string is:
```
\u0437\u0434\u0440\u0430\u0432\u0435\u0439
```
**Expected behavior**
The encoded string should be `здравей`
**Kirby Version**
3.2.3
**Additional context**
This can be fixed by simply adding [a flag](https://www.php.net/manual/en/json.constants.php) in `Json::encode()` (tested):
```php
json_encode($data, JSON_UNESCAPED_UNICODE);
```
|
1.0
|
Json::encode escapes unicode entities - **To Reproduce**
Steps to reproduce the behavior:
1. Put this in a template:
```php
echo \Kirby\Data\Json::encode('здравей');
```
2. Echoed string is:
```
\u0437\u0434\u0440\u0430\u0432\u0435\u0439
```
**Expected behavior**
The encoded string should be `здравей`
**Kirby Version**
3.2.3
**Additional context**
This can be fixed by simply adding [a flag](https://www.php.net/manual/en/json.constants.php) in `Json::encode()` (tested):
```php
json_encode($data, JSON_UNESCAPED_UNICODE);
```
|
non_process
|
json encode escapes unicode entities to reproduce steps to reproduce the behavior put this in a template php echo kirby data json encode здравей echoed string is expected behavior the encoded string should be здравей kirby version additional context this can be fixed by simply adding in json encode tested php json encode data json unescaped unicode
| 0
|
11,113
| 13,957,681,439
|
IssuesEvent
|
2020-10-24 08:07:25
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
DE: request regarding XML schema validation of metadata records
|
DE - Germany Geoportal Harvesting process
|
Related to issue #3563 note-47 to note-50
Dear Daniele,
We are planning to change our schema-validation-file for the harvest process (from apiso.xsd version 1.0.0 to apiso.xsd version 1.0.1). To be on the save site and don't "loose" any records during the harvest process for the INSPIRE Geoportal, it is important for us to know, which schema you are use for the INSPIRE Geoportal.
Angelo Quaglia wrote us, that during the harvest process a validation or check for data sets (gmd) and services (srv) is against both schemata http://schemas.opengis.net/iso/19139/20060504/ and http://schemas.opengis.net/iso/19139/20070417/. Referring to this we have a couple of questions
How long you will support both schemata http://schemas.opengis.net/iso/19139/20060504/ and http://schemas.opengis.net/iso/19139/20070417/?
Are you planning to stop supporting the schema http://schemas.opengis.net/iso/19139/20060504/ in the (near) future?
Thanks in advance and best regards,
Anja (on behalf of Coordination Office SDI Germany)
|
1.0
|
DE: request regarding XML schema validation of metadata records - Related to issue #3563 note-47 to note-50
Dear Daniele,
We are planning to change our schema-validation-file for the harvest process (from apiso.xsd version 1.0.0 to apiso.xsd version 1.0.1). To be on the save site and don't "loose" any records during the harvest process for the INSPIRE Geoportal, it is important for us to know, which schema you are use for the INSPIRE Geoportal.
Angelo Quaglia wrote us, that during the harvest process a validation or check for data sets (gmd) and services (srv) is against both schemata http://schemas.opengis.net/iso/19139/20060504/ and http://schemas.opengis.net/iso/19139/20070417/. Referring to this we have a couple of questions
How long you will support both schemata http://schemas.opengis.net/iso/19139/20060504/ and http://schemas.opengis.net/iso/19139/20070417/?
Are you planning to stop supporting the schema http://schemas.opengis.net/iso/19139/20060504/ in the (near) future?
Thanks in advance and best regards,
Anja (on behalf of Coordination Office SDI Germany)
|
process
|
de request regarding xml schema validation of metadata records related to issue note to note dear daniele we are planning to change our schema validation file for the harvest process from apiso xsd version to apiso xsd version to be on the save site and don t quot loose quot any records during the harvest process for the inspire geoportal it is important for us to know which schema you are use for the inspire geoportal angelo quaglia wrote us that during the harvest process a validation or check for data sets gmd and services srv is against both schemata and referring to this we have a couple of questions how long you will support both schemata and are you planning to stop supporting the schema in the near future thanks in advance and best regards anja on behalf of coordination office sdi germany
| 1
|
344,787
| 10,349,640,108
|
IssuesEvent
|
2019-09-04 23:18:11
|
oslc-op/jira-migration-landfill
|
https://api.github.com/repos/oslc-op/jira-migration-landfill
|
closed
|
literal_value of the oslc_where syntax is not well-defined
|
Core: Query Priority: High Xtra: Jira
|
The spec is not clear on how to interpret the literals w/o the xsd data type.
E.g.
The terms boolean and decimal are short forms for typed literals. For example, true is a short form for "true"^xsd:booleancode>, 42 is a short form for "42"xsd:integer and 3.14159 is a short form for "3.14159"^xsd:decimal.
does not specify how I am supposed to know whether 42 is an integer but 3.14 is a decimal (or a single-precision float?), let alone how I am supposed to ensure that ‘true‘ is a boolean True, not a "true" string literal.
---
_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-134 (opened by @berezovskyi; previously assigned to @oslc-bot)_
|
1.0
|
literal_value of the oslc_where syntax is not well-defined - The spec is not clear on how to interpret the literals w/o the xsd data type.
E.g.
The terms boolean and decimal are short forms for typed literals. For example, true is a short form for "true"^xsd:booleancode>, 42 is a short form for "42"xsd:integer and 3.14159 is a short form for "3.14159"^xsd:decimal.
does not specify how I am supposed to know whether 42 is an integer but 3.14 is a decimal (or a single-precision float?), let alone how I am supposed to ensure that ‘true‘ is a boolean True, not a "true" string literal.
---
_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-134 (opened by @berezovskyi; previously assigned to @oslc-bot)_
|
non_process
|
literal value of the oslc where syntax is not well defined the spec is not clear on how to interpret the literals w o the xsd data type e g the terms boolean and decimal are short forms for typed literals for example true is a short form for true xsd booleancode is a short form for xsd integer and is a short form for xsd decimal does not specify how i am supposed to know whether is an integer but is a decimal or a single precision float let alone how i am supposed to ensure that ‘true‘ is a boolean true not a true string literal migrated from opened by berezovskyi previously assigned to oslc bot
| 0
|
11,592
| 14,447,380,996
|
IssuesEvent
|
2020-12-08 03:37:45
|
A01731346/5a
|
https://api.github.com/repos/A01731346/5a
|
closed
|
fill_size_estimating_template
|
process-dashboard
|
- Llenado de template de estimación de líneas de código en process dashboard
- Correr el PROBE Wizard
|
1.0
|
fill_size_estimating_template - - Llenado de template de estimación de líneas de código en process dashboard
- Correr el PROBE Wizard
|
process
|
fill size estimating template llenado de template de estimación de líneas de código en process dashboard correr el probe wizard
| 1
|
301,800
| 26,101,935,283
|
IssuesEvent
|
2022-12-27 08:18:30
|
wazuh/wazuh
|
https://api.github.com/repos/wazuh/wazuh
|
opened
|
Release 4.4.0 - Alpha 2 - E2E UX tests - Wazuh Indexer
|
type/test/manual release test/4.4.0
|
The following issue aims to run the specified test for the current release candidate, report the results, and open new issues for any encountered errors.
## Test information
| | |
|-------------------------|--------------------------------------------|
| **Test name** | Wazuh Indexer |
| **Category** | Installation |
| **Deployment option** | Single Indexer and agent, Multi Server: Step by step |
| **Main release issue** | https://github.com/wazuh/wazuh/issues/15749 |
| **Release candidate #** | Alpha 2 |
| **Previous issue** | https://github.com/wazuh/wazuh/issues/15534 |
## Environment
| **Component** | **OS** | **Installation** | **Type** |
|-|-|-|-|
| Wazuh dashboard | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-dashboard/step-by-step.html) | - |
| Wazuh indexer | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-indexer/step-by-step.html) | Single node |
| Wazuh server | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-server/step-by-step.html) | Multi node |
| Wazuh agent | Amazon Linux 2 | [Installing Wazuh agents](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-agent/index.html) | - |
## Test description
Best effort to test Wazuh indexer package. Think critically and at least review/test:
- Wazuh indexer package specs
- Indexer package size
- Indexer package metadata (description)
- Indexer package digital signature
- Installed files location, size and permissions
- Installation footprint (check that no unnecessary files are modified/broken in the file system. For example that operating system files do keep their right owner/pemissions and that the installer did not break the system.)
- Installed Wazuh indexer service
- Wazuh indexer logs when installed
- Wazuh indexer templates and indices created
- Wazuh indexer configuration (e.g. replicas are expected to be zero by default, how many shards per index,...) Try to compare and find anomalies with the previous Wazuh indexer version using appropiate E2E UX issue. Write down and report as much information as possible to allow comparison between versions using this issue.
- Wazuh indexer cluster node communication and configuration
- Wazuh indexer cluster status
- Wazuh indexer packages uninstallation procedure
## Test report procedure
All test results must have one of the following statuses:
| | |
|---------------------------------|--------------------------------------------|
| :green_circle: | All checks passed. |
| :red_circle: | There is at least one failed result. |
| :yellow_circle: | There is at least one expected failure or skipped test and no failures. |
Any failing test must be properly addressed with a new issue, detailing the error and the possible cause.
An extended report of the test results must be attached as a ZIP or TXT file. Please attach any documents, screenshots, or tables to the issue update with the results. This report can be used by the auditors to dig deeper into any possible failures and details.
## Conclusions
| **Status** | **Test** | **Failure type** | **Notes** |
|----------------|-------------|---------------------|----------------|
| ⚫ | [Environment installation]() | | |
| ⚫ | [Wazuh indexer package information]() | | |
| ⚫ | [Installed files location, size and permissions]() | | |
| ⚫ | [Installation footprint]() | | |
| ⚫ | [Wazuh indexer service]() | | |
| ⚫ | [Wazuh indexer installation logs]() | | Found some warnings that don't affect the performance |
| ⚫ | [Wazuh indexer indices, templates, and shards]() | | |
| ⚫ | [Wazuh indexer cluster status]() | | |
| ⚫ | [Uninstall procedure]() | | |
| ⚫ | [E2E dataflow]() | | |
## Auditors validation
The definition of done for this one is the validation of the conclusions and the test results from all auditors.
- [ ] @wazuh/cicd
- [ ] @davidjiglesias
|
2.0
|
Release 4.4.0 - Alpha 2 - E2E UX tests - Wazuh Indexer - The following issue aims to run the specified test for the current release candidate, report the results, and open new issues for any encountered errors.
## Test information
| | |
|-------------------------|--------------------------------------------|
| **Test name** | Wazuh Indexer |
| **Category** | Installation |
| **Deployment option** | Single Indexer and agent, Multi Server: Step by step |
| **Main release issue** | https://github.com/wazuh/wazuh/issues/15749 |
| **Release candidate #** | Alpha 2 |
| **Previous issue** | https://github.com/wazuh/wazuh/issues/15534 |
## Environment
| **Component** | **OS** | **Installation** | **Type** |
|-|-|-|-|
| Wazuh dashboard | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-dashboard/step-by-step.html) | - |
| Wazuh indexer | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-indexer/step-by-step.html) | Single node |
| Wazuh server | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-server/step-by-step.html) | Multi node |
| Wazuh agent | Amazon Linux 2 | [Installing Wazuh agents](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-agent/index.html) | - |
## Test description
Best effort to test Wazuh indexer package. Think critically and at least review/test:
- Wazuh indexer package specs
- Indexer package size
- Indexer package metadata (description)
- Indexer package digital signature
- Installed files location, size and permissions
- Installation footprint (check that no unnecessary files are modified/broken in the file system. For example that operating system files do keep their right owner/pemissions and that the installer did not break the system.)
- Installed Wazuh indexer service
- Wazuh indexer logs when installed
- Wazuh indexer templates and indices created
- Wazuh indexer configuration (e.g. replicas are expected to be zero by default, how many shards per index,...) Try to compare and find anomalies with the previous Wazuh indexer version using appropiate E2E UX issue. Write down and report as much information as possible to allow comparison between versions using this issue.
- Wazuh indexer cluster node communication and configuration
- Wazuh indexer cluster status
- Wazuh indexer packages uninstallation procedure
## Test report procedure
All test results must have one of the following statuses:
| | |
|---------------------------------|--------------------------------------------|
| :green_circle: | All checks passed. |
| :red_circle: | There is at least one failed result. |
| :yellow_circle: | There is at least one expected failure or skipped test and no failures. |
Any failing test must be properly addressed with a new issue, detailing the error and the possible cause.
An extended report of the test results must be attached as a ZIP or TXT file. Please attach any documents, screenshots, or tables to the issue update with the results. This report can be used by the auditors to dig deeper into any possible failures and details.
## Conclusions
| **Status** | **Test** | **Failure type** | **Notes** |
|----------------|-------------|---------------------|----------------|
| ⚫ | [Environment installation]() | | |
| ⚫ | [Wazuh indexer package information]() | | |
| ⚫ | [Installed files location, size and permissions]() | | |
| ⚫ | [Installation footprint]() | | |
| ⚫ | [Wazuh indexer service]() | | |
| ⚫ | [Wazuh indexer installation logs]() | | Found some warnings that don't affect the performance |
| ⚫ | [Wazuh indexer indices, templates, and shards]() | | |
| ⚫ | [Wazuh indexer cluster status]() | | |
| ⚫ | [Uninstall procedure]() | | |
| ⚫ | [E2E dataflow]() | | |
## Auditors validation
The definition of done for this one is the validation of the conclusions and the test results from all auditors.
- [ ] @wazuh/cicd
- [ ] @davidjiglesias
|
non_process
|
release alpha ux tests wazuh indexer the following issue aims to run the specified test for the current release candidate report the results and open new issues for any encountered errors test information test name wazuh indexer category installation deployment option single indexer and agent multi server step by step main release issue release candidate alpha previous issue environment component os installation type wazuh dashboard amazon linux wazuh indexer amazon linux single node wazuh server amazon linux multi node wazuh agent amazon linux test description best effort to test wazuh indexer package think critically and at least review test wazuh indexer package specs indexer package size indexer package metadata description indexer package digital signature installed files location size and permissions installation footprint check that no unnecessary files are modified broken in the file system for example that operating system files do keep their right owner pemissions and that the installer did not break the system installed wazuh indexer service wazuh indexer logs when installed wazuh indexer templates and indices created wazuh indexer configuration e g replicas are expected to be zero by default how many shards per index try to compare and find anomalies with the previous wazuh indexer version using appropiate ux issue write down and report as much information as possible to allow comparison between versions using this issue wazuh indexer cluster node communication and configuration wazuh indexer cluster status wazuh indexer packages uninstallation procedure test report procedure all test results must have one of the following statuses green circle all checks passed red circle there is at least one failed result yellow circle there is at least one expected failure or skipped test and no failures any failing test must be properly addressed with a new issue detailing the error and the possible cause an extended report of the test results must be attached as a zip or txt file please attach any documents screenshots or tables to the issue update with the results this report can be used by the auditors to dig deeper into any possible failures and details conclusions status test failure type notes ⚫ ⚫ ⚫ ⚫ ⚫ ⚫ found some warnings that don t affect the performance ⚫ ⚫ ⚫ ⚫ auditors validation the definition of done for this one is the validation of the conclusions and the test results from all auditors wazuh cicd davidjiglesias
| 0
|
13,867
| 16,623,137,571
|
IssuesEvent
|
2021-06-03 05:57:28
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Refactor incorrectly converts from int64 to int8
|
Bug Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
When refactoring **_non-fid_** fields within a Geopackage attribute table, the output table's _**fid**_ field is also incorrectly converted from **Integer64** to **Integer8**, without the user requesting. This error only happens when the table contains records; empty tables correctly output the _**fid**_ as **Integer64**. This problem occurs with layer attribute tables as well as standalone attribute tables.
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome -->
1. Refactor a Geopackage table
2. Review the output table's properties
3. Note that the output table's _**fid**_ field is now **Integer8**, when it should be **Integer64**
I am attaching screenshots of input and output table properties.
Input table:

Refactored table:

**QGIS and OS versions**
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
<!--StartFragment-->
<meta http-equiv="Content-Type" content="text/html; charset=utf-8"><style type="text/css">
p, li { white-space: pre-wrap; }
</style>
QGIS version | 3.18.3-Zürich | QGIS code revision | 735cc85be9
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | batch_hillshader-master; changeDataSource; DissolveWithStats; ee_plugin; GlobeBuilder; GroupStats; LAStools; latlontools; mapswipetool_plugin; mmqgis; MultiDistanceBuffer; NNJoin; PointConnector; pointstopaths; qdraw; qfieldsync; Qgis2threejs; qgis2web; QGISSortAndNumber-master; qgis_gee_data_catalog; QRectangleCreator; quick_map_services; refFunctions; SemiAutomaticClassificationPlugin; Serval; shapetools; SpreadsheetLayers; ViewshedAnalysis; db_manager; MetaSearch; processing
<!--EndFragment-->
**Additional context**
I first reported this problem in Stackexchange at: https://gis.stackexchange.com/questions/397984/qgis-refactor-fields-curiosity
<!-- Add any other context about the problem here. -->
|
1.0
|
Refactor incorrectly converts from int64 to int8 -
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
When refactoring **_non-fid_** fields within a Geopackage attribute table, the output table's _**fid**_ field is also incorrectly converted from **Integer64** to **Integer8**, without the user requesting. This error only happens when the table contains records; empty tables correctly output the _**fid**_ as **Integer64**. This problem occurs with layer attribute tables as well as standalone attribute tables.
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome -->
1. Refactor a Geopackage table
2. Review the output table's properties
3. Note that the output table's _**fid**_ field is now **Integer8**, when it should be **Integer64**
I am attaching screenshots of input and output table properties.
Input table:

Refactored table:

**QGIS and OS versions**
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
<!--StartFragment-->
<meta http-equiv="Content-Type" content="text/html; charset=utf-8"><style type="text/css">
p, li { white-space: pre-wrap; }
</style>
QGIS version | 3.18.3-Zürich | QGIS code revision | 735cc85be9
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | batch_hillshader-master; changeDataSource; DissolveWithStats; ee_plugin; GlobeBuilder; GroupStats; LAStools; latlontools; mapswipetool_plugin; mmqgis; MultiDistanceBuffer; NNJoin; PointConnector; pointstopaths; qdraw; qfieldsync; Qgis2threejs; qgis2web; QGISSortAndNumber-master; qgis_gee_data_catalog; QRectangleCreator; quick_map_services; refFunctions; SemiAutomaticClassificationPlugin; Serval; shapetools; SpreadsheetLayers; ViewshedAnalysis; db_manager; MetaSearch; processing
<!--EndFragment-->
**Additional context**
I first reported this problem in Stackexchange at: https://gis.stackexchange.com/questions/397984/qgis-refactor-fields-curiosity
<!-- Add any other context about the problem here. -->
|
process
|
refactor incorrectly converts from to bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug when refactoring non fid fields within a geopackage attribute table the output table s fid field is also incorrectly converted from to without the user requesting this error only happens when the table contains records empty tables correctly output the fid as this problem occurs with layer attribute tables as well as standalone attribute tables how to reproduce refactor a geopackage table review the output table s properties note that the output table s fid field is now when it should be i am attaching screenshots of input and output table properties input table refactored table qgis and os versions about click in the table ctrl a and then ctrl c finally paste here p li white space pre wrap qgis version zürich qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins batch hillshader master changedatasource dissolvewithstats ee plugin globebuilder groupstats lastools latlontools mapswipetool plugin mmqgis multidistancebuffer nnjoin pointconnector pointstopaths qdraw qfieldsync qgissortandnumber master qgis gee data catalog qrectanglecreator quick map services reffunctions semiautomaticclassificationplugin serval shapetools spreadsheetlayers viewshedanalysis db manager metasearch processing additional context i first reported this problem in stackexchange at
| 1
|
278,349
| 21,075,277,652
|
IssuesEvent
|
2022-04-02 03:39:20
|
Shopify/shopify-cli
|
https://api.github.com/repos/Shopify/shopify-cli
|
closed
|
Update the blog regarding CLI
|
area:documentation no-issue-activity
|
Per feedback provided on #762
> Finding the install instructions was a bit of a maze
> The blog leads you to shopify.dev which doesn't actually mention CLI but refers to "tools" down the page - OK, click on that - there is a link to the CLI source on GitHub, with a documentation link in the readme — click — then click again on install instructions - finally.
>
> This was unnecessarily hard. Ideally, a landing page on shopify.dev that explains how to install on various platforms, linked directly from the blog post.
|
1.0
|
Update the blog regarding CLI - Per feedback provided on #762
> Finding the install instructions was a bit of a maze
> The blog leads you to shopify.dev which doesn't actually mention CLI but refers to "tools" down the page - OK, click on that - there is a link to the CLI source on GitHub, with a documentation link in the readme — click — then click again on install instructions - finally.
>
> This was unnecessarily hard. Ideally, a landing page on shopify.dev that explains how to install on various platforms, linked directly from the blog post.
|
non_process
|
update the blog regarding cli per feedback provided on finding the install instructions was a bit of a maze the blog leads you to shopify dev which doesn t actually mention cli but refers to tools down the page ok click on that there is a link to the cli source on github with a documentation link in the readme — click — then click again on install instructions finally this was unnecessarily hard ideally a landing page on shopify dev that explains how to install on various platforms linked directly from the blog post
| 0
|
231,417
| 7,632,155,584
|
IssuesEvent
|
2018-05-05 11:56:28
|
pzahemszky/sudoku
|
https://api.github.com/repos/pzahemszky/sudoku
|
opened
|
Separate primary and secondary peers
|
enhancement good first issue low priority
|
There could be a simple two-level hierarchy between peers: those operations that are more likely to be successful should be placed in front of the secondary operations in the queue of `remove_rearrange`.
In particular, for the triple `(dig, row, col)` and corresponding `row_slice` and `col_slice` objects the below operations should be prioritised before any of the other ones.
- `Operation('digcol', dig, c)`, if `c` is within `col_slice`
- `Operation('digrow', dig, r)`, if `r` is within `row_slice`
- `Operation('digbox', dig, b)`, if `b` intersects with `row` or `col`
|
1.0
|
Separate primary and secondary peers - There could be a simple two-level hierarchy between peers: those operations that are more likely to be successful should be placed in front of the secondary operations in the queue of `remove_rearrange`.
In particular, for the triple `(dig, row, col)` and corresponding `row_slice` and `col_slice` objects the below operations should be prioritised before any of the other ones.
- `Operation('digcol', dig, c)`, if `c` is within `col_slice`
- `Operation('digrow', dig, r)`, if `r` is within `row_slice`
- `Operation('digbox', dig, b)`, if `b` intersects with `row` or `col`
|
non_process
|
separate primary and secondary peers there could be a simple two level hierarchy between peers those operations that are more likely to be successful should be placed in front of the secondary operations in the queue of remove rearrange in particular for the triple dig row col and corresponding row slice and col slice objects the below operations should be prioritised before any of the other ones operation digcol dig c if c is within col slice operation digrow dig r if r is within row slice operation digbox dig b if b intersects with row or col
| 0
|
242,465
| 18,545,121,060
|
IssuesEvent
|
2021-10-21 20:59:20
|
ReznikovRoman/airbnb-clone
|
https://api.github.com/repos/ReznikovRoman/airbnb-clone
|
opened
|
[FEATURE] Improve documentation
|
documentation feature cleanup/optimization
|
**Description**
- Change README.md file: makefile.env file is required (used by pre-commit)
- Add docker-compose.yml file: populate database with fake data, run tests, etc.
- Add guidelines.md file: specify project style guide:
- Code style
- Naming conventions
- Project structure (models, views, services, etc.)
**Possible solution**
-
**Additional context**
-
|
1.0
|
[FEATURE] Improve documentation - **Description**
- Change README.md file: makefile.env file is required (used by pre-commit)
- Add docker-compose.yml file: populate database with fake data, run tests, etc.
- Add guidelines.md file: specify project style guide:
- Code style
- Naming conventions
- Project structure (models, views, services, etc.)
**Possible solution**
-
**Additional context**
-
|
non_process
|
improve documentation description change readme md file makefile env file is required used by pre commit add docker compose yml file populate database with fake data run tests etc add guidelines md file specify project style guide code style naming conventions project structure models views services etc possible solution additional context
| 0
|
22,510
| 31,562,740,230
|
IssuesEvent
|
2023-09-03 12:57:05
|
nextflow-io/nextflow
|
https://api.github.com/repos/nextflow-io/nextflow
|
closed
|
make inputs read-only
|
lang/processes good first issue
|
I've run into several hard-to-trace pipeline bugs caused by tasks inadvertently modifying input files that were staged in as symlinks or hardlinks. It would be good if Nextflow could make such inputs read-only before task execution, and restore their mode afterwards.
|
1.0
|
make inputs read-only - I've run into several hard-to-trace pipeline bugs caused by tasks inadvertently modifying input files that were staged in as symlinks or hardlinks. It would be good if Nextflow could make such inputs read-only before task execution, and restore their mode afterwards.
|
process
|
make inputs read only i ve run into several hard to trace pipeline bugs caused by tasks inadvertently modifying input files that were staged in as symlinks or hardlinks it would be good if nextflow could make such inputs read only before task execution and restore their mode afterwards
| 1
|
1,410
| 3,971,742,637
|
IssuesEvent
|
2016-05-04 13:10:47
|
openvstorage/openvstorage-health-check
|
https://api.github.com/repos/openvstorage/openvstorage-health-check
|
closed
|
Exception in halted volumes when volume is detached/unreachable
|
priority_critical process_duplicate type_bug
|
```
[INFO] Checking vPool 'env1newvpool':
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda>
__call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw)
File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__
return self.run(*args, **kwargs)
File "healthcheck.py", line 85, in check_attended
return HealthCheckController.execute_check()
File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda>
__call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw)
File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__
return self.run(*args, **kwargs)
File "healthcheck.py", line 132, in execute_check
HealthCheckController.check_openvstorage()
File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda>
__call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw)
File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__
return self.run(*args, **kwargs)
File "healthcheck.py", line 185, in check_openvstorage
ovs.check_for_halted_volumes()
File "/opt/OpenvStorage/ovs/extensions/healthcheck/openvstorage/openvstoragecluster_health_check.py", line 947, in check_for_halted_volumes
if int(self.utility.convert_xml_to_json(voldrv_client.info_volume(volume))["boost_serialization"]
volumedriver.storagerouter.storagerouterclient.MaxRedirectsExceededException: volumeInfo
```
|
1.0
|
Exception in halted volumes when volume is detached/unreachable - ```
[INFO] Checking vPool 'env1newvpool':
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda>
__call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw)
File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__
return self.run(*args, **kwargs)
File "healthcheck.py", line 85, in check_attended
return HealthCheckController.execute_check()
File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda>
__call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw)
File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__
return self.run(*args, **kwargs)
File "healthcheck.py", line 132, in execute_check
HealthCheckController.check_openvstorage()
File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda>
__call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw)
File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__
return self.run(*args, **kwargs)
File "healthcheck.py", line 185, in check_openvstorage
ovs.check_for_halted_volumes()
File "/opt/OpenvStorage/ovs/extensions/healthcheck/openvstorage/openvstoragecluster_health_check.py", line 947, in check_for_halted_volumes
if int(self.utility.convert_xml_to_json(voldrv_client.info_volume(volume))["boost_serialization"]
volumedriver.storagerouter.storagerouterclient.MaxRedirectsExceededException: volumeInfo
```
|
process
|
exception in halted volumes when volume is detached unreachable checking vpool traceback most recent call last file line in file usr lib dist packages celery local py line in call lambda x a kw x get current object a kw file usr lib dist packages celery app task py line in call return self run args kwargs file healthcheck py line in check attended return healthcheckcontroller execute check file usr lib dist packages celery local py line in call lambda x a kw x get current object a kw file usr lib dist packages celery app task py line in call return self run args kwargs file healthcheck py line in execute check healthcheckcontroller check openvstorage file usr lib dist packages celery local py line in call lambda x a kw x get current object a kw file usr lib dist packages celery app task py line in call return self run args kwargs file healthcheck py line in check openvstorage ovs check for halted volumes file opt openvstorage ovs extensions healthcheck openvstorage openvstoragecluster health check py line in check for halted volumes if int self utility convert xml to json voldrv client info volume volume volumedriver storagerouter storagerouterclient maxredirectsexceededexception volumeinfo
| 1
|
13,730
| 5,435,976,267
|
IssuesEvent
|
2017-03-05 21:20:26
|
docker/docker
|
https://api.github.com/repos/docker/docker
|
closed
|
Possible docker 1.12.5 problem with Centos 7.3 upgrade
|
area/builder area/distribution area/runtime status/more-info-needed version/1.12
|
**Description**
After running yum upgrade (to Centos 7.3) the Centos distribution docker 1.12.5 is unable to complete a docker build due to an oci runtime error apparently due to a missing /var/lib/docker/devicemapper/mnt/..../rootfs directory.
Reverting back to a VM snapshot from ~ two weeks ago when docker was working normally after the snapshot was made and, with or without removing all containers / images / volumes, repeating the yum upgrade yields the same result upon trying docker build.
**Steps to reproduce the issue:**
1. Have a Dockerfile containing:
"""
From alpine
COPY . /
RUN /docker-build/build
CMD ["/bin/sleep","infinity"]
"""
2. Have a docker-build/build file containing:
"""
#!/bin/sh
set -ex
apk update
apk add bash
apk add coreutils
apk add go
apk add git
apk add make
apk add xz
"""
3. Attempt to build the image with docker build -t golang-build:latest .
**Describe the results you received:**
Docker pulled the base alpine normally, says it copied the build script into the image but threw the oci runtime error while executing the build script:
"""
$ docker build -t golang-build:latest .
Sending build context to Docker daemon 4.608 kB
Step 1 : FROM alpine
Trying to pull repository docker.io/library/alpine ...
latest: Pulling from docker.io/library/alpine
627beaf3eaaf: Pull complete
Digest: sha256:58e1a1bb75db1b5a24a462dd5e2915277ea06438c3f105138f97eb53149673c4
---> 4a415e366388
Step 2 : COPY . /
---> 641277111855
Removing intermediate container f49d4e90b877
Step 3 : RUN /docker-build/build
---> Running in 56e0411caa59
container_linux.go:247: starting container process caused "process_linux.go:359: container init caused \"rootfs_linux.go:54: mounting \\\"/var/lib/docker/containers/56e0411caa59d37303a221bd06cca92810250db517db7e06e92167cef383696f/secrets\\\" to rootfs \\\"/var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs\\\" at \\\"/run/secrets\\\" caused \\\"lstat /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs/run/secrets: not a directory\\\"\""
invalid header field value "oci runtime error: container_linux.go:247: starting container process caused \"process_linux.go:359: container init caused \\\"rootfs_linux.go:54: mounting \\\\\\\"/var/lib/docker/containers/56e0411caa59d37303a221bd06cca92810250db517db7e06e92167cef383696f/secrets\\\\\\\" to rootfs \\\\\\\"/var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs\\\\\\\" at \\\\\\\"/run/secrets\\\\\\\" caused \\\\\\\"lstat /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs/run/secrets: not a directory\\\\\\\"\\\"\"\n"
"""
The /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/ directory does not contain a rootfs directory
$ docker ps -a
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
56e0411caa59 641277111855 "/bin/sh -c /docker-b" About a minute ago Created elegant_bohr
$ docker images -a
REPOSITORY TAG IMAGE ID CREATED SIZE
`<none>` `<none>` 641277111855 2 minutes ago 3.984 MB
docker.io/alpine latest 4a415e366388 43 hours ago 3.984 MB
**Describe the results you expected:**
Previously docker built the image and docker images -a would show docker.io/alpine and the build image with the specified name and tag would appear and there would be nothing in docker ps -a
**Additional information you deem important (e.g. issue happens only occasionally):**
Since the yum upgrade to Centos 7.3 I have been unable to do any docker builds due to the problem above. The name of the container appearing due to the docker build changes on every attempt. The os user is in the docker group to perform all docker operations as a non-privileged user.
The closest existing issue I could find is https://github.com/docker/docker/issues/28109 but that appears to be a /sys/fs/cgroup problem which is not being reported here.
**Output of `docker version`:**
```
Client:
Version: 1.12.5
API version: 1.24
Package version: docker-common-1.12.5-14.el7.centos.x86_64
Go version: go1.7.4
Git commit: 047e51b/1.12.5
Built: Mon Jan 23 15:35:13 2017
OS/Arch: linux/amd64
Server:
Version: 1.12.5
API version: 1.24
Package version: docker-common-1.12.5-14.el7.centos.x86_64
Go version: go1.7.4
Git commit: 047e51b/1.12.5
Built: Mon Jan 23 15:35:13 2017
OS/Arch: linux/amd64
```
**Output of `docker info`:**
```
Containers: 1
Running: 0
Paused: 0
Stopped: 1
Images: 2
Server Version: 1.12.5
Storage Driver: devicemapper
Pool Name: centos-docker--pool
Pool Blocksize: 524.3 kB
Base Device Size: 10.74 GB
Backing Filesystem: xfs
Data file:
Metadata file:
Data Space Used: 60.29 MB
Data Space Total: 27.92 GB
Data Space Available: 27.86 GB
Metadata Space Used: 73.73 kB
Metadata Space Total: 109.1 MB
Metadata Space Available: 109 MB
Thin Pool Minimum Free Space: 2.791 GB
Udev Sync Supported: true
Deferred Removal Enabled: true
Deferred Deletion Enabled: false
Deferred Deleted Device Count: 0
Library Version: 1.02.135-RHEL7 (2016-11-16)
Logging Driver: journald
Cgroup Driver: systemd
Plugins:
Volume: local
Network: null bridge host overlay
Swarm: inactive
Runtimes: docker-runc runc
Default Runtime: docker-runc
Security Options: seccomp selinux
Kernel Version: 3.10.0-514.10.2.el7.x86_64
Operating System: CentOS Linux 7 (Core)
OSType: linux
Architecture: x86_64
Number of Docker Hooks: 2
CPUs: 4
Total Memory: 3.702 GiB
Name: localhost.localdomain
ID: ITFX:BTFW:TKNH:XZ77:LXGN:2SD5:SGPE:ZC3B:COAB:PHMY:O2NR:Y3LZ
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Insecure Registries:
127.0.0.0/8
Registries: docker.io (secure)
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
VirtualBox 5.0.32 running on Windows 7 64bit pro
VM is Centos 7.3
$ rpm --query centos-release
centos-release-7-3.1611.el7.centos.x86_64
Virtualbox guest additions are installed and appear to be working normally
Disks are not full
$ df
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/mapper/centos-root 20961280 5822412 15138868 28% /
devtmpfs 1925036 0 1925036 0% /dev
tmpfs 1940864 80 1940784 1% /dev/shm
tmpfs 1940864 8912 1931952 1% /run
tmpfs 1940864 0 1940864 0% /sys/fs/cgroup
/dev/sda1 508580 314972 193608 62% /boot
/dev/mapper/centos-home 10475520 1051912 9423608 11% /home
tmpfs 388176 4 388172 1% /run/user/992
VmShare 966214652 288862344 677352308 30% /media/sf_VmShare
tmpfs 388176 24 388152 1% /run/user/1000
$ sudo lvdisplay
--- Logical volume ---
LV Path /dev/centos/root
LV Name root
VG Name centos
LV UUID V3PdeT-eF3c-yd2l-rNNp-mSjE-KBIZ-7hlSdF
LV Write Access read/write
LV Creation host, time localhost.localdomain, 2017-02-10 22:41:57 +0000
LV Status available
` # open 1`
LV Size 20.00 GiB
Current LE 5120
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 8192
Block device 253:0
--- Logical volume ---
LV Path /dev/centos/home
LV Name home
VG Name centos
LV UUID cLdgQy-gRpY-vhiU-73hL-t1xP-2dyO-x1z4Fv
LV Write Access read/write
LV Creation host, time localhost.localdomain, 2017-02-10 22:41:59 +0000
LV Status available
`# open 1`
LV Size 10.00 GiB
Current LE 2560
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 8192
Block device 253:2
--- Logical volume ---
LV Path /dev/centos/swap
LV Name swap
VG Name centos
LV UUID PqR0HU-TbW3-Ew4C-gET1-f7aP-OVRB-5xhx5H
LV Write Access read/write
LV Creation host, time localhost.localdomain, 2017-02-10 22:42:00 +0000
LV Status available
`# open 2`
LV Size 4.00 GiB
Current LE 1024
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 8192
Block device 253:1
--- Logical volume ---
LV Name docker-pool
VG Name centos
LV UUID TtSvhh-Yb8c-8yGo-dnHk-tWwm-4ujD-MejUJX
LV Write Access read/write
LV Creation host, time localhost.localdomain, 2017-02-10 23:02:07 +0000
LV Pool metadata docker-pool_tmeta
LV Pool data docker-pool_tdata
LV Status available
`# open 0`
LV Size 26.00 GiB
Allocated pool data 0.22%
Allocated metadata 0.07%
Current LE 6656
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 8192
Block device 253:5
|
1.0
|
Possible docker 1.12.5 problem with Centos 7.3 upgrade - **Description**
After running yum upgrade (to Centos 7.3) the Centos distribution docker 1.12.5 is unable to complete a docker build due to an oci runtime error apparently due to a missing /var/lib/docker/devicemapper/mnt/..../rootfs directory.
Reverting back to a VM snapshot from ~ two weeks ago when docker was working normally after the snapshot was made and, with or without removing all containers / images / volumes, repeating the yum upgrade yields the same result upon trying docker build.
**Steps to reproduce the issue:**
1. Have a Dockerfile containing:
"""
From alpine
COPY . /
RUN /docker-build/build
CMD ["/bin/sleep","infinity"]
"""
2. Have a docker-build/build file containing:
"""
#!/bin/sh
set -ex
apk update
apk add bash
apk add coreutils
apk add go
apk add git
apk add make
apk add xz
"""
3. Attempt to build the image with docker build -t golang-build:latest .
**Describe the results you received:**
Docker pulled the base alpine normally, says it copied the build script into the image but threw the oci runtime error while executing the build script:
"""
$ docker build -t golang-build:latest .
Sending build context to Docker daemon 4.608 kB
Step 1 : FROM alpine
Trying to pull repository docker.io/library/alpine ...
latest: Pulling from docker.io/library/alpine
627beaf3eaaf: Pull complete
Digest: sha256:58e1a1bb75db1b5a24a462dd5e2915277ea06438c3f105138f97eb53149673c4
---> 4a415e366388
Step 2 : COPY . /
---> 641277111855
Removing intermediate container f49d4e90b877
Step 3 : RUN /docker-build/build
---> Running in 56e0411caa59
container_linux.go:247: starting container process caused "process_linux.go:359: container init caused \"rootfs_linux.go:54: mounting \\\"/var/lib/docker/containers/56e0411caa59d37303a221bd06cca92810250db517db7e06e92167cef383696f/secrets\\\" to rootfs \\\"/var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs\\\" at \\\"/run/secrets\\\" caused \\\"lstat /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs/run/secrets: not a directory\\\"\""
invalid header field value "oci runtime error: container_linux.go:247: starting container process caused \"process_linux.go:359: container init caused \\\"rootfs_linux.go:54: mounting \\\\\\\"/var/lib/docker/containers/56e0411caa59d37303a221bd06cca92810250db517db7e06e92167cef383696f/secrets\\\\\\\" to rootfs \\\\\\\"/var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs\\\\\\\" at \\\\\\\"/run/secrets\\\\\\\" caused \\\\\\\"lstat /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs/run/secrets: not a directory\\\\\\\"\\\"\"\n"
"""
The /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/ directory does not contain a rootfs directory
$ docker ps -a
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
56e0411caa59 641277111855 "/bin/sh -c /docker-b" About a minute ago Created elegant_bohr
$ docker images -a
REPOSITORY TAG IMAGE ID CREATED SIZE
`<none>` `<none>` 641277111855 2 minutes ago 3.984 MB
docker.io/alpine latest 4a415e366388 43 hours ago 3.984 MB
**Describe the results you expected:**
Previously docker built the image and docker images -a would show docker.io/alpine and the build image with the specified name and tag would appear and there would be nothing in docker ps -a
**Additional information you deem important (e.g. issue happens only occasionally):**
Since the yum upgrade to Centos 7.3 I have been unable to do any docker builds due to the problem above. The name of the container appearing due to the docker build changes on every attempt. The os user is in the docker group to perform all docker operations as a non-privileged user.
The closest existing issue I could find is https://github.com/docker/docker/issues/28109 but that appears to be a /sys/fs/cgroup problem which is not being reported here.
**Output of `docker version`:**
```
Client:
Version: 1.12.5
API version: 1.24
Package version: docker-common-1.12.5-14.el7.centos.x86_64
Go version: go1.7.4
Git commit: 047e51b/1.12.5
Built: Mon Jan 23 15:35:13 2017
OS/Arch: linux/amd64
Server:
Version: 1.12.5
API version: 1.24
Package version: docker-common-1.12.5-14.el7.centos.x86_64
Go version: go1.7.4
Git commit: 047e51b/1.12.5
Built: Mon Jan 23 15:35:13 2017
OS/Arch: linux/amd64
```
**Output of `docker info`:**
```
Containers: 1
Running: 0
Paused: 0
Stopped: 1
Images: 2
Server Version: 1.12.5
Storage Driver: devicemapper
Pool Name: centos-docker--pool
Pool Blocksize: 524.3 kB
Base Device Size: 10.74 GB
Backing Filesystem: xfs
Data file:
Metadata file:
Data Space Used: 60.29 MB
Data Space Total: 27.92 GB
Data Space Available: 27.86 GB
Metadata Space Used: 73.73 kB
Metadata Space Total: 109.1 MB
Metadata Space Available: 109 MB
Thin Pool Minimum Free Space: 2.791 GB
Udev Sync Supported: true
Deferred Removal Enabled: true
Deferred Deletion Enabled: false
Deferred Deleted Device Count: 0
Library Version: 1.02.135-RHEL7 (2016-11-16)
Logging Driver: journald
Cgroup Driver: systemd
Plugins:
Volume: local
Network: null bridge host overlay
Swarm: inactive
Runtimes: docker-runc runc
Default Runtime: docker-runc
Security Options: seccomp selinux
Kernel Version: 3.10.0-514.10.2.el7.x86_64
Operating System: CentOS Linux 7 (Core)
OSType: linux
Architecture: x86_64
Number of Docker Hooks: 2
CPUs: 4
Total Memory: 3.702 GiB
Name: localhost.localdomain
ID: ITFX:BTFW:TKNH:XZ77:LXGN:2SD5:SGPE:ZC3B:COAB:PHMY:O2NR:Y3LZ
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Insecure Registries:
127.0.0.0/8
Registries: docker.io (secure)
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
VirtualBox 5.0.32 running on Windows 7 64bit pro
VM is Centos 7.3
$ rpm --query centos-release
centos-release-7-3.1611.el7.centos.x86_64
Virtualbox guest additions are installed and appear to be working normally
Disks are not full
$ df
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/mapper/centos-root 20961280 5822412 15138868 28% /
devtmpfs 1925036 0 1925036 0% /dev
tmpfs 1940864 80 1940784 1% /dev/shm
tmpfs 1940864 8912 1931952 1% /run
tmpfs 1940864 0 1940864 0% /sys/fs/cgroup
/dev/sda1 508580 314972 193608 62% /boot
/dev/mapper/centos-home 10475520 1051912 9423608 11% /home
tmpfs 388176 4 388172 1% /run/user/992
VmShare 966214652 288862344 677352308 30% /media/sf_VmShare
tmpfs 388176 24 388152 1% /run/user/1000
$ sudo lvdisplay
--- Logical volume ---
LV Path /dev/centos/root
LV Name root
VG Name centos
LV UUID V3PdeT-eF3c-yd2l-rNNp-mSjE-KBIZ-7hlSdF
LV Write Access read/write
LV Creation host, time localhost.localdomain, 2017-02-10 22:41:57 +0000
LV Status available
` # open 1`
LV Size 20.00 GiB
Current LE 5120
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 8192
Block device 253:0
--- Logical volume ---
LV Path /dev/centos/home
LV Name home
VG Name centos
LV UUID cLdgQy-gRpY-vhiU-73hL-t1xP-2dyO-x1z4Fv
LV Write Access read/write
LV Creation host, time localhost.localdomain, 2017-02-10 22:41:59 +0000
LV Status available
`# open 1`
LV Size 10.00 GiB
Current LE 2560
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 8192
Block device 253:2
--- Logical volume ---
LV Path /dev/centos/swap
LV Name swap
VG Name centos
LV UUID PqR0HU-TbW3-Ew4C-gET1-f7aP-OVRB-5xhx5H
LV Write Access read/write
LV Creation host, time localhost.localdomain, 2017-02-10 22:42:00 +0000
LV Status available
`# open 2`
LV Size 4.00 GiB
Current LE 1024
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 8192
Block device 253:1
--- Logical volume ---
LV Name docker-pool
VG Name centos
LV UUID TtSvhh-Yb8c-8yGo-dnHk-tWwm-4ujD-MejUJX
LV Write Access read/write
LV Creation host, time localhost.localdomain, 2017-02-10 23:02:07 +0000
LV Pool metadata docker-pool_tmeta
LV Pool data docker-pool_tdata
LV Status available
`# open 0`
LV Size 26.00 GiB
Allocated pool data 0.22%
Allocated metadata 0.07%
Current LE 6656
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 8192
Block device 253:5
|
non_process
|
possible docker problem with centos upgrade description after running yum upgrade to centos the centos distribution docker is unable to complete a docker build due to an oci runtime error apparently due to a missing var lib docker devicemapper mnt rootfs directory reverting back to a vm snapshot from two weeks ago when docker was working normally after the snapshot was made and with or without removing all containers images volumes repeating the yum upgrade yields the same result upon trying docker build steps to reproduce the issue have a dockerfile containing from alpine copy run docker build build cmd have a docker build build file containing bin sh set ex apk update apk add bash apk add coreutils apk add go apk add git apk add make apk add xz attempt to build the image with docker build t golang build latest describe the results you received docker pulled the base alpine normally says it copied the build script into the image but threw the oci runtime error while executing the build script docker build t golang build latest sending build context to docker daemon kb step from alpine trying to pull repository docker io library alpine latest pulling from docker io library alpine pull complete digest step copy removing intermediate container step run docker build build running in container linux go starting container process caused process linux go container init caused rootfs linux go mounting var lib docker containers secrets to rootfs var lib docker devicemapper mnt rootfs at run secrets caused lstat var lib docker devicemapper mnt rootfs run secrets not a directory invalid header field value oci runtime error container linux go starting container process caused process linux go container init caused rootfs linux go mounting var lib docker containers secrets to rootfs var lib docker devicemapper mnt rootfs at run secrets caused lstat var lib docker devicemapper mnt rootfs run secrets not a directory n the var lib docker devicemapper mnt directory does not contain a rootfs directory docker ps a container id image command created status ports names bin sh c docker b about a minute ago created elegant bohr docker images a repository tag image id created size minutes ago mb docker io alpine latest hours ago mb describe the results you expected previously docker built the image and docker images a would show docker io alpine and the build image with the specified name and tag would appear and there would be nothing in docker ps a additional information you deem important e g issue happens only occasionally since the yum upgrade to centos i have been unable to do any docker builds due to the problem above the name of the container appearing due to the docker build changes on every attempt the os user is in the docker group to perform all docker operations as a non privileged user the closest existing issue i could find is but that appears to be a sys fs cgroup problem which is not being reported here output of docker version client version api version package version docker common centos go version git commit built mon jan os arch linux server version api version package version docker common centos go version git commit built mon jan os arch linux output of docker info containers running paused stopped images server version storage driver devicemapper pool name centos docker pool pool blocksize kb base device size gb backing filesystem xfs data file metadata file data space used mb data space total gb data space available gb metadata space used kb metadata space total mb metadata space available mb thin pool minimum free space gb udev sync supported true deferred removal enabled true deferred deletion enabled false deferred deleted device count library version logging driver journald cgroup driver systemd plugins volume local network null bridge host overlay swarm inactive runtimes docker runc runc default runtime docker runc security options seccomp selinux kernel version operating system centos linux core ostype linux architecture number of docker hooks cpus total memory gib name localhost localdomain id itfx btfw tknh lxgn sgpe coab phmy docker root dir var lib docker debug mode client false debug mode server false registry insecure registries registries docker io secure additional environment details aws virtualbox physical etc virtualbox running on windows pro vm is centos rpm query centos release centos release centos virtualbox guest additions are installed and appear to be working normally disks are not full df filesystem blocks used available use mounted on dev mapper centos root devtmpfs dev tmpfs dev shm tmpfs run tmpfs sys fs cgroup dev boot dev mapper centos home home tmpfs run user vmshare media sf vmshare tmpfs run user sudo lvdisplay logical volume lv path dev centos root lv name root vg name centos lv uuid rnnp msje kbiz lv write access read write lv creation host time localhost localdomain lv status available open lv size gib current le segments allocation inherit read ahead sectors auto currently set to block device logical volume lv path dev centos home lv name home vg name centos lv uuid cldgqy grpy vhiu lv write access read write lv creation host time localhost localdomain lv status available open lv size gib current le segments allocation inherit read ahead sectors auto currently set to block device logical volume lv path dev centos swap lv name swap vg name centos lv uuid ovrb lv write access read write lv creation host time localhost localdomain lv status available open lv size gib current le segments allocation inherit read ahead sectors auto currently set to block device logical volume lv name docker pool vg name centos lv uuid ttsvhh dnhk twwm mejujx lv write access read write lv creation host time localhost localdomain lv pool metadata docker pool tmeta lv pool data docker pool tdata lv status available open lv size gib allocated pool data allocated metadata current le segments allocation inherit read ahead sectors auto currently set to block device
| 0
|
1,533
| 4,119,268,774
|
IssuesEvent
|
2016-06-08 14:25:27
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
opened
|
NTR: epidermal growth factor receptor signaling pathway involved in heart process
|
BHF-UCL miRNA New term request RNA processes signaling
|
Dear Editors,
I'd like to request a new term re: PMID:23069713
In this study inhibition of epidermal growth factor receptor (EGFR) was shown to promote cardiogenic differentiation of human Mesenchymal Stem Cells (hMSCs) and the transplantation of hMSCs, in which EGFR was inhibited, resulted in enhancement of heart functions. hMSCs transfected with microRNA-133a (miR-133a), which targets EGFR, expressed cardiac-specific markers and luciferase assays demonstrated that EGFR is silenced by miR-133a.
The new requested term: 'epidermal growth factor receptor signaling pathway involved in heart process'.
I am subsequently planning to request regulation terms to create the following annotation:
hsa-miR-133a-3p GO term: negative regulation of epidermal growth factor receptor signaling pathway involved in heart process
AE: *regulates(cardiocyte differentiation), occurs_in(mesenchymal stem cell)
Thank you,
Barbara
GOC:BHF, GOC:BHF_miRNA and GOC:bc
@rachhuntley
@RLovering
*When available I intend to change the ‘regulates’ relation to ‘positively regulates’ (The relation has been requested by @rachhuntley).
|
1.0
|
NTR: epidermal growth factor receptor signaling pathway involved in heart process - Dear Editors,
I'd like to request a new term re: PMID:23069713
In this study inhibition of epidermal growth factor receptor (EGFR) was shown to promote cardiogenic differentiation of human Mesenchymal Stem Cells (hMSCs) and the transplantation of hMSCs, in which EGFR was inhibited, resulted in enhancement of heart functions. hMSCs transfected with microRNA-133a (miR-133a), which targets EGFR, expressed cardiac-specific markers and luciferase assays demonstrated that EGFR is silenced by miR-133a.
The new requested term: 'epidermal growth factor receptor signaling pathway involved in heart process'.
I am subsequently planning to request regulation terms to create the following annotation:
hsa-miR-133a-3p GO term: negative regulation of epidermal growth factor receptor signaling pathway involved in heart process
AE: *regulates(cardiocyte differentiation), occurs_in(mesenchymal stem cell)
Thank you,
Barbara
GOC:BHF, GOC:BHF_miRNA and GOC:bc
@rachhuntley
@RLovering
*When available I intend to change the ‘regulates’ relation to ‘positively regulates’ (The relation has been requested by @rachhuntley).
|
process
|
ntr epidermal growth factor receptor signaling pathway involved in heart process dear editors i d like to request a new term re pmid in this study inhibition of epidermal growth factor receptor egfr was shown to promote cardiogenic differentiation of human mesenchymal stem cells hmscs and the transplantation of hmscs in which egfr was inhibited resulted in enhancement of heart functions hmscs transfected with microrna mir which targets egfr expressed cardiac specific markers and luciferase assays demonstrated that egfr is silenced by mir the new requested term epidermal growth factor receptor signaling pathway involved in heart process i am subsequently planning to request regulation terms to create the following annotation hsa mir go term negative regulation of epidermal growth factor receptor signaling pathway involved in heart process ae regulates cardiocyte differentiation occurs in mesenchymal stem cell thank you barbara goc bhf goc bhf mirna and goc bc rachhuntley rlovering when available i intend to change the ‘regulates’ relation to ‘positively regulates’ the relation has been requested by rachhuntley
| 1
|
473,413
| 13,641,998,933
|
IssuesEvent
|
2020-09-25 14:55:16
|
OpenNebula/one
|
https://api.github.com/repos/OpenNebula/one
|
closed
|
create cli flags for install_gems
|
Category: Packages Community Priority: Low Status: Accepted Type: Backlog
|
---
Author Name: **Rogier Mars** (Rogier Mars)
Original Redmine Issue: 4980, https://dev.opennebula.org/issues/4980
Original Date: 2017-01-12
---
Hi,
Would it be possible to create flags to set the OS and make the script non-interactive? This would make it easier to run the script from configmanagement like ansible. Now we have to wrap it in expect, and that causes other issues.
For example:
```
install_gems --os=centos --silent --force
install_gems --os=ubuntu --silent --force
```
|
1.0
|
create cli flags for install_gems - ---
Author Name: **Rogier Mars** (Rogier Mars)
Original Redmine Issue: 4980, https://dev.opennebula.org/issues/4980
Original Date: 2017-01-12
---
Hi,
Would it be possible to create flags to set the OS and make the script non-interactive? This would make it easier to run the script from configmanagement like ansible. Now we have to wrap it in expect, and that causes other issues.
For example:
```
install_gems --os=centos --silent --force
install_gems --os=ubuntu --silent --force
```
|
non_process
|
create cli flags for install gems author name rogier mars rogier mars original redmine issue original date hi would it be possible to create flags to set the os and make the script non interactive this would make it easier to run the script from configmanagement like ansible now we have to wrap it in expect and that causes other issues for example install gems os centos silent force install gems os ubuntu silent force
| 0
|
13,815
| 16,577,454,616
|
IssuesEvent
|
2021-05-31 07:19:30
|
paul-buerkner/brms
|
https://api.github.com/repos/paul-buerkner/brms
|
closed
|
loo moment_match crashes R if save_all_pars not specified
|
bug post-processing
|
As mentioned in #1126 loo with moment matching doesn't work without `save_all_pars=save_pars(all = TRUE)`. But it seems worse than just not having an appropriate warning message, as it can crash R if tried without the parameter.
Example:
With `save_all_pars`
```{r}
library(brms)
m <- brm(yield ~ N*P*K, npk, save_all_pars = save_pars(all = TRUE))
loo(m, moment_match = TRUE)
```
Yields:
```{r}
Computed from 4000 by 24 log-likelihood matrix
Estimate SE
elpd_loo -81.2 3.1
p_loo 7.5 1.3
looic 162.4 6.3
------
Monte Carlo SE of elpd_loo is 0.1.
Pareto k diagnostic values:
Count Pct. Min. n_eff
(-Inf, 0.5] (good) 20 83.3% 377
(0.5, 0.7] (ok) 4 16.7% 538
(0.7, 1] (bad) 0 0.0% <NA>
(1, Inf) (very bad) 0 0.0% <NA>
All Pareto k estimates are ok (k < 0.7).
See help('pareto-k-diagnostic') for details.
```
Whereas,
```{r}
library(brms)
m <- brm(yield ~ N*P*K, npk)
loo(m, moment_match = TRUE)
```
Yields:
```
*** caught segfault ***
address (nil), cause 'memory not mapped'
Traceback:
1: .External(list(name = "CppMethod__invoke_notvoid", address = <pointer: 0x564515d9c940>, dll = list(name = "Rcpp", path = "/lib/R/library/Rcpp/libs/Rcpp.so", dynamicLookup = TRUE, handle = <pointer: 0x564515fcbf70>, info = <pointer: 0x564513e69b30>), numParameters = -1L), <\
pointer: 0x564524ed9590>, <pointer: 0x56451dc4f180>, .pointer, ...)
2: object@.MISC$stan_fit_instance$unconstrain_pars(pars)
3: .local(object, ...)
4: rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton))
5: rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton))
6: FUN(newX[, i], ...)
7: apply(pars, 1, FUN = function(theta) { rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton))})
8: unconstrain_pars_stanfit(x$fit, pars = pars, ...)
9: unconstrain_pars(x, pars = pars, ...)
10: loo::loo_moment_match.default(x, loo = loo, post_draws = as.matrix, log_lik_i = .log_lik_i, unconstrain_pars = .unconstrain_pars, log_prob_upars = .log_prob_upars, log_lik_i_upars = .log_lik_i_upars, k_threshold = k_threshold, newdata = newdata, resp = resp, ...)
11: doTryCatch(return(expr), name, parentenv, handler)
12: tryCatchOne(expr, names, parentenv, handlers[[1L]])
13: tryCatchList(expr, classes, parentenv, handlers)
14: tryCatch(expr, error = function(e) { call <- conditionCall(e) if (!is.null(call)) { if (identical(call[[1L]], quote(doTryCatch))) call <- sys.call(-4L) dcall <- deparse(call)[1L] prefix <- paste("Error in", dcall, ": ") LONG <- 75L sm <- strsplit(conditionMessage(e), "\n")[[1L]] \
w <- 14L + nchar(dcall, type = "w") + nchar(sm[1L], type = "w") if (is.na(w)) w <- 14L + nchar(dcall, type = "b") + nchar(sm[1L], type = "b") if (w > LONG) prefix <- paste0(prefix, "\n ") } else prefix <- "Error : " msg <- paste0(prefix, conditionMessage(e), "\n") \
.Internal(seterrmessage(msg[1L])) if (!silent && isTRUE(getOption("show.error.messages"))) { cat(msg, file = outFile) .Internal(printDeferredWarnings()) } invisible(structure(msg, class = "try-error", condition = e))})
15: try(loo::loo_moment_match.default(x, loo = loo, post_draws = as.matrix, log_lik_i = .log_lik_i, unconstrain_pars = .unconstrain_pars, log_prob_upars = .log_prob_upars, log_lik_i_upars = .log_lik_i_upars, k_threshold = k_threshold, newdata = newdata, resp = resp, ...))
16: loo_moment_match.brmsfit(x = .x1, loo = .x2, newdata = .x3, resp = .x4, k_threshold = .x5, check = .x6)
17: loo_moment_match(x = .x1, loo = .x2, newdata = .x3, resp = .x4, k_threshold = .x5, check = .x6)
18: eval(expr, envir, ...)
19: eval(expr, envir, ...)
20: eval2(call, envir = args, enclos = envir)
21: do_call("loo_moment_match", moment_match_args)
22: .loo(x = .x1, newdata = .x2, resp = .x3, model_name = .x4, pointwise = .x5, k_threshold = .x6, save_psis = .x7, moment_match = .x8, reloo = .x9, moment_match_args = .x10, reloo_args = .x11)
23: eval(expr, envir, ...)
24: eval(expr, envir, ...)
25: eval2(call, envir = args, enclos = envir)
26: do_call(paste0(".", criterion), args)
27: .fun(criterion = .x1, pointwise = .x2, resp = .x3, k_threshold = .x4, save_psis = .x5, moment_match = .x6, reloo = .x7, moment_match_args = .x8, reloo_args = .x9, x = .x10, model_name = .x11, use_stored = .x12)
28: eval(expr, envir, ...)
29: eval(expr, envir, ...)
30: eval2(call, envir = args, enclos = envir)
31: do_call(compute_loo, args)
32: .fun(models = .x1, criterion = .x2, pointwise = .x3, compare = .x4, resp = .x5, k_threshold = .x6, save_psis = .x7, moment_match = .x8, reloo = .x9, moment_match_args = .x10, reloo_args = .x11)
33: eval(expr, envir, ...)
34: eval(expr, envir, ...)
35: eval2(call, envir = args, enclos = envir)
36: do_call(compute_loolist, args)
37: loo.brmsfit(m, moment_match = TRUE)
38: loo(m, moment_match = TRUE)
An irrecoverable exception occurred. R is aborting now ...
```
Session Info:
```{r}
R version 4.0.5 (2021-03-31)
Platform: x86_64-pc-linux-gnu (64-bit)
Running under: Ubuntu 18.04.5 LTS
Matrix products: default
BLAS: /lib/R/lib/libRblas.so
LAPACK: /lib/R/lib/libRlapack.so
locale:
[1] LC_CTYPE=en_GB.utf8 LC_NUMERIC=C
[3] LC_TIME=en_DK.utf8 LC_COLLATE=en_GB.utf8
[5] LC_MONETARY=en_GB.utf8 LC_MESSAGES=en_GB.utf8
[7] LC_PAPER=fi_FI.utf8 LC_NAME=C
[9] LC_ADDRESS=C LC_TELEPHONE=C
[11] LC_MEASUREMENT=en_GB.utf8 LC_IDENTIFICATION=C
attached base packages:
[1] stats graphics grDevices utils datasets methods base
other attached packages:
[1] brms_2.15.5 Rcpp_1.0.6
loaded via a namespace (and not attached):
[1] nlme_3.1-152 matrixStats_0.58.0 xts_0.12.1
[4] threejs_0.3.3 rstan_2.26.1 backports_1.2.1
[7] tools_4.0.5 utf8_1.2.1 R6_2.5.0
[10] DT_0.18 DBI_1.1.1 mgcv_1.8-35
[13] projpred_2.0.2 colorspace_2.0-1 tidyselect_1.1.1
[16] gridExtra_2.3 prettyunits_1.1.1 processx_3.5.2
[19] Brobdingnag_1.2-6 emmeans_1.6.0 curl_4.3.1
[22] compiler_4.0.5 cli_2.5.0 shinyjs_2.0.0
[25] colourpicker_1.1.0 scales_1.1.1 dygraphs_1.1.1.6
[28] mvtnorm_1.1-1 ggridges_0.5.3 callr_3.7.0
[31] stringr_1.4.0 digest_0.6.27 StanHeaders_2.26.1
[34] minqa_1.2.4 base64enc_0.1-3 pkgconfig_2.0.3
[37] htmltools_0.5.1.1 lme4_1.1-27 fastmap_1.1.0
[40] htmlwidgets_1.5.3 rlang_0.4.11 shiny_1.6.0
[43] generics_0.1.0 zoo_1.8-9 jsonlite_1.7.2
[46] crosstalk_1.1.1 gtools_3.8.2 dplyr_1.0.6
[49] inline_0.3.18 magrittr_2.0.1 loo_2.4.1
[52] bayesplot_1.8.0 Matrix_1.3-2 munsell_0.5.0
[55] fansi_0.4.2 abind_1.4-5 lifecycle_1.0.0
[58] stringi_1.6.2 MASS_7.3-53.1 pkgbuild_1.2.0
[61] plyr_1.8.6 grid_4.0.5 parallel_4.0.5
[64] promises_1.2.0.1 crayon_1.4.1 miniUI_0.1.1.1
[67] lattice_0.20-41 splines_4.0.5 ps_1.6.0
[70] pillar_1.6.1 igraph_1.2.6 boot_1.3-27
[73] estimability_1.3 markdown_1.1 shinystan_2.5.0
[76] codetools_0.2-18 reshape2_1.4.4 stats4_4.0.5
[79] rstantools_2.1.1 glue_1.4.2 V8_3.4.2
[82] RcppParallel_5.1.4 vctrs_0.3.8 nloptr_1.2.2.2
[85] httpuv_1.6.1 gtable_0.3.0 purrr_0.3.4
[88] assertthat_0.2.1 ggplot2_3.3.3 mime_0.10
[91] xtable_1.8-4 coda_0.19-4 later_1.2.0
[94] rsconnect_0.8.17 tibble_3.1.2 shinythemes_1.2.0
[97] gamm4_0.2-6 ellipsis_0.3.2 bridgesampling_1.1-2
```
|
1.0
|
loo moment_match crashes R if save_all_pars not specified - As mentioned in #1126 loo with moment matching doesn't work without `save_all_pars=save_pars(all = TRUE)`. But it seems worse than just not having an appropriate warning message, as it can crash R if tried without the parameter.
Example:
With `save_all_pars`
```{r}
library(brms)
m <- brm(yield ~ N*P*K, npk, save_all_pars = save_pars(all = TRUE))
loo(m, moment_match = TRUE)
```
Yields:
```{r}
Computed from 4000 by 24 log-likelihood matrix
Estimate SE
elpd_loo -81.2 3.1
p_loo 7.5 1.3
looic 162.4 6.3
------
Monte Carlo SE of elpd_loo is 0.1.
Pareto k diagnostic values:
Count Pct. Min. n_eff
(-Inf, 0.5] (good) 20 83.3% 377
(0.5, 0.7] (ok) 4 16.7% 538
(0.7, 1] (bad) 0 0.0% <NA>
(1, Inf) (very bad) 0 0.0% <NA>
All Pareto k estimates are ok (k < 0.7).
See help('pareto-k-diagnostic') for details.
```
Whereas,
```{r}
library(brms)
m <- brm(yield ~ N*P*K, npk)
loo(m, moment_match = TRUE)
```
Yields:
```
*** caught segfault ***
address (nil), cause 'memory not mapped'
Traceback:
1: .External(list(name = "CppMethod__invoke_notvoid", address = <pointer: 0x564515d9c940>, dll = list(name = "Rcpp", path = "/lib/R/library/Rcpp/libs/Rcpp.so", dynamicLookup = TRUE, handle = <pointer: 0x564515fcbf70>, info = <pointer: 0x564513e69b30>), numParameters = -1L), <\
pointer: 0x564524ed9590>, <pointer: 0x56451dc4f180>, .pointer, ...)
2: object@.MISC$stan_fit_instance$unconstrain_pars(pars)
3: .local(object, ...)
4: rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton))
5: rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton))
6: FUN(newX[, i], ...)
7: apply(pars, 1, FUN = function(theta) { rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton))})
8: unconstrain_pars_stanfit(x$fit, pars = pars, ...)
9: unconstrain_pars(x, pars = pars, ...)
10: loo::loo_moment_match.default(x, loo = loo, post_draws = as.matrix, log_lik_i = .log_lik_i, unconstrain_pars = .unconstrain_pars, log_prob_upars = .log_prob_upars, log_lik_i_upars = .log_lik_i_upars, k_threshold = k_threshold, newdata = newdata, resp = resp, ...)
11: doTryCatch(return(expr), name, parentenv, handler)
12: tryCatchOne(expr, names, parentenv, handlers[[1L]])
13: tryCatchList(expr, classes, parentenv, handlers)
14: tryCatch(expr, error = function(e) { call <- conditionCall(e) if (!is.null(call)) { if (identical(call[[1L]], quote(doTryCatch))) call <- sys.call(-4L) dcall <- deparse(call)[1L] prefix <- paste("Error in", dcall, ": ") LONG <- 75L sm <- strsplit(conditionMessage(e), "\n")[[1L]] \
w <- 14L + nchar(dcall, type = "w") + nchar(sm[1L], type = "w") if (is.na(w)) w <- 14L + nchar(dcall, type = "b") + nchar(sm[1L], type = "b") if (w > LONG) prefix <- paste0(prefix, "\n ") } else prefix <- "Error : " msg <- paste0(prefix, conditionMessage(e), "\n") \
.Internal(seterrmessage(msg[1L])) if (!silent && isTRUE(getOption("show.error.messages"))) { cat(msg, file = outFile) .Internal(printDeferredWarnings()) } invisible(structure(msg, class = "try-error", condition = e))})
15: try(loo::loo_moment_match.default(x, loo = loo, post_draws = as.matrix, log_lik_i = .log_lik_i, unconstrain_pars = .unconstrain_pars, log_prob_upars = .log_prob_upars, log_lik_i_upars = .log_lik_i_upars, k_threshold = k_threshold, newdata = newdata, resp = resp, ...))
16: loo_moment_match.brmsfit(x = .x1, loo = .x2, newdata = .x3, resp = .x4, k_threshold = .x5, check = .x6)
17: loo_moment_match(x = .x1, loo = .x2, newdata = .x3, resp = .x4, k_threshold = .x5, check = .x6)
18: eval(expr, envir, ...)
19: eval(expr, envir, ...)
20: eval2(call, envir = args, enclos = envir)
21: do_call("loo_moment_match", moment_match_args)
22: .loo(x = .x1, newdata = .x2, resp = .x3, model_name = .x4, pointwise = .x5, k_threshold = .x6, save_psis = .x7, moment_match = .x8, reloo = .x9, moment_match_args = .x10, reloo_args = .x11)
23: eval(expr, envir, ...)
24: eval(expr, envir, ...)
25: eval2(call, envir = args, enclos = envir)
26: do_call(paste0(".", criterion), args)
27: .fun(criterion = .x1, pointwise = .x2, resp = .x3, k_threshold = .x4, save_psis = .x5, moment_match = .x6, reloo = .x7, moment_match_args = .x8, reloo_args = .x9, x = .x10, model_name = .x11, use_stored = .x12)
28: eval(expr, envir, ...)
29: eval(expr, envir, ...)
30: eval2(call, envir = args, enclos = envir)
31: do_call(compute_loo, args)
32: .fun(models = .x1, criterion = .x2, pointwise = .x3, compare = .x4, resp = .x5, k_threshold = .x6, save_psis = .x7, moment_match = .x8, reloo = .x9, moment_match_args = .x10, reloo_args = .x11)
33: eval(expr, envir, ...)
34: eval(expr, envir, ...)
35: eval2(call, envir = args, enclos = envir)
36: do_call(compute_loolist, args)
37: loo.brmsfit(m, moment_match = TRUE)
38: loo(m, moment_match = TRUE)
An irrecoverable exception occurred. R is aborting now ...
```
Session Info:
```{r}
R version 4.0.5 (2021-03-31)
Platform: x86_64-pc-linux-gnu (64-bit)
Running under: Ubuntu 18.04.5 LTS
Matrix products: default
BLAS: /lib/R/lib/libRblas.so
LAPACK: /lib/R/lib/libRlapack.so
locale:
[1] LC_CTYPE=en_GB.utf8 LC_NUMERIC=C
[3] LC_TIME=en_DK.utf8 LC_COLLATE=en_GB.utf8
[5] LC_MONETARY=en_GB.utf8 LC_MESSAGES=en_GB.utf8
[7] LC_PAPER=fi_FI.utf8 LC_NAME=C
[9] LC_ADDRESS=C LC_TELEPHONE=C
[11] LC_MEASUREMENT=en_GB.utf8 LC_IDENTIFICATION=C
attached base packages:
[1] stats graphics grDevices utils datasets methods base
other attached packages:
[1] brms_2.15.5 Rcpp_1.0.6
loaded via a namespace (and not attached):
[1] nlme_3.1-152 matrixStats_0.58.0 xts_0.12.1
[4] threejs_0.3.3 rstan_2.26.1 backports_1.2.1
[7] tools_4.0.5 utf8_1.2.1 R6_2.5.0
[10] DT_0.18 DBI_1.1.1 mgcv_1.8-35
[13] projpred_2.0.2 colorspace_2.0-1 tidyselect_1.1.1
[16] gridExtra_2.3 prettyunits_1.1.1 processx_3.5.2
[19] Brobdingnag_1.2-6 emmeans_1.6.0 curl_4.3.1
[22] compiler_4.0.5 cli_2.5.0 shinyjs_2.0.0
[25] colourpicker_1.1.0 scales_1.1.1 dygraphs_1.1.1.6
[28] mvtnorm_1.1-1 ggridges_0.5.3 callr_3.7.0
[31] stringr_1.4.0 digest_0.6.27 StanHeaders_2.26.1
[34] minqa_1.2.4 base64enc_0.1-3 pkgconfig_2.0.3
[37] htmltools_0.5.1.1 lme4_1.1-27 fastmap_1.1.0
[40] htmlwidgets_1.5.3 rlang_0.4.11 shiny_1.6.0
[43] generics_0.1.0 zoo_1.8-9 jsonlite_1.7.2
[46] crosstalk_1.1.1 gtools_3.8.2 dplyr_1.0.6
[49] inline_0.3.18 magrittr_2.0.1 loo_2.4.1
[52] bayesplot_1.8.0 Matrix_1.3-2 munsell_0.5.0
[55] fansi_0.4.2 abind_1.4-5 lifecycle_1.0.0
[58] stringi_1.6.2 MASS_7.3-53.1 pkgbuild_1.2.0
[61] plyr_1.8.6 grid_4.0.5 parallel_4.0.5
[64] promises_1.2.0.1 crayon_1.4.1 miniUI_0.1.1.1
[67] lattice_0.20-41 splines_4.0.5 ps_1.6.0
[70] pillar_1.6.1 igraph_1.2.6 boot_1.3-27
[73] estimability_1.3 markdown_1.1 shinystan_2.5.0
[76] codetools_0.2-18 reshape2_1.4.4 stats4_4.0.5
[79] rstantools_2.1.1 glue_1.4.2 V8_3.4.2
[82] RcppParallel_5.1.4 vctrs_0.3.8 nloptr_1.2.2.2
[85] httpuv_1.6.1 gtable_0.3.0 purrr_0.3.4
[88] assertthat_0.2.1 ggplot2_3.3.3 mime_0.10
[91] xtable_1.8-4 coda_0.19-4 later_1.2.0
[94] rsconnect_0.8.17 tibble_3.1.2 shinythemes_1.2.0
[97] gamm4_0.2-6 ellipsis_0.3.2 bridgesampling_1.1-2
```
|
process
|
loo moment match crashes r if save all pars not specified as mentioned in loo with moment matching doesn t work without save all pars save pars all true but it seems worse than just not having an appropriate warning message as it can crash r if tried without the parameter example with save all pars r library brms m brm yield n p k npk save all pars save pars all true loo m moment match true yields r computed from by log likelihood matrix estimate se elpd loo p loo looic monte carlo se of elpd loo is pareto k diagnostic values count pct min n eff inf good ok bad inf very bad all pareto k estimates are ok k see help pareto k diagnostic for details whereas r library brms m brm yield n p k npk loo m moment match true yields caught segfault address nil cause memory not mapped traceback external list name cppmethod invoke notvoid address dll list name rcpp path lib r library rcpp libs rcpp so dynamiclookup true handle info numparameters pointer pointer object misc stan fit instance unconstrain pars pars local object rstan unconstrain pars x pars rstan relist theta skeleton rstan unconstrain pars x pars rstan relist theta skeleton fun newx apply pars fun function theta rstan unconstrain pars x pars rstan relist theta skeleton unconstrain pars stanfit x fit pars pars unconstrain pars x pars pars loo loo moment match default x loo loo post draws as matrix log lik i log lik i unconstrain pars unconstrain pars log prob upars log prob upars log lik i upars log lik i upars k threshold k threshold newdata newdata resp resp dotrycatch return expr name parentenv handler trycatchone expr names parentenv handlers trycatchlist expr classes parentenv handlers trycatch expr error function e call conditioncall e if is null call if identical call quote dotrycatch call sys call dcall deparse call prefix paste error in dcall long sm strsplit conditionmessage e n w long prefix prefix n else prefix error msg prefix conditionmessage e n internal seterrmessage msg if silent istrue getoption show error messages cat msg file outfile internal printdeferredwarnings invisible structure msg class try error condition e try loo loo moment match default x loo loo post draws as matrix log lik i log lik i unconstrain pars unconstrain pars log prob upars log prob upars log lik i upars log lik i upars k threshold k threshold newdata newdata resp resp loo moment match brmsfit x loo newdata resp k threshold check loo moment match x loo newdata resp k threshold check eval expr envir eval expr envir call envir args enclos envir do call loo moment match moment match args loo x newdata resp model name pointwise k threshold save psis moment match reloo moment match args reloo args eval expr envir eval expr envir call envir args enclos envir do call criterion args fun criterion pointwise resp k threshold save psis moment match reloo moment match args reloo args x model name use stored eval expr envir eval expr envir call envir args enclos envir do call compute loo args fun models criterion pointwise compare resp k threshold save psis moment match reloo moment match args reloo args eval expr envir eval expr envir call envir args enclos envir do call compute loolist args loo brmsfit m moment match true loo m moment match true an irrecoverable exception occurred r is aborting now session info r r version platform pc linux gnu bit running under ubuntu lts matrix products default blas lib r lib librblas so lapack lib r lib librlapack so locale lc ctype en gb lc numeric c lc time en dk lc collate en gb lc monetary en gb lc messages en gb lc paper fi fi lc name c lc address c lc telephone c lc measurement en gb lc identification c attached base packages stats graphics grdevices utils datasets methods base other attached packages brms rcpp loaded via a namespace and not attached nlme matrixstats xts threejs rstan backports tools dt dbi mgcv projpred colorspace tidyselect gridextra prettyunits processx brobdingnag emmeans curl compiler cli shinyjs colourpicker scales dygraphs mvtnorm ggridges callr stringr digest stanheaders minqa pkgconfig htmltools fastmap htmlwidgets rlang shiny generics zoo jsonlite crosstalk gtools dplyr inline magrittr loo bayesplot matrix munsell fansi abind lifecycle stringi mass pkgbuild plyr grid parallel promises crayon miniui lattice splines ps pillar igraph boot estimability markdown shinystan codetools rstantools glue rcppparallel vctrs nloptr httpuv gtable purrr assertthat mime xtable coda later rsconnect tibble shinythemes ellipsis bridgesampling
| 1
|
10,967
| 3,152,375,513
|
IssuesEvent
|
2015-09-16 13:35:36
|
galenframework/galen
|
https://api.github.com/repos/galenframework/galen
|
closed
|
Add COUNT spec for multiple objects
|
c1 enhancement p2 ready for test
|
When using multiple object identification, would be great to have a COUNT spec
Example:
```
=====================================
menu-items-* css .menuitem
=====================================
menu-items-*
count: 5
```
|
1.0
|
Add COUNT spec for multiple objects - When using multiple object identification, would be great to have a COUNT spec
Example:
```
=====================================
menu-items-* css .menuitem
=====================================
menu-items-*
count: 5
```
|
non_process
|
add count spec for multiple objects when using multiple object identification would be great to have a count spec example menu items css menuitem menu items count
| 0
|
75,229
| 9,829,284,390
|
IssuesEvent
|
2019-06-15 19:17:24
|
paul-buerkner/brms
|
https://api.github.com/repos/paul-buerkner/brms
|
closed
|
Number of multiple imputation in "missing values" vignette
|
documentation
|
In the vignette "Handle Missing Values with brms", section "Imputation before model fitting", there are m = 5 multiply imputed datasets used for mixing their posterior draws. However, according to the paper cited below, m = 5 is not enough for reliable posterior inferences. The authors recommend to choose a larger number of multiply imputed datasets. They don't recommend a specific number meant by "large", but m = 100 is doing well in their simulation study. In contrast, m = 20 is not enough.
Xiang Zhou & Jerome P. Reiter (2010) A Note on Bayesian Inference After Multiple Imputation, The American Statistician, 64:2, 159-163, DOI: 10.1198/tast.2010.09109
I understand that m = 100 increases the computational burden heavily, but perhaps a remark should be added to the vignette that m = 5 might not be enough for reliable posterior inferences.
|
1.0
|
Number of multiple imputation in "missing values" vignette - In the vignette "Handle Missing Values with brms", section "Imputation before model fitting", there are m = 5 multiply imputed datasets used for mixing their posterior draws. However, according to the paper cited below, m = 5 is not enough for reliable posterior inferences. The authors recommend to choose a larger number of multiply imputed datasets. They don't recommend a specific number meant by "large", but m = 100 is doing well in their simulation study. In contrast, m = 20 is not enough.
Xiang Zhou & Jerome P. Reiter (2010) A Note on Bayesian Inference After Multiple Imputation, The American Statistician, 64:2, 159-163, DOI: 10.1198/tast.2010.09109
I understand that m = 100 increases the computational burden heavily, but perhaps a remark should be added to the vignette that m = 5 might not be enough for reliable posterior inferences.
|
non_process
|
number of multiple imputation in missing values vignette in the vignette handle missing values with brms section imputation before model fitting there are m multiply imputed datasets used for mixing their posterior draws however according to the paper cited below m is not enough for reliable posterior inferences the authors recommend to choose a larger number of multiply imputed datasets they don t recommend a specific number meant by large but m is doing well in their simulation study in contrast m is not enough xiang zhou jerome p reiter a note on bayesian inference after multiple imputation the american statistician doi tast i understand that m increases the computational burden heavily but perhaps a remark should be added to the vignette that m might not be enough for reliable posterior inferences
| 0
|
332,059
| 10,083,740,195
|
IssuesEvent
|
2019-07-25 14:16:54
|
getkirby/kirby
|
https://api.github.com/repos/getkirby/kirby
|
closed
|
KirbyTag gets escaped twice when using escape()
|
missing: discussion 🗣 missing: information ❓ priority: minor 🔜
|
**Describe the bug**
When using `escape()` on a KirbyText, KirbyTags get escaped twice.
**To Reproduce**
Given a field text with a textarea and the following content:
```
foo & bar
(link: http://example.com text: foo & bar)
```
In my template/snippet I am using the following line:
`<?= $page->text()->escape()->kirbytext() ?>`
The output is the following:
```
<p>foo & bar</p>
<p><a href="http://example.com">foo &amp; bar</a></p>
```
**Expected behavior**
The KirbyTag Link shouldn't be escaped twice.
**Kirby Version**
3.1.4
|
1.0
|
KirbyTag gets escaped twice when using escape() - **Describe the bug**
When using `escape()` on a KirbyText, KirbyTags get escaped twice.
**To Reproduce**
Given a field text with a textarea and the following content:
```
foo & bar
(link: http://example.com text: foo & bar)
```
In my template/snippet I am using the following line:
`<?= $page->text()->escape()->kirbytext() ?>`
The output is the following:
```
<p>foo & bar</p>
<p><a href="http://example.com">foo &amp; bar</a></p>
```
**Expected behavior**
The KirbyTag Link shouldn't be escaped twice.
**Kirby Version**
3.1.4
|
non_process
|
kirbytag gets escaped twice when using escape describe the bug when using escape on a kirbytext kirbytags get escaped twice to reproduce given a field text with a textarea and the following content foo bar link text foo bar in my template snippet i am using the following line text escape kirbytext the output is the following foo amp bar expected behavior the kirbytag link shouldn t be escaped twice kirby version
| 0
|
10,423
| 13,215,849,266
|
IssuesEvent
|
2020-08-17 01:25:55
|
nion-software/nionswift
|
https://api.github.com/repos/nion-software/nionswift
|
opened
|
Add ability to designate dependent/source data to be used in displays, computations
|
f - processing f - user-interface feature stage - planning type - enhancement
|
For example, if background subtraction is applied to a line plot, the user could process the original data (re-binning or smoothing, for example) and then designate the resulting data to be used anywhere the original data is used (computations, displays, maybe more).
Other ideas are: if the designated replacement is deleted, it reverts to the original source; some specialized operations may have the option to auto-designate without explicitly doing so (UI needed).
|
1.0
|
Add ability to designate dependent/source data to be used in displays, computations - For example, if background subtraction is applied to a line plot, the user could process the original data (re-binning or smoothing, for example) and then designate the resulting data to be used anywhere the original data is used (computations, displays, maybe more).
Other ideas are: if the designated replacement is deleted, it reverts to the original source; some specialized operations may have the option to auto-designate without explicitly doing so (UI needed).
|
process
|
add ability to designate dependent source data to be used in displays computations for example if background subtraction is applied to a line plot the user could process the original data re binning or smoothing for example and then designate the resulting data to be used anywhere the original data is used computations displays maybe more other ideas are if the designated replacement is deleted it reverts to the original source some specialized operations may have the option to auto designate without explicitly doing so ui needed
| 1
|
7,120
| 10,266,291,252
|
IssuesEvent
|
2019-08-22 21:02:10
|
automotive-edge-computing-consortium/AECC
|
https://api.github.com/repos/automotive-edge-computing-consortium/AECC
|
opened
|
Good place to keep the Issues list
|
priority:High status:Open type:Process
|
Looking for good place to store this issue list. Perhaps this is part of ticketing system.
Currently this issue list doesn't include sensitive item. But it will have sensitive information such as launching new SIGs. So better to be stored in work space where limited people (chairs or sponsor members) have access right.
This is the place...
|
1.0
|
Good place to keep the Issues list - Looking for good place to store this issue list. Perhaps this is part of ticketing system.
Currently this issue list doesn't include sensitive item. But it will have sensitive information such as launching new SIGs. So better to be stored in work space where limited people (chairs or sponsor members) have access right.
This is the place...
|
process
|
good place to keep the issues list looking for good place to store this issue list perhaps this is part of ticketing system currently this issue list doesn t include sensitive item but it will have sensitive information such as launching new sigs so better to be stored in work space where limited people chairs or sponsor members have access right this is the place
| 1
|
189,861
| 14,525,777,071
|
IssuesEvent
|
2020-12-14 13:24:10
|
fourMs/MGT-python
|
https://api.github.com/repos/fourMs/MGT-python
|
closed
|
Make sure the ffmpeg commands always use the -y flag
|
bug testing
|
Because if not and the destination file happens to exist already the process will just quit without doing anything.
|
1.0
|
Make sure the ffmpeg commands always use the -y flag - Because if not and the destination file happens to exist already the process will just quit without doing anything.
|
non_process
|
make sure the ffmpeg commands always use the y flag because if not and the destination file happens to exist already the process will just quit without doing anything
| 0
|
20,843
| 27,612,216,142
|
IssuesEvent
|
2023-03-09 16:43:24
|
influxdata/telegraf
|
https://api.github.com/repos/influxdata/telegraf
|
closed
|
processors.converter - convert time
|
help wanted feature request plugin/processor size/m
|
### Use Case
The idea is to improve `processors.converter` and allow it to manage timestamps.
Currently, there is no "easy" way to override/set the timestamp of a point by getting it from a tag/field, the only option is to use `processors.starlark`.
This is a rare use case, but still, I think it's worth having it
### Expected behavior
Being able to override/set the time of a point base on a field/tag value using `processors.converter`.
As done for "generic" input plugins I just expect to specify the source tag/field and time format
### Actual behavior
The only way to override/set the time of a point base on a field/tag value is to use `processors.starlark`
### Additional info
_No response_
|
1.0
|
processors.converter - convert time - ### Use Case
The idea is to improve `processors.converter` and allow it to manage timestamps.
Currently, there is no "easy" way to override/set the timestamp of a point by getting it from a tag/field, the only option is to use `processors.starlark`.
This is a rare use case, but still, I think it's worth having it
### Expected behavior
Being able to override/set the time of a point base on a field/tag value using `processors.converter`.
As done for "generic" input plugins I just expect to specify the source tag/field and time format
### Actual behavior
The only way to override/set the time of a point base on a field/tag value is to use `processors.starlark`
### Additional info
_No response_
|
process
|
processors converter convert time use case the idea is to improve processors converter and allow it to manage timestamps currently there is no easy way to override set the timestamp of a point by getting it from a tag field the only option is to use processors starlark this is a rare use case but still i think it s worth having it expected behavior being able to override set the time of a point base on a field tag value using processors converter as done for generic input plugins i just expect to specify the source tag field and time format actual behavior the only way to override set the time of a point base on a field tag value is to use processors starlark additional info no response
| 1
|
521,769
| 15,115,337,783
|
IssuesEvent
|
2021-02-09 04:12:47
|
openmsupply/mobile
|
https://api.github.com/repos/openmsupply/mobile
|
opened
|
Add cumulative breach calculation logic to BreachManager
|
Docs: not needed Feature Module: vaccines Priority: high
|
## Is your feature request related to a problem? Please describe.
Add cumulative breach calculation logic to mobile app
## Describe the solution you'd like
- Add methods to BreachManager to calculate cumulative duration above/below configured thresholds
- When consecutive breach calculation is performed, call above method to do cumulative breach calculation
## Implementation
N/A
## Describe alternatives you've considered
Not sure, CCE app has this calculation done in SQL but think we would want it to be DB agnostic?
## Additional context
N/A
|
1.0
|
Add cumulative breach calculation logic to BreachManager - ## Is your feature request related to a problem? Please describe.
Add cumulative breach calculation logic to mobile app
## Describe the solution you'd like
- Add methods to BreachManager to calculate cumulative duration above/below configured thresholds
- When consecutive breach calculation is performed, call above method to do cumulative breach calculation
## Implementation
N/A
## Describe alternatives you've considered
Not sure, CCE app has this calculation done in SQL but think we would want it to be DB agnostic?
## Additional context
N/A
|
non_process
|
add cumulative breach calculation logic to breachmanager is your feature request related to a problem please describe add cumulative breach calculation logic to mobile app describe the solution you d like add methods to breachmanager to calculate cumulative duration above below configured thresholds when consecutive breach calculation is performed call above method to do cumulative breach calculation implementation n a describe alternatives you ve considered not sure cce app has this calculation done in sql but think we would want it to be db agnostic additional context n a
| 0
|
5,712
| 8,567,916,927
|
IssuesEvent
|
2018-11-10 16:33:43
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
Wrong phone number on website / Web Form does not deliver mail
|
status-inprocess type-bug website-general
|
<img width="467" alt="screen shot 2018-07-06 at 4 39 24 pm" src="https://user-images.githubusercontent.com/5417918/43462686-c5509e8c-94a4-11e8-9d51-71aafb574283.png">
|
1.0
|
Wrong phone number on website / Web Form does not deliver mail - <img width="467" alt="screen shot 2018-07-06 at 4 39 24 pm" src="https://user-images.githubusercontent.com/5417918/43462686-c5509e8c-94a4-11e8-9d51-71aafb574283.png">
|
process
|
wrong phone number on website web form does not deliver mail img width alt screen shot at pm src
| 1
|
33,204
| 4,818,098,226
|
IssuesEvent
|
2016-11-04 15:29:37
|
infiniteautomation/ma-core-public
|
https://api.github.com/repos/infiniteautomation/ma-core-public
|
closed
|
Log4j data source - ALL level doesn't work
|
Bug Ready for Testing
|
The ALL level fails to match any messages whether using Regex or not.
This has been fixed, with validation added to saving a point with regex, a bug fix on points that were edited a few times, and some i18n issues.
|
1.0
|
Log4j data source - ALL level doesn't work - The ALL level fails to match any messages whether using Regex or not.
This has been fixed, with validation added to saving a point with regex, a bug fix on points that were edited a few times, and some i18n issues.
|
non_process
|
data source all level doesn t work the all level fails to match any messages whether using regex or not this has been fixed with validation added to saving a point with regex a bug fix on points that were edited a few times and some issues
| 0
|
38,494
| 8,850,936,038
|
IssuesEvent
|
2019-01-08 14:34:56
|
STEllAR-GROUP/phylanx
|
https://api.github.com/repos/STEllAR-GROUP/phylanx
|
closed
|
PhySL parser silently converts the type to node_data<double>
|
category: PhySL submodule: backend type: defect
|
Using square brackets to indicate a vector of integers in PhySL produces a `node_data<double>` instead of `node_data<std::int64_t>`.
```
cout(repeat(2,hstack(3)))
cout(repeat(2,[3]))
```
```console
[2, 2, 2]
physl: exception caught:
<unknown>: primitive_argument_type does not hold an integer value type (type held: 'phylanx::ir::node_data<double>'): HPX(bad_parameter)
```
```scheme
define(a,[3])
cout(repeat(2,a))
```
```console
physl: exception caught:
<unknown>: primitive_argument_type does not hold an integer value type (type held: 'phylanx::ir::node_data<double>'): HPX(bad_parameter)
```
|
1.0
|
PhySL parser silently converts the type to node_data<double> - Using square brackets to indicate a vector of integers in PhySL produces a `node_data<double>` instead of `node_data<std::int64_t>`.
```
cout(repeat(2,hstack(3)))
cout(repeat(2,[3]))
```
```console
[2, 2, 2]
physl: exception caught:
<unknown>: primitive_argument_type does not hold an integer value type (type held: 'phylanx::ir::node_data<double>'): HPX(bad_parameter)
```
```scheme
define(a,[3])
cout(repeat(2,a))
```
```console
physl: exception caught:
<unknown>: primitive_argument_type does not hold an integer value type (type held: 'phylanx::ir::node_data<double>'): HPX(bad_parameter)
```
|
non_process
|
physl parser silently converts the type to node data using square brackets to indicate a vector of integers in physl produces a node data instead of node data cout repeat hstack cout repeat console physl exception caught primitive argument type does not hold an integer value type type held phylanx ir node data hpx bad parameter scheme define a cout repeat a console physl exception caught primitive argument type does not hold an integer value type type held phylanx ir node data hpx bad parameter
| 0
|
52,661
| 13,224,887,074
|
IssuesEvent
|
2020-08-17 20:02:56
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
h5booker should refuse to book if some framenames need sanitizing (Trac #118)
|
Migrated from Trac booking defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/118">https://code.icecube.wisc.edu/projects/icecube/ticket/118</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-04-14T19:16:34",
"_ts": "1302808594000000",
"description": "",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"time": "2008-09-03T00:00:00",
"component": "booking",
"summary": "h5booker should refuse to book if some framenames need sanitizing",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
h5booker should refuse to book if some framenames need sanitizing (Trac #118) -
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/118">https://code.icecube.wisc.edu/projects/icecube/ticket/118</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-04-14T19:16:34",
"_ts": "1302808594000000",
"description": "",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"time": "2008-09-03T00:00:00",
"component": "booking",
"summary": "h5booker should refuse to book if some framenames need sanitizing",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
non_process
|
should refuse to book if some framenames need sanitizing trac migrated from json status closed changetime ts description reporter troy cc resolution wont or cant fix time component booking summary should refuse to book if some framenames need sanitizing priority major keywords milestone owner troy type defect
| 0
|
20,487
| 27,144,858,344
|
IssuesEvent
|
2023-02-16 19:06:34
|
mehta-lab/microDL
|
https://api.github.com/repos/mehta-lab/microDL
|
closed
|
Housecleaning for unused code in gunpowder implementation
|
preprocessing
|
The following sections of code will be deleted in gunpowder implementation
- Tensorflow networks module
- Inference dataset
- tensorflow dataset
- resizing
|
1.0
|
Housecleaning for unused code in gunpowder implementation - The following sections of code will be deleted in gunpowder implementation
- Tensorflow networks module
- Inference dataset
- tensorflow dataset
- resizing
|
process
|
housecleaning for unused code in gunpowder implementation the following sections of code will be deleted in gunpowder implementation tensorflow networks module inference dataset tensorflow dataset resizing
| 1
|
17,143
| 22,690,449,054
|
IssuesEvent
|
2022-07-04 19:22:44
|
threefoldtech/grid_weblets
|
https://api.github.com/repos/threefoldtech/grid_weblets
|
closed
|
UX regarding the node selection
|
process_wontfix
|
It is very confusing for the user that when selecting the node to deploy anything, you have too many options in Weblets:

You can have a "Manual Selection" which currently doesn't work,
You can select the farm and node,
But actually you need to click the "Apply Filters and Suggested Nodes" button in order to search for available nodes from which you need to select to provision a resource
|
1.0
|
UX regarding the node selection - It is very confusing for the user that when selecting the node to deploy anything, you have too many options in Weblets:

You can have a "Manual Selection" which currently doesn't work,
You can select the farm and node,
But actually you need to click the "Apply Filters and Suggested Nodes" button in order to search for available nodes from which you need to select to provision a resource
|
process
|
ux regarding the node selection it is very confusing for the user that when selecting the node to deploy anything you have too many options in weblets you can have a manual selection which currently doesn t work you can select the farm and node but actually you need to click the apply filters and suggested nodes button in order to search for available nodes from which you need to select to provision a resource
| 1
|
870
| 3,329,827,475
|
IssuesEvent
|
2015-11-11 05:45:35
|
uavis/mlproj
|
https://api.github.com/repos/uavis/mlproj
|
closed
|
Automate Skull removal
|
MS Lesion Preprocessing
|
Input: .nrrd (labels is in .nrrd as well)
.nrrd(.nhdr, .raw) -> .nii.gz -> .nrrd -> Matlab
Tool: bet2 to remove the skull. Issue: bet2 only takes .nii.gz
Right now, we’re converting the format with ITKSNAP:
```
$ /usr/work/vincent/itksnap-3.2.0-20141023-Linux-x86_64/bin/itksnap
# The nrrd produced by itk has irrelevant chars. To remove it:
$ sed -i '13,61d' UNC_train_Case01_T2_mask.nrrd % in place removal of line 13-61
```
Strip the skull with BET2:
```
/usr/work/share/fsl/bin
```
may need to set up the environmental variable before starting it:
```
$ export FSLDIR=/usr/work/share/fsl
$ . ${FSLDIR}/etc/fslconf/fsl.sh
```
```
Example:
vincent@richdale:/usr/data/medical_images/MSlesion08/tmp$ /usr/work/share/fsl/bin/bet2 UNC_train_Case01_T2_original.nii.gz UNC_train_Case01_T2_output -m
Arguments:
“UNC_train_Case01_T2_original.nii.gz” is the input scan,
“UNC_train_Case01_T2_output” is the output file without the extension,
-m : option to generate a mask (1 for brain tissue, 0 otherwise)
Output:
UNC_train_Case01_T2_output_mask.nii.gz (the mask generated by -m)
UNC_train_Case01_T2_output.nii.gz
```
Open task:
- [x] Find a converter (command line or matlab package)
- [x] write script to automate the process
|
1.0
|
Automate Skull removal - Input: .nrrd (labels is in .nrrd as well)
.nrrd(.nhdr, .raw) -> .nii.gz -> .nrrd -> Matlab
Tool: bet2 to remove the skull. Issue: bet2 only takes .nii.gz
Right now, we’re converting the format with ITKSNAP:
```
$ /usr/work/vincent/itksnap-3.2.0-20141023-Linux-x86_64/bin/itksnap
# The nrrd produced by itk has irrelevant chars. To remove it:
$ sed -i '13,61d' UNC_train_Case01_T2_mask.nrrd % in place removal of line 13-61
```
Strip the skull with BET2:
```
/usr/work/share/fsl/bin
```
may need to set up the environmental variable before starting it:
```
$ export FSLDIR=/usr/work/share/fsl
$ . ${FSLDIR}/etc/fslconf/fsl.sh
```
```
Example:
vincent@richdale:/usr/data/medical_images/MSlesion08/tmp$ /usr/work/share/fsl/bin/bet2 UNC_train_Case01_T2_original.nii.gz UNC_train_Case01_T2_output -m
Arguments:
“UNC_train_Case01_T2_original.nii.gz” is the input scan,
“UNC_train_Case01_T2_output” is the output file without the extension,
-m : option to generate a mask (1 for brain tissue, 0 otherwise)
Output:
UNC_train_Case01_T2_output_mask.nii.gz (the mask generated by -m)
UNC_train_Case01_T2_output.nii.gz
```
Open task:
- [x] Find a converter (command line or matlab package)
- [x] write script to automate the process
|
process
|
automate skull removal input nrrd labels is in nrrd as well nrrd nhdr raw nii gz nrrd matlab tool to remove the skull issue only takes nii gz right now we’re converting the format with itksnap usr work vincent itksnap linux bin itksnap the nrrd produced by itk has irrelevant chars to remove it sed i unc train mask nrrd in place removal of line strip the skull with usr work share fsl bin may need to set up the environmental variable before starting it export fsldir usr work share fsl fsldir etc fslconf fsl sh example vincent richdale usr data medical images tmp usr work share fsl bin unc train original nii gz unc train output m arguments “unc train original nii gz” is the input scan “unc train output” is the output file without the extension m option to generate a mask for brain tissue otherwise output unc train output mask nii gz the mask generated by m unc train output nii gz open task find a converter command line or matlab package write script to automate the process
| 1
|
911
| 3,373,517,198
|
IssuesEvent
|
2015-11-24 08:13:49
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
reopened
|
Not all types of 'eval' function calls wrapped
|
AREA: client SYSTEM: resource processing TYPE: bug
|
We have to wrap:
```javascript
var temp = eval;
temp('some code');
```
to
```javascript
var temp = eval === window.eval ? script => eval(__proc$Script(script)) : eval;
temp('some code');
```
|
1.0
|
Not all types of 'eval' function calls wrapped - We have to wrap:
```javascript
var temp = eval;
temp('some code');
```
to
```javascript
var temp = eval === window.eval ? script => eval(__proc$Script(script)) : eval;
temp('some code');
```
|
process
|
not all types of eval function calls wrapped we have to wrap javascript var temp eval temp some code to javascript var temp eval window eval script eval proc script script eval temp some code
| 1
|
17,687
| 23,532,206,709
|
IssuesEvent
|
2022-08-19 16:26:55
|
googleapis/google-cloud-node
|
https://api.github.com/repos/googleapis/google-cloud-node
|
closed
|
Warning: a recent release failed
|
type: process
|
The following release PRs may have failed:
* #3282 - The release job failed -- check the build log.
* #3270 - The release job failed -- check the build log.
* #3281 - The release job failed -- check the build log.
* #3280 - The release job failed -- check the build log.
|
1.0
|
Warning: a recent release failed - The following release PRs may have failed:
* #3282 - The release job failed -- check the build log.
* #3270 - The release job failed -- check the build log.
* #3281 - The release job failed -- check the build log.
* #3280 - The release job failed -- check the build log.
|
process
|
warning a recent release failed the following release prs may have failed the release job failed check the build log the release job failed check the build log the release job failed check the build log the release job failed check the build log
| 1
|
46,919
| 19,553,175,971
|
IssuesEvent
|
2022-01-03 03:12:03
|
PreMiD/Presences
|
https://api.github.com/repos/PreMiD/Presences
|
opened
|
Sushiroll | sushiroll.co.id
|
💳 Paid Service 🌍 Region Restricted
|
### Discussed in https://github.com/PreMiD/Presences/discussions/4591
<div type='discussions-op-text'>
<sup>Originally posted by **rayyplay** April 12, 2021</sup>
**Prerequisites and essential questions** <!--- Required, please answer the following questions as honestly as possible by changing the "[ ]" to "[x]" or by marking it after creating the issue (easier), not marking a question counts as "No". -->
- [x] Is it a popular site?
- [x] Is the website older than 2 months? <!--- It is necessary for the website to be older than 2 months. -->
- [x] Is the site locked to a specific country/region?
- [x] Is the site a paid service? (e.g. Netflix, Hulu)
- [ ] Does the website feature NSFW content? (e.g. porn, etc...)
- [ ] Are you a donator/patron?
- [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority?
**What's your Discord username?** rayyplay#4999 <!--- Optional, unless you are a donator/patron. Ex. Clyde#0000 -->
**What's the name of the service?** sushiroll.co.id | Sushiroll <!--- Required, Ex. www.youtube.com | YouTube -->
**What should the Presence display?** Watching Sushiroll, title anime, duration, play/pause/stop <!--- Required, make sure to be as clear as possible on what should be added. -->
**If possible, please provide a logo for the service (512x512 minimum)**

<!--- Optional, it is recommended to upload the image here instead of using a 3rd-party host. -->
</div>
|
1.0
|
Sushiroll | sushiroll.co.id - ### Discussed in https://github.com/PreMiD/Presences/discussions/4591
<div type='discussions-op-text'>
<sup>Originally posted by **rayyplay** April 12, 2021</sup>
**Prerequisites and essential questions** <!--- Required, please answer the following questions as honestly as possible by changing the "[ ]" to "[x]" or by marking it after creating the issue (easier), not marking a question counts as "No". -->
- [x] Is it a popular site?
- [x] Is the website older than 2 months? <!--- It is necessary for the website to be older than 2 months. -->
- [x] Is the site locked to a specific country/region?
- [x] Is the site a paid service? (e.g. Netflix, Hulu)
- [ ] Does the website feature NSFW content? (e.g. porn, etc...)
- [ ] Are you a donator/patron?
- [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority?
**What's your Discord username?** rayyplay#4999 <!--- Optional, unless you are a donator/patron. Ex. Clyde#0000 -->
**What's the name of the service?** sushiroll.co.id | Sushiroll <!--- Required, Ex. www.youtube.com | YouTube -->
**What should the Presence display?** Watching Sushiroll, title anime, duration, play/pause/stop <!--- Required, make sure to be as clear as possible on what should be added. -->
**If possible, please provide a logo for the service (512x512 minimum)**

<!--- Optional, it is recommended to upload the image here instead of using a 3rd-party host. -->
</div>
|
non_process
|
sushiroll sushiroll co id discussed in originally posted by rayyplay april prerequisites and essential questions is it a popular site is the website older than months is the site locked to a specific country region is the site a paid service e g netflix hulu does the website feature nsfw content e g porn etc are you a donator patron do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority what s your discord username rayyplay what s the name of the service sushiroll co id sushiroll what should the presence display watching sushiroll title anime duration play pause stop if possible please provide a logo for the service minimum
| 0
|
3,838
| 6,802,496,946
|
IssuesEvent
|
2017-11-02 20:25:15
|
WikiWatershed/model-my-watershed
|
https://api.github.com/repos/WikiWatershed/model-my-watershed
|
closed
|
Only Log Geoprocessing API Requests Not From the UI
|
1 Geoprocessing API
|
We should only log requests if they're coming from an actual API user, not a user on the app's frontend.
Depends on #2270
Depends on #2190
|
1.0
|
Only Log Geoprocessing API Requests Not From the UI - We should only log requests if they're coming from an actual API user, not a user on the app's frontend.
Depends on #2270
Depends on #2190
|
process
|
only log geoprocessing api requests not from the ui we should only log requests if they re coming from an actual api user not a user on the app s frontend depends on depends on
| 1
|
7,309
| 10,449,281,510
|
IssuesEvent
|
2019-09-19 08:04:00
|
stekylsha/CISC210Lab
|
https://api.github.com/repos/stekylsha/CISC210Lab
|
opened
|
Source Configuration Management
|
Software Process
|
### Story
As a software developer
I want to understand SCM, in particular git
So that I will be able to control management of software releases, concurrent development, and regression reversion.
### Acceptance Criteria
Demonstrate:
- Understanding of SCM
- Why? How?
- Using git, ability to ...
- clone git repository
- check out specific branch
- create new branch
- commit changes
- understand the different phases of change commit
- merge changes from another branch
- create pull request
- Describe a simple branching strategy for development and release
|
1.0
|
Source Configuration Management - ### Story
As a software developer
I want to understand SCM, in particular git
So that I will be able to control management of software releases, concurrent development, and regression reversion.
### Acceptance Criteria
Demonstrate:
- Understanding of SCM
- Why? How?
- Using git, ability to ...
- clone git repository
- check out specific branch
- create new branch
- commit changes
- understand the different phases of change commit
- merge changes from another branch
- create pull request
- Describe a simple branching strategy for development and release
|
process
|
source configuration management story as a software developer i want to understand scm in particular git so that i will be able to control management of software releases concurrent development and regression reversion acceptance criteria demonstrate understanding of scm why how using git ability to clone git repository check out specific branch create new branch commit changes understand the different phases of change commit merge changes from another branch create pull request describe a simple branching strategy for development and release
| 1
|
196,430
| 14,859,384,245
|
IssuesEvent
|
2021-01-18 18:23:17
|
kiriDevs/kiriCore
|
https://api.github.com/repos/kiriDevs/kiriCore
|
reopened
|
Write tests for MessageService
|
tests
|
Although more complex, as you need a map from at least three prefixes and some preset messages to test it fully, it would make sense to test the MessageService aswell, as it belongs to the API part and is also commonly used by kiriCore itself, making it a crucial part of the Plugin.
|
1.0
|
Write tests for MessageService - Although more complex, as you need a map from at least three prefixes and some preset messages to test it fully, it would make sense to test the MessageService aswell, as it belongs to the API part and is also commonly used by kiriCore itself, making it a crucial part of the Plugin.
|
non_process
|
write tests for messageservice although more complex as you need a map from at least three prefixes and some preset messages to test it fully it would make sense to test the messageservice aswell as it belongs to the api part and is also commonly used by kiricore itself making it a crucial part of the plugin
| 0
|
656,244
| 21,724,229,865
|
IssuesEvent
|
2022-05-11 05:42:45
|
jordan-sullivan/flashcards-2.5
|
https://api.github.com/repos/jordan-sullivan/flashcards-2.5
|
opened
|
Round class and tests
|
high priority
|
Round
Your Round class will be the object that takes in responses and records these guesses (as well as if they are correct or incorrect). The currentCard should be the first Card in the Deck (the array of Cards) at the start of the Round
Your Round class should meet the following requirements:
returnCurrentCard: method that returns the current card being played
|
1.0
|
Round class and tests - Round
Your Round class will be the object that takes in responses and records these guesses (as well as if they are correct or incorrect). The currentCard should be the first Card in the Deck (the array of Cards) at the start of the Round
Your Round class should meet the following requirements:
returnCurrentCard: method that returns the current card being played
|
non_process
|
round class and tests round your round class will be the object that takes in responses and records these guesses as well as if they are correct or incorrect the currentcard should be the first card in the deck the array of cards at the start of the round your round class should meet the following requirements returncurrentcard method that returns the current card being played
| 0
|
144,641
| 5,543,635,312
|
IssuesEvent
|
2017-03-22 17:20:55
|
jmatheson/OnMe-Backend
|
https://api.github.com/repos/jmatheson/OnMe-Backend
|
closed
|
Join Auth Credentials
|
high priority
|
Introduce new rule for `/api/join-` methods. Server should reject join operations if there's an existing user with requested credentials. In case user tries adding credential which belongs to other user, server should response with an error. Example response could look like:
{
"data" : null,
"error" : "Service already belong to another user.",
"code" : 1001
}
Also, make sure that user is unable to `unjoin` his last authentication server. In case user has only `twitter` linked, he should not be able to `unjoin` it. In case user tries so, server should reject such operation and example response might look like.
{
"data" : null,
"error" : "Illegal operation. Last authentication service can not be unjoined",
"code" : 1002
}
|
1.0
|
Join Auth Credentials - Introduce new rule for `/api/join-` methods. Server should reject join operations if there's an existing user with requested credentials. In case user tries adding credential which belongs to other user, server should response with an error. Example response could look like:
{
"data" : null,
"error" : "Service already belong to another user.",
"code" : 1001
}
Also, make sure that user is unable to `unjoin` his last authentication server. In case user has only `twitter` linked, he should not be able to `unjoin` it. In case user tries so, server should reject such operation and example response might look like.
{
"data" : null,
"error" : "Illegal operation. Last authentication service can not be unjoined",
"code" : 1002
}
|
non_process
|
join auth credentials introduce new rule for api join methods server should reject join operations if there s an existing user with requested credentials in case user tries adding credential which belongs to other user server should response with an error example response could look like data null error service already belong to another user code also make sure that user is unable to unjoin his last authentication server in case user has only twitter linked he should not be able to unjoin it in case user tries so server should reject such operation and example response might look like data null error illegal operation last authentication service can not be unjoined code
| 0
|
9,323
| 12,338,906,602
|
IssuesEvent
|
2020-05-14 17:12:35
|
unicode-org/icu4x
|
https://api.github.com/repos/unicode-org/icu4x
|
opened
|
Add chair and vice chair to README
|
C-process T-task
|
I volunteered as Chair; @zbraniecki and @nciric volunteered as Vice-Chairs. Document this.
|
1.0
|
Add chair and vice chair to README - I volunteered as Chair; @zbraniecki and @nciric volunteered as Vice-Chairs. Document this.
|
process
|
add chair and vice chair to readme i volunteered as chair zbraniecki and nciric volunteered as vice chairs document this
| 1
|
222,720
| 17,469,615,443
|
IssuesEvent
|
2021-08-06 23:35:57
|
caforbes/git_fst
|
https://api.github.com/repos/caforbes/git_fst
|
opened
|
big T after long vowel
|
info needed testing
|
write tests for big T after long vowels, since those patterns are different than after short V. Requires documentation of various alternations (e.g. amg_oot, amg_oodit)
|
1.0
|
big T after long vowel - write tests for big T after long vowels, since those patterns are different than after short V. Requires documentation of various alternations (e.g. amg_oot, amg_oodit)
|
non_process
|
big t after long vowel write tests for big t after long vowels since those patterns are different than after short v requires documentation of various alternations e g amg oot amg oodit
| 0
|
225,527
| 17,269,150,363
|
IssuesEvent
|
2021-07-22 17:19:40
|
AyatoKirishima/Portfolio
|
https://api.github.com/repos/AyatoKirishima/Portfolio
|
closed
|
Barre de navigation
|
documentation enhancement invalid
|
A faire :
- [x] Créer structure site
- [x] Editer noms & liens
- [x] Editer "logo" ou "accueil"

Image : Accueil > Photographies > Plus de photographies
|
1.0
|
Barre de navigation - A faire :
- [x] Créer structure site
- [x] Editer noms & liens
- [x] Editer "logo" ou "accueil"

Image : Accueil > Photographies > Plus de photographies
|
non_process
|
barre de navigation a faire créer structure site editer noms liens editer logo ou accueil image accueil photographies plus de photographies
| 0
|
6,651
| 9,770,118,046
|
IssuesEvent
|
2019-06-06 10:10:06
|
dzhw/zofar
|
https://api.github.com/repos/dzhw/zofar
|
closed
|
translation process / multilingual surveys
|
category: service.processes prio: 9999 status: discussion type: backlog.item
|
optimizing the process to implement the translated survey
|
1.0
|
translation process / multilingual surveys - optimizing the process to implement the translated survey
|
process
|
translation process multilingual surveys optimizing the process to implement the translated survey
| 1
|
21,793
| 30,299,017,715
|
IssuesEvent
|
2023-07-10 03:19:57
|
winter-telescope/mirar
|
https://api.github.com/repos/winter-telescope/mirar
|
opened
|
[BUG] ZOGY assumes reference images have ZP key
|
bug processors Critical
|
**Describe the bug**
Similar to #463. The reference image generators do not add the 'ZP' key to the reference images, however downstream it is assumed that the key is in there. You cannot even optionally specify a different key.
I think we should consider a combination of:
- Allowing the choice of ref ZP key
- Picking a reasonable default that matches the photometric calibration processor
|
1.0
|
[BUG] ZOGY assumes reference images have ZP key - **Describe the bug**
Similar to #463. The reference image generators do not add the 'ZP' key to the reference images, however downstream it is assumed that the key is in there. You cannot even optionally specify a different key.
I think we should consider a combination of:
- Allowing the choice of ref ZP key
- Picking a reasonable default that matches the photometric calibration processor
|
process
|
zogy assumes reference images have zp key describe the bug similar to the reference image generators do not add the zp key to the reference images however downstream it is assumed that the key is in there you cannot even optionally specify a different key i think we should consider a combination of allowing the choice of ref zp key picking a reasonable default that matches the photometric calibration processor
| 1
|
298,333
| 25,817,649,044
|
IssuesEvent
|
2022-12-12 06:54:17
|
Azure/azure-sdk-for-net
|
https://api.github.com/repos/Azure/azure-sdk-for-net
|
opened
|
[Storage] [DataMovement] Multiple single transfer tests won't run recorded
|
Storage Client test-reliability
|
### Library name
Azure.Storage.DataMovement; Azure.Storage.DataMovement.Blobs
### Please describe the feature.
Tests for pushing multiple single transfers is not running well under recorded.
Making this issue to investigate the reason behind that.
|
1.0
|
[Storage] [DataMovement] Multiple single transfer tests won't run recorded - ### Library name
Azure.Storage.DataMovement; Azure.Storage.DataMovement.Blobs
### Please describe the feature.
Tests for pushing multiple single transfers is not running well under recorded.
Making this issue to investigate the reason behind that.
|
non_process
|
multiple single transfer tests won t run recorded library name azure storage datamovement azure storage datamovement blobs please describe the feature tests for pushing multiple single transfers is not running well under recorded making this issue to investigate the reason behind that
| 0
|
21,291
| 28,487,810,661
|
IssuesEvent
|
2023-04-18 09:04:10
|
JoTec2002/TINF21C_AAS_Management
|
https://api.github.com/repos/JoTec2002/TINF21C_AAS_Management
|
opened
|
Admin dashboard edit
|
in Process frontend
|
Admin dashboard will be able to show the actual accounts and the associated roles.
|
1.0
|
Admin dashboard edit - Admin dashboard will be able to show the actual accounts and the associated roles.
|
process
|
admin dashboard edit admin dashboard will be able to show the actual accounts and the associated roles
| 1
|
765,627
| 26,854,475,260
|
IssuesEvent
|
2023-02-03 13:38:20
|
KinsonDigital/Velaptor
|
https://api.github.com/repos/KinsonDigital/Velaptor
|
closed
|
🔬Research how to add different type layered rendering
|
high priority 🔬research
|
### Complete The Item Below
- [X] I have updated the title without removing the 🔬 emoji.
### What To Research
Currently, layered rendering has been implemented for all the different types of rendering such as textures, lines, rect, text etc.
When rending on different layers for layers in its own rendering category, everything works fine. But, when rendering a rectangle and text, they do not take layering into account between the 2 different types.
**Example:**
Suppose we had a single rectangle and a single piece of text that was being rendered onto the screen. The rectangle was set to render on layer 10 and the text was set to render on layer 5, the text would still render on top of the rectangle.
The reason for this is that the layers are only taken into account for that rendering category. The categories themselves are rendered in a specific order in the `Renderer` class and this is not configurable.
Research needs to be done on how to consider layers between categories. This research needs to be done to assess the number of changes that will be required and the risk involved in making this happen.
The results from this research will need to be heavily documented and then an issue created to implement this feature.
> **Note** Need to look into forcing the text to render after the button face for button controls
### Research Results
After doing some prototyping, it was decided that the best course of action would be to do the following:
1. Class name refactoring
- Rename the batch manager class to something like `IBatchManagerProxy`
- Rename the batching services to something like `IBatchManager`. Basically replace the word _**Service**_ with **Manager**
2. Create 4 new **Carbonate** events. One for each batch type.
- These events are going to be pushed out by the `BatchManager` to send data to be rendered between the 4 different renderers. These rendered are coming with issue #451.
3. For performance reasons, the renderers and the `BatchManager` will use pointers to push and receive data to be rendered.
4. An `internal` and `sealed` service will be created to wrap the `Unsafe.AsPointer()` and `Unsafe.AsRef<T>` dotnet types to maintain testability.
5. The `BatchManager` will contain all 4 arrays of data of 1000 items each for each render type.
6. A new `readonly struct` will be created called `RenderItem<T>` which will hold the batch time and metadata.
- The metadata will be the layer, date and time of the add, the pointer to the array data item, and the batch type. All of this information is used to properly sort what should be rendered and in what order before telling the renderers what to render.
7. A new `readonly struct` will be created called `ItemToRender` which will hold data to be sent to a renderer.
- This `struct` holds data such as the batch type, layer, date/time, and a pointer to the data for the renderer to easily get the data to send to the GPU.
### Acceptance Criteria
**This issue is finished when:**
- [x] Research complete and issues created _(if needed)_.
- [x] If any issues were created, they have been added to the _**Related Work**_ section below.
### ToDo Items
- [x] Priority label added to this issue. Refer to the _**Priority Type Labels**_ section below.
- [x] Issue linked to the correct project.
### Issue Dependencies
_No response_
### Related Work
_No response_
### Additional Information:
**_<details closed><summary>Priority Type Labels</summary>_**
| Priority Type | Label |
|---------------------|-------------------|
| Low Priority | `low priority` |
| Medium Priority | `medium priority` |
| High Priority | `high priority` |
</details>
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct.
|
1.0
|
🔬Research how to add different type layered rendering - ### Complete The Item Below
- [X] I have updated the title without removing the 🔬 emoji.
### What To Research
Currently, layered rendering has been implemented for all the different types of rendering such as textures, lines, rect, text etc.
When rending on different layers for layers in its own rendering category, everything works fine. But, when rendering a rectangle and text, they do not take layering into account between the 2 different types.
**Example:**
Suppose we had a single rectangle and a single piece of text that was being rendered onto the screen. The rectangle was set to render on layer 10 and the text was set to render on layer 5, the text would still render on top of the rectangle.
The reason for this is that the layers are only taken into account for that rendering category. The categories themselves are rendered in a specific order in the `Renderer` class and this is not configurable.
Research needs to be done on how to consider layers between categories. This research needs to be done to assess the number of changes that will be required and the risk involved in making this happen.
The results from this research will need to be heavily documented and then an issue created to implement this feature.
> **Note** Need to look into forcing the text to render after the button face for button controls
### Research Results
After doing some prototyping, it was decided that the best course of action would be to do the following:
1. Class name refactoring
- Rename the batch manager class to something like `IBatchManagerProxy`
- Rename the batching services to something like `IBatchManager`. Basically replace the word _**Service**_ with **Manager**
2. Create 4 new **Carbonate** events. One for each batch type.
- These events are going to be pushed out by the `BatchManager` to send data to be rendered between the 4 different renderers. These rendered are coming with issue #451.
3. For performance reasons, the renderers and the `BatchManager` will use pointers to push and receive data to be rendered.
4. An `internal` and `sealed` service will be created to wrap the `Unsafe.AsPointer()` and `Unsafe.AsRef<T>` dotnet types to maintain testability.
5. The `BatchManager` will contain all 4 arrays of data of 1000 items each for each render type.
6. A new `readonly struct` will be created called `RenderItem<T>` which will hold the batch time and metadata.
- The metadata will be the layer, date and time of the add, the pointer to the array data item, and the batch type. All of this information is used to properly sort what should be rendered and in what order before telling the renderers what to render.
7. A new `readonly struct` will be created called `ItemToRender` which will hold data to be sent to a renderer.
- This `struct` holds data such as the batch type, layer, date/time, and a pointer to the data for the renderer to easily get the data to send to the GPU.
### Acceptance Criteria
**This issue is finished when:**
- [x] Research complete and issues created _(if needed)_.
- [x] If any issues were created, they have been added to the _**Related Work**_ section below.
### ToDo Items
- [x] Priority label added to this issue. Refer to the _**Priority Type Labels**_ section below.
- [x] Issue linked to the correct project.
### Issue Dependencies
_No response_
### Related Work
_No response_
### Additional Information:
**_<details closed><summary>Priority Type Labels</summary>_**
| Priority Type | Label |
|---------------------|-------------------|
| Low Priority | `low priority` |
| Medium Priority | `medium priority` |
| High Priority | `high priority` |
</details>
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct.
|
non_process
|
🔬research how to add different type layered rendering complete the item below i have updated the title without removing the 🔬 emoji what to research currently layered rendering has been implemented for all the different types of rendering such as textures lines rect text etc when rending on different layers for layers in its own rendering category everything works fine but when rendering a rectangle and text they do not take layering into account between the different types example suppose we had a single rectangle and a single piece of text that was being rendered onto the screen the rectangle was set to render on layer and the text was set to render on layer the text would still render on top of the rectangle the reason for this is that the layers are only taken into account for that rendering category the categories themselves are rendered in a specific order in the renderer class and this is not configurable research needs to be done on how to consider layers between categories this research needs to be done to assess the number of changes that will be required and the risk involved in making this happen the results from this research will need to be heavily documented and then an issue created to implement this feature note need to look into forcing the text to render after the button face for button controls research results after doing some prototyping it was decided that the best course of action would be to do the following class name refactoring rename the batch manager class to something like ibatchmanagerproxy rename the batching services to something like ibatchmanager basically replace the word service with manager create new carbonate events one for each batch type these events are going to be pushed out by the batchmanager to send data to be rendered between the different renderers these rendered are coming with issue for performance reasons the renderers and the batchmanager will use pointers to push and receive data to be rendered an internal and sealed service will be created to wrap the unsafe aspointer and unsafe asref dotnet types to maintain testability the batchmanager will contain all arrays of data of items each for each render type a new readonly struct will be created called renderitem which will hold the batch time and metadata the metadata will be the layer date and time of the add the pointer to the array data item and the batch type all of this information is used to properly sort what should be rendered and in what order before telling the renderers what to render a new readonly struct will be created called itemtorender which will hold data to be sent to a renderer this struct holds data such as the batch type layer date time and a pointer to the data for the renderer to easily get the data to send to the gpu acceptance criteria this issue is finished when research complete and issues created if needed if any issues were created they have been added to the related work section below todo items priority label added to this issue refer to the priority type labels section below issue linked to the correct project issue dependencies no response related work no response additional information priority type labels priority type label low priority low priority medium priority medium priority high priority high priority code of conduct i agree to follow this project s code of conduct
| 0
|
52,473
| 7,765,327,904
|
IssuesEvent
|
2018-06-02 02:00:08
|
reactjs/react-modal
|
https://api.github.com/repos/reactjs/react-modal
|
opened
|
Sprint for react-modal v4.
|
dependencies discussion documentation help wanted react testing
|
It's time for react-modal to get ready for `react 16.3+`, this includes:
## Checklist
1. Deprecate old lifecycle methods and remove polyfills.
2. Decide how long we can maintain react-modal versions below `v4` (backports or another approach).
3. Improve documentation where missing.
4. Add more examples.
If you want to give it a try on any item on the list, create a new issue and assign to you.
Appreciate ideas regarding this update and help with all the listed items.
|
1.0
|
Sprint for react-modal v4. - It's time for react-modal to get ready for `react 16.3+`, this includes:
## Checklist
1. Deprecate old lifecycle methods and remove polyfills.
2. Decide how long we can maintain react-modal versions below `v4` (backports or another approach).
3. Improve documentation where missing.
4. Add more examples.
If you want to give it a try on any item on the list, create a new issue and assign to you.
Appreciate ideas regarding this update and help with all the listed items.
|
non_process
|
sprint for react modal it s time for react modal to get ready for react this includes checklist deprecate old lifecycle methods and remove polyfills decide how long we can maintain react modal versions below backports or another approach improve documentation where missing add more examples if you want to give it a try on any item on the list create a new issue and assign to you appreciate ideas regarding this update and help with all the listed items
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.