Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
10,823
3,436,596,271
IssuesEvent
2015-12-12 14:39:16
lintool/warcbase
https://api.github.com/repos/lintool/warcbase
closed
Redo Documentation to Account for getContentString, getContentBytes, etc.
documentation
I've just redone the ["Analysis of Site Link Structure"](http://lintool.github.io/warcbase-docs/Spark-Analysis-of-Site-Link-Structure/) walkthrough in the docs to account for our API revisions. Currently, all these scripts will crash as they're referring to depreciated code. Will doublecheck that it works and can do others, such as <http://lintool.github.io/warcbase-docs/Spark-Extracting-Domain-Level-Plain-Text/>.
1.0
Redo Documentation to Account for getContentString, getContentBytes, etc. - I've just redone the ["Analysis of Site Link Structure"](http://lintool.github.io/warcbase-docs/Spark-Analysis-of-Site-Link-Structure/) walkthrough in the docs to account for our API revisions. Currently, all these scripts will crash as they're referring to depreciated code. Will doublecheck that it works and can do others, such as <http://lintool.github.io/warcbase-docs/Spark-Extracting-Domain-Level-Plain-Text/>.
non_process
redo documentation to account for getcontentstring getcontentbytes etc i ve just redone the walkthrough in the docs to account for our api revisions currently all these scripts will crash as they re referring to depreciated code will doublecheck that it works and can do others such as
0
123,930
26,357,838,033
IssuesEvent
2023-01-11 11:03:47
vegaprotocol/specs
https://api.github.com/repos/vegaprotocol/specs
closed
0059-STKG-simple_staking_and_delegating - spec needs updating to align with implementation
ac-code-remediation
Specs seems out of date and requires a full review against what has been delivered Once completed we can review again for missing ACs
1.0
0059-STKG-simple_staking_and_delegating - spec needs updating to align with implementation - Specs seems out of date and requires a full review against what has been delivered Once completed we can review again for missing ACs
non_process
stkg simple staking and delegating spec needs updating to align with implementation specs seems out of date and requires a full review against what has been delivered once completed we can review again for missing acs
0
116,387
4,701,053,082
IssuesEvent
2016-10-12 20:24:29
Innovate-Inc/CRS
https://api.github.com/repos/Innovate-Inc/CRS
opened
Review CRS ArcGIS Server Environment
CRS Review GIS Team MVP Priority: Medium
Set up a time to test and finishing configuring ArcGIS Server and SQL Server connectivity.
1.0
Review CRS ArcGIS Server Environment - Set up a time to test and finishing configuring ArcGIS Server and SQL Server connectivity.
non_process
review crs arcgis server environment set up a time to test and finishing configuring arcgis server and sql server connectivity
0
6,109
4,155,302,782
IssuesEvent
2016-06-16 14:33:22
sixeco/ProjectAlbatross
https://api.github.com/repos/sixeco/ProjectAlbatross
opened
Relevance Assessment
usability user story
Als Usabiltiy - Fachmann möchte ich ein Relevance Assessment durchführen, um zu sehen wie gut meine Suche ranked.
True
Relevance Assessment - Als Usabiltiy - Fachmann möchte ich ein Relevance Assessment durchführen, um zu sehen wie gut meine Suche ranked.
non_process
relevance assessment als usabiltiy fachmann möchte ich ein relevance assessment durchführen um zu sehen wie gut meine suche ranked
0
2,726
5,612,442,533
IssuesEvent
2017-04-03 05:11:30
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Insertrow modal placeholders have commas when dynamically generating the table.
enhancement inprocess
This is probably because the <TableHeaderColumn/> gets its child through a map operation. `this.props.keys.map( (key) => <TableHeaderColumn key={key} width={key==="sno"?"50px":"150px"} dataField={key}> {key.toUpperCase()} </TableHeaderColumn> )` ![screenshot from 2017-03-14 00-45-15](https://cloud.githubusercontent.com/assets/24292251/23871685/7519823e-0851-11e7-9d56-083842eb0357.png)
1.0
Insertrow modal placeholders have commas when dynamically generating the table. - This is probably because the <TableHeaderColumn/> gets its child through a map operation. `this.props.keys.map( (key) => <TableHeaderColumn key={key} width={key==="sno"?"50px":"150px"} dataField={key}> {key.toUpperCase()} </TableHeaderColumn> )` ![screenshot from 2017-03-14 00-45-15](https://cloud.githubusercontent.com/assets/24292251/23871685/7519823e-0851-11e7-9d56-083842eb0357.png)
process
insertrow modal placeholders have commas when dynamically generating the table this is probably because the gets its child through a map operation this props keys map key key touppercase
1
21,076
28,019,955,588
IssuesEvent
2023-03-28 04:02:59
0xPolygonMiden/miden-vm
https://api.github.com/repos/0xPolygonMiden/miden-vm
closed
Replace `mtree_cwm` instruction with `mtree_merge`
assembly processor
Once we integrate `MerkleStore` into the `AdviceProvider` (#774) `mtree_cwm` instruction will no longer be necessary - so, we should remove it. At the same time, we need to introduce a new instruction which would merge two nodes in the advice provider using [merge_roots()](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs#L255) function. We could call this instruction `mtree_merge` and it could work as follows: - Call `merge_roots()` method on the `AdviceProvider` using the top two words on the stack as inputs. - Execute an equivalent of `hmerge` instruction to hash the top two words on the stack.
1.0
Replace `mtree_cwm` instruction with `mtree_merge` - Once we integrate `MerkleStore` into the `AdviceProvider` (#774) `mtree_cwm` instruction will no longer be necessary - so, we should remove it. At the same time, we need to introduce a new instruction which would merge two nodes in the advice provider using [merge_roots()](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs#L255) function. We could call this instruction `mtree_merge` and it could work as follows: - Call `merge_roots()` method on the `AdviceProvider` using the top two words on the stack as inputs. - Execute an equivalent of `hmerge` instruction to hash the top two words on the stack.
process
replace mtree cwm instruction with mtree merge once we integrate merklestore into the adviceprovider mtree cwm instruction will no longer be necessary so we should remove it at the same time we need to introduce a new instruction which would merge two nodes in the advice provider using function we could call this instruction mtree merge and it could work as follows call merge roots method on the adviceprovider using the top two words on the stack as inputs execute an equivalent of hmerge instruction to hash the top two words on the stack
1
90,310
3,814,366,915
IssuesEvent
2016-03-28 12:52:54
minetest/minetest
https://api.github.com/repos/minetest/minetest
closed
Incomplete mapblocks being saved to database
@ Mapgen Can't fix Low priority
How to reproduce: (with a Lua mapgen) 1) Manipulate the mapgen, so it throws an error while building the terrain 2) Fix the error 3) Rejoin and enjoy the TNT-effect Basically, the database should not save incomplete mapblocks, but in this case, it does.
1.0
Incomplete mapblocks being saved to database - How to reproduce: (with a Lua mapgen) 1) Manipulate the mapgen, so it throws an error while building the terrain 2) Fix the error 3) Rejoin and enjoy the TNT-effect Basically, the database should not save incomplete mapblocks, but in this case, it does.
non_process
incomplete mapblocks being saved to database how to reproduce with a lua mapgen manipulate the mapgen so it throws an error while building the terrain fix the error rejoin and enjoy the tnt effect basically the database should not save incomplete mapblocks but in this case it does
0
60,231
8,409,312,424
IssuesEvent
2018-10-12 06:49:17
bio-phys/MDBenchmark
https://api.github.com/repos/bio-phys/MDBenchmark
opened
Update documentation for 2.0
documentation good first issue
We just merged the documentation into `develop` (PR #101). Before releasing version 2.0, we should update the documentation to reflect the new functionality. This should actually be only minor changes, because the biggest part was to write the initial documentation in the first place.
1.0
Update documentation for 2.0 - We just merged the documentation into `develop` (PR #101). Before releasing version 2.0, we should update the documentation to reflect the new functionality. This should actually be only minor changes, because the biggest part was to write the initial documentation in the first place.
non_process
update documentation for we just merged the documentation into develop pr before releasing version we should update the documentation to reflect the new functionality this should actually be only minor changes because the biggest part was to write the initial documentation in the first place
0
202,976
15,863,576,389
IssuesEvent
2021-04-08 12:56:39
cornellius-gp/gpytorch
https://api.github.com/repos/cornellius-gp/gpytorch
reopened
How to use Adam optimizer instead of SGD in the example given in the document SVDKL
documentation
Hi, I am using the [example](https://docs.gpytorch.ai/en/v1.2.1/examples/06_PyTorch_NN_Integration_DKL/Deep_Kernel_Learning_DenseNet_CIFAR_Tutorial.html) and whilst using SGD optimzer , my accuracy stuck at 50% on my custom dataset. I would like to use Adam instead but I know from the [link](https://docs.gpytorch.ai/en/stable/examples/04_Variational_and_Approximate_GPs/Natural_Gradient_Descent.html#Difference-#2:-Two-optimizers---one-for-the-variational-parameters;-one-for-the-hyperparameters) that Adam can be used for hyper-parameters so how can I use it in deep kernel learning example for variational parameters. Is using Adam as follows without including variational parameters, the right way? `optimizer = torch.optim.Adam([ {'params': model.parameters()}, {'params': likelihood.parameters()}, ], lr=0.1)`
1.0
How to use Adam optimizer instead of SGD in the example given in the document SVDKL - Hi, I am using the [example](https://docs.gpytorch.ai/en/v1.2.1/examples/06_PyTorch_NN_Integration_DKL/Deep_Kernel_Learning_DenseNet_CIFAR_Tutorial.html) and whilst using SGD optimzer , my accuracy stuck at 50% on my custom dataset. I would like to use Adam instead but I know from the [link](https://docs.gpytorch.ai/en/stable/examples/04_Variational_and_Approximate_GPs/Natural_Gradient_Descent.html#Difference-#2:-Two-optimizers---one-for-the-variational-parameters;-one-for-the-hyperparameters) that Adam can be used for hyper-parameters so how can I use it in deep kernel learning example for variational parameters. Is using Adam as follows without including variational parameters, the right way? `optimizer = torch.optim.Adam([ {'params': model.parameters()}, {'params': likelihood.parameters()}, ], lr=0.1)`
non_process
how to use adam optimizer instead of sgd in the example given in the document svdkl hi i am using the and whilst using sgd optimzer my accuracy stuck at on my custom dataset i would like to use adam instead but i know from the that adam can be used for hyper parameters so how can i use it in deep kernel learning example for variational parameters is using adam as follows without including variational parameters the right way optimizer torch optim adam params model parameters params likelihood parameters lr
0
137,546
11,140,629,213
IssuesEvent
2019-12-21 15:52:37
bitfocus/companion-module-requests
https://api.github.com/repos/bitfocus/companion-module-requests
closed
H2R Support
Needs testers Software Stale
Module Support for a H2R that does cut and fill and Chroma key messages and lower thrids https://heretorecord.com/graphics#download https://heretorecord.com/graphics/docs#osc-control Just adding OSC support in a module
1.0
H2R Support - Module Support for a H2R that does cut and fill and Chroma key messages and lower thrids https://heretorecord.com/graphics#download https://heretorecord.com/graphics/docs#osc-control Just adding OSC support in a module
non_process
support module support for a that does cut and fill and chroma key messages and lower thrids just adding osc support in a module
0
203,409
15,366,343,815
IssuesEvent
2021-03-02 01:08:08
gardners/surveysystem
https://api.github.com/repos/gardners/surveysystem
closed
backend, extend test_serialiser
Priority: MEDIUM backend tests
housekeeping: `test_serialiser.c` can be used for unit tests in general, rename and improve This is the first step to better unit test ing
1.0
backend, extend test_serialiser - housekeeping: `test_serialiser.c` can be used for unit tests in general, rename and improve This is the first step to better unit test ing
non_process
backend extend test serialiser housekeeping test serialiser c can be used for unit tests in general rename and improve this is the first step to better unit test ing
0
622
3,089,032,404
IssuesEvent
2015-08-25 19:32:08
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Can no longer publish to XHTML image with data protocol [DOT 2.x develop branch]
bug P2 preprocess
In my DITA topic I refer to an embedded image like: ```xml <image href="data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAjUAAAC9CAIAAADECxtTAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAAAJcEhZcwAAEnQAABJ0Ad5mH3gAABp8SURBVHhe7Zxbjlw3kkC9K6nWI6n20EsQ5K7f+ehNGLBRCxkDPXZ9CGiggQa0APV4ZgBPkEHyksHHfWRmJbPyHARuJ8lgRJCXjKgsVfuHPwEAAObD1qcfAAAAro2rR1qWEtoLAABwLahPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNCfQIAgBmhPgEAwIxQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNygfr0/CjT9/H4HOZek07cjdgWzSkCBwB4i0iOdU9tJLT3VPYXquun+zpmG9PL00MYUShRAACXQBKse2ojob2nY3O5kKXzxqhw5XxfFSgTz3oBgxXkvT88vYTGZXgFFwBwaSS/uqc2Etp7OuP65Gl8ybpqZqE+XRq3gxd+xa/gAgAujuRX99RGQnvPwJZsXutcM7dU0awVVMrTLnT7LvqCX8EFALwCcpHdUxsJ7T0Dde1ppPNa6YrJpQpmGDDFaRfp+/Tl3u8ruACA10EusntqI6G9Z2BTfWr8FvB6iX9LfYJDZFt7qeLxCi4A4LWQi+ye2kho7xnYVJ8aWtfLLtSny1Du60Ve7yu4AIBXRC6ye2ojob1nYFN94vvTW6fa1PMXj1dwAQCvi1xk99RGQnvPwKb6VCsNkostZtvS0PZZg/rUiNORrahVaZNCObged8ddxXoBbYcVQrBujm1oP4z1ZbQ9tuZ1YjvmYvsSKuzUlT0z8QVtMbIyD+DOkevintpIaO8ZqDNHnQMa2aWdKJJiHE4do3u+c1YVjYmlGrcKrbRXT/K0l2lMVGFbelYslYGHp+cq1sCmDU1aWbyru9XAOAtzst7SSlt9THfO+hIMaWr1XoyPQBxeRvOAt747gLtE7oh7aiOhvWcgv4uKvZG1RvOid/LHMruZHQ7MquJZDXjDirrYuUI5vR220M6GQ5qBJTt2tBGakGnlEeQVtRFZPuzpRb8olu5NcKe4OLgER6bSfi92Xm85aUp7kwHAI3fEPbWR0N4zYPKK0L7YC83U0EkpQm7CXvZDs6qYBlaVVQWHKlVJdG1yOVwOdpJon0ZghX0bnY2tVCjdD4aEat3t2Eu13L0NvZ5/xMWeJRQxmNE8ul7UdjN1rNpiAFiQO+Ke2kho7xmweWVMO6WURsyNLpJKMXZsVh2xTSH7FXINOzhIkYLZEDPc2a4uq5GvuC/nm8nl4DjwTuiV1uJiMBTY5uL4EopBM7HwvYyV9pruqmUAwIJcG/fURkJ7z4C9om2aqSRSJh6rWo4u1/3YLKGK2KaQ0xSGg2VYgonbjlvPK1jf9baPNFacm+HS9trCAsZ9plQZqJa+xcXxJZRD1nQ5mqxW71rIp8r4zjcIcF/IlXFPbSS09wzUV3T3jSxNjFPDMnpsllBFbAM+TWE4aBJkZdqO22WtYH3X0wcBnBRbNbkXeaa4761scXHCEkr/1nQ5MY1W/iJV8ADQQm6Le2ojob1noMoru+9mbWFEtH5slmM1E56mMBzcm0Ct5xWs7zqDD1TWJg9jrwYbrjtUG6ZUS9/g4vgSOkF0SNMqgxnbdwDgbpGb4p7aSGjvGagv9s6cai1svNbHZjmqiG3ApymMZ9uEZuI2wyduZWtXuuGtpv+hwursFikWiaEyUK193cUJSzDbsiV6pTJZsPcFAtwZckvcUxsJ7T0DVb47Oalum39slqOK2E49TWFtthkvR4vB7UkyYn03LHTDOyG5C6uzSzL1oFgZsJu+wcWqRl/BbkvlfUS1pzm7LAHcGXJF3FMbCe09A/Xd3HshuxlzyLFZjtWZpymsB1Zq9OZW0zZgfTeKRF9lbfKwgqwXj0ShmmwMrXs2uDi+BDtz9/ZXBhL9nQC4e+SGuKc2Etp7Bup7ufdmV0ljk4Vjsxyrqeg0hdXZnlwrKix9R1Oa9d2wY1WW6OyOriX3Yrx6HZ0llHrZ1lQGqm3b4GIYotAfr4y3X9uY2ojngCWAO0EuiHtqI6G9Z6BKx/uvY22ik9xEcxk5NkuoJtqAT1NYnb3QyGe9NWzD+q6tjTK4GbOTjfFyuFpKcyFGK9+ZykC1bVtcHF9C9draKxCy42ROlme7IQCQC+Ke2kho7xmor2M/H/eobfRyz7AQbJrVmGcDPk1hdbZD8+jZE5f1XTsYJnAzaAIvjRvTZmo+LkPBklXKbVQGqm074mLHEqr3Vq3R4ewvRsvWQmmrpQEADrkg7qmNhPaeTpU0mrd6lUZyKA15BXvRzzTrvAqrs6PGkX1awfquXIzTtxkvZ5dTe4taCLOlP9mpdJKZeqSxPVtcnHUJjtyCV8inqcHKUOmpMQwAitwQ99RGQntPp3WnD13IZnLIqfOVcGRWNcfGe5rC2uxs/NBGjah870vQjlwln9zrj5SmcxY3fZ3uf2i9F0JJuZKjS2jtn8FMSxbtVi6u2o4AwCN3xD21kdDeU+nd52OJd5AdBrd876yGfhnuqkIjTy4K9ewyikG4hgObaI1r2o+GirgHO5rpRa2sqx9Wc2mleksl+Gjsah3jBheeo0sYvp/GluXa2XDsbswAgAy5Ju6pjYT2HmRwhw2jTNCmSlOb7vi2WcO4H55+X1MQk40sGpDh4eCWGGo2rT5hjevsdu8a9WK2zCt8Nd9+rmEsFi67R2fdReTYEoRqYm+exBIisFM2egK4b+SuuKc2EtoLr06dM9fYUeOPVSIAgOsgaco9tZHQXrgGtoqssb1AUZ8A4JaQNOWe2khoL1yNHUVqe5WhPgHALSFpyj21kdBeeH2WGjL6XpRXGuoTALxNJE25pzYS2guvTaog66UjqfL7PQB4m0iack9tJLQXXpnlryM216c9NYb6BAC3hKQp99RGQnvhtcn+fG/8tSgo7qsw1CcAuCUkTbmnNhLaC1cgLyLNGhVr2O7qYsuTsP13gwAAr40kKffURkJ74XoM/o9Q+4vKwJiHMgUAEyLZyT21kdBeAACAa0F9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNCfQIAgBmhPgEAwIxQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8T8fwoG//w9BKa2zk88Q54eXqQzXl8Ds1d7NhXXgHA2ZFL5Z7aSGjvLePSxVmyhc87CZfmyp7kRPOg54jjZHXv5MMT74Sj9Sm9z/V93aEKANuRS+We2kho702SVY9zZYuQfrIUF3oqB67/2E/qHh/71qhfnh6T5q6JFyKPZ8x2zWvj3/O2fd2hevfczgGAKyN5zT21kdDemyOUjcjebCFTwidLnf+1DNpa9Px4Uobak+MkgEVzguRYxDNku+bVoT5dggsegP4VhptEX+gbqU8BrR37s4XOCo0Sb7Kwp7WwLFAnlqcdOc5oXj05Ho58bnYEe1PruiaX2yh/falPbwp9odQnh0z59u2bPEM7R23Wv+LLC1RVnmIgnqKULSOPzy9PT2EoXd04bL+fKYVZr78yMcTq6W+Kn5sbqf3nnhdDdTy2M9ppavbMBrrBL7PyPaxIexPw01w7Gq48LmRzlzD8YlIzzs7d5JH5wXz60GGh5yYXO6bWKtcNHY8qSkeY4Ye8cm7ZLC+bv3C2qIrOpJqbzzZHu2VqGG+GFpHx7uWFm0VfKPXJIVP+5ZEPoWshXZbUFPIecVs49FEU9zK7joumG4la3ocz6zsKAxkvL1lBCl2DiZWD1r74KY6Hx8cH1VbdxY5vl4a02YiniMA3/FhDs2/W0w1eGqW7ZU5GsJdmLcuUvfJ9RaQVfn4WbKldjJaqopgZzePTkDKbFZWG91rMEJXYLkJaNNWI41He6NOLG3n/l7+ETunTOTpBNNRcYazkPFE1DoAbtZvzezP+wlmBqPVvLtww+kKpTw6Z8s+IfA69Eb0x4SZJw/3Qbnv8J8UPpggKVXPXzPen8rJ21mDG+hPlc2FiyRUVaiQby636z8UCdZdDV+6y2eqsamxWRotQs+C7e1hROjRxV6MldrQMtxhdGu5TviRpFw6yVXSwGoUjQdrt+H1rcV4uVWmplPN7oZ0pqmqs8Fc4acVfIyrh0rauLdw0+kKpTw6Z8g+PfAhdOdk1k4/+f5fbFHsa6DxH1OgF6DVTZ9kqMWPdif5jTTPU2p/Odsr1WJlKWuOOxX90WWiOzS6TC9TS9pfsNDOt0mcrggU76ttp84rR0Hiq/u3fD9R4K3ERgbblZCEOt/+Zc/GTzHjzqaWUprXZ8WspR6O/vVEVVpbhgqDbir+JqI1uLtws+kKpTw6Z8vXrV3mGtkUvk9wX+ZDfIPGz9ORoIC6MNDWSXcwlTN+ZmmWrxIx1J7qPHRMVtT/f46OO6wwDnmW0Nbe39kJzbNZ9MoMFqukZqTkn2XgRgGlZ7Khvp7dYjPqGUppzAz0HukmJZDmMaNsn/szX82Om5+htdW4kUQQ9XFGDc0RV+HCNvr9W/D1Ec3h54SbRF0p9cuis0GjiLbt/eEqWtSf9o81CedXLK5rQ7qTXmNNZgxnrTvQfN97w2t/S4z8ZQ7l+ObfR6qenvln/aTV4tZH5s7hXlA0WAZiWxY6WERWjsVEZLOdsRg/W8pNPZj83VrrzrcWZt2FclxNMdGaw5gxRFWO+YSJcaMU/QJSF0IA3gb7Qe6hPektG510Gw6cuajo3HJxZq+UNLa9o+aNmdgnLScZEQct8c6K3boxUP+oqtT83OXT4we5oOdcY0qnRZTk4NjsIvruHNXmcQiMA42LBjvp28lOMLg3/KYtmsIoRajD7WcjbMZaKECrXrW1pzWivqIlqnBJVOTjcnFb8Y0Q/fII3gb7QN1Wf9Do48qPd7t2NuzLldXKGzQUT1F3sD5dYPD8//fX30obTDCH5SWmkbJUsY/4vA0YTg+/QdmMjmyaPZJuldgqzi6HF5cvTX//mZ8bBfO3SlWu6v2kYmo2TQ08WfO1+idTgbSyjSwB1y1KNZrZCaA7Xkav6z5nL7irGmMCrtqCeorXgR1T8Vrcm2CWVKmawyalRLT7yAxC13WgKoGHa2IY3jrxr99RGQntvDj27hniU46A573up/xm4819ryYN5fA630Ht31zS2hRBQ1pPpO5oRR4WHp7+tTixD0b4ar+V+VRk06yRQ7K8xtMQjk/prt5qOkdle8O09rMh3w/nL22ubbOd6lmjC32+WfwwtiGrWjmbbq1jBnjUxUs0tDceY4x+XBuKscknPpcp4NxZOicoPxmZjUxfPzfgF7S8jgDeLvGv31EZCe+Gu8Defiw8As0B9ggD1CQCmgvoEAeoTAEwF9QmE4rf91CgAmAFJR+6pjYT2AgAAXAvqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9Qnx7//5/8QBEEmkZCY7h7qk8McDgRBkCtKSEx3D/XJYQ4HgiDIFSUkpruH+uQwhwNBEOSiIjlWMJ1JQmK6e7QSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+ZhBfvoo7+H957//r+m/unz/45cPLrQff/1jutiacnMBzynTHsgDcvUjIc4F05kkJKa7R7bIPbWR0N6b5PnRv3fHw9NL6FzDHI4k3//47fP7YM3z6ad4mr///Cn0Cec+5eHyeNOzpYPvf//xXQjtNtL9zQU8ocx8IA/IDEdC/ZvOJCEx3T2yRe6pjYT23hwvTw/+pWc8PoexIeZwGNFS9O7Lb7bfVa8zX9dfv3xKBif+/uTL9tzpPt/JEwPOTd2bTHUgz/girn6GXXaiPq0hW+Se2kho740h1Sl9Y1q+RW36EqVnQrTT+chFf9p6hfrkf0pdDP76xV2gK6aDnsxfn8xOnhKwMXVXMtWBPO+LmLM+pZ6QmO4e2RD31EZCe2+M56e8FKUKteUbVDocgn7O5dXqk7n/1KfDYrbulICnfQuvIFMdyPN6n7A+5T0hMd09siHuqY2E9t408Xd9+74/ffv2TZ7aTLKxPvmf7wIffv7t85dfFs30+26hcyX8L08Szmy6kHFo+dcvkS02RdLvZBb7H5fARAZ2ekPNu91T9quQDdHlWO9JwqhQraVnOS0tzg3vot7JPODwWYnWwpTKdW1KOvMXnTqb0lv7cEXuLUe/WfCO8gC0wihW510XmtHX4KUnqdfuV+E+xKEjB1KkHfky3Zldmt5U+0V0PPa2PZfmGS68+FlbTsuxMFxPlmrkc558QmK6e2RD3FMbCe29acL3p21/I5GOyL888kF7VPT8jetT/lnEneCUjNw/X4WbHI5yeStc/x+//Zrdf+3U5rv37oiHK73PZsoC7z98fK+pylyVgZ3RUHW3m8r/+d//le72h4+fZAPdtlRxiuQLX3KE1+xYXpYm+yOW0/40d9IEHJTz+F1P8JKkbcrHEPN+SF7azCUNudFy7Wt79enDx+yNv//0wZ+rtMBgfxiG7mEelZuelt9/s0maa9fm4QOpsiHyaCe+Ndmcxos47cgl42mocO2MhysfX8Si6XtyzSNh6Gj6bDJPSEx3j2yIe2ojob23TChP2/48YqlP/4zIZ+0U0WOnhzUXf8TDhdFDnHTckP/+lOuo+GtQ9CRppYPyJvsjvt9mGZjrcDdqYGfsIg8mNvvK/qqnRFaLzXQ2f3Ut+8+N/ZGmzWXlqOtxUWUK4rQTYW4q2Mk0Q/AxQxmp1z5eUf7GzVC+hNUwQqLMFMTyli01YrYxD6+KZ5PN1ci9i7iKzEUcCjbHHlePnIgxbpvlBvZOyylhuCGfZOQZkk6WdkJiuntkQ9xTGwntvV30l3t7/75cpvzDk06Jih7WcX0S8UdTKM5rOOgVKQfl0koH8axn9+cUm7HHKQ/sjF20L3NFUPa3tBmbStOa7vbYcm9/zFA9uvTE3PHTx2KLcim8ZLElGeT3eu3bVxSbjWS9JYwiUcrcj5telpFWPJnNMp6a2uZq5L0lx6Hofehx9ciJGOO5eEeeVJ9UuTotp4ShmumzyTwhMd09siHuqY2E9t4qWp02fnXypFPy9evXdESSNC+V63endrm9IsvJLi5SoTOQ/AaaZpUODtqMPe7aDOyMXZi7vaK8JVlkOik81z+03NsfM1SPqiQd9yN8K0+pFF58nOYk5AEbqde+fUWx2apPG8LwjkJKFf2kPA7ASCueuBWHDuRq5L0lx6Hofeix3vZamkdCf8SU8PLdU0ne89NyShjOU5Zq5HOefEJiuntkQ9xTGwntvU38b/a2f3XypCMi6Odc6sMa+t2vJsJdykXvWzjNfu74qiTJb6BpVungoM28Z2Bn7MLc7RXlDclCRFODkpTHlnv7Y4bq0dgZfq8oyiZj5lJ48fGYk2B85VKvffuKYrP7/Wk1DL+lMl0mLqd0HICRVjxxK6p4tthcjby3ZKM29nigPtlmFWfztJwShgwJvZ6QmO4e2RD31EZCe28RV53K4hT/7Dz+QV/re1U6H/rBiDm7S7+cv3iC3fHNT3M8muFYl3NltHlq8xtomkU6OMGmiEtba3bGLqq7PVTekCzyO5/L2HJvf8xQPZrEqwmNHzKSNLx09jP1JKnXvn1FsdmqT9vC0Oz57n2xt+MAjLTiKbcixLPV5mrkvSXHoeT91CNnjIe9KgtPq44KWbE/IQxvyiac1BMS090jG+Ke2kho780RS1CJFqRlrFGg0vnoiR61/CCaq6sntTzf8Zrp3KjsJpYHOkmyKdM/hz9LTReyvE77bC43TZvpzgzsjIbKYFaU/dAgWcScVbIhDL+W0N/LZbqTdcAqIb+UaciINaXxFOGFGGpprn3jikTiFyDXbB+AtTByC0kGARgxa8/Da8ezweY4ch3VHfPBK24JJpiBx9xIT2z8WpBiM7j++Mv3n39cYmudlr1h+FW4ROFntX8gFgmJ6e6RLXJPbSS097ZoF6elHK1/fxpLOMGJeArD6B+/fNa/Ig0USaGY28+G4QII73/8j/DJ8eHn2O+ICWubTb3VHz66W+FZcoHKwE5zaAnSkWWWlrLexsBKkBb7w6wSjWRb3diffCezP0YX7PIHfxmhkpsqclnAZv8kg7Wvrujdl19c9oysH4BOGC5Ftra9GUAtlziQIoPIQ9nwuP8ToXxO2z5+EdHjliPXPMP5xGWBpYXmadkVhvZTnzYiW+Se2kho7/1gDsdbEn8fVlLw1aX+76rJnf/Q+o3fecXlqfLnDATpyXlPiy9P1KcVtBJRn+z5eDMyf31yP+NX117q0+fh72fOIrI56VsagozlvKfFlyfq0wpaiahP9ny8GZm/PvkIf8h/TX/przXx3wyE7m/nEETlQqdFLZrOJCEx3T2yRe6pjYT23g/mcLwNyX+VL8z8RSFLAcply0Z0R3FC1uVCp8XbpD6tIFvkntpIaO/9YA4HgiDIFSUkpruH+uQwhwNBEOSKEhLT3UN9cpjDgSAIckUJienuoT45zOFAEAS5ooTEdPdQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZ6dYnAACA6+LqkZYlAACAqaA+AQDAjFCfAABgPv788/8BAVd3+F5c4/IAAAAASUVORK5CYII="/> ``` Publishing to XHTML works with DITA OT 1.8. But publishing with DITA OT 2.x fails: >BUILD FAILED D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line: D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.base\build_preprocess.xml:438: java.net.MalformedURLException: unknown protocol: data at java.net.URL.<init>(Unknown Source) at java.net.URL.<init>(Unknown Source) at org.apache.tools.ant.types.resources.URLResource.getURL(URLResource.java:153) at org.apache.tools.ant.types.resources.URLResource.isExists(URLResource.java:226) at org.apache.tools.ant.types.resources.URLResource.isExists(URLResource.java:204) at org.apache.tools.ant.taskdefs.Copy.execute(Copy.java:519) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393) at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38) at org.apache.tools.ant.Project.executeTargets(Project.java:1248) at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441) at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) Such data referenced image resources should not be copied anywhere because they are already there.
1.0
Can no longer publish to XHTML image with data protocol [DOT 2.x develop branch] - In my DITA topic I refer to an embedded image like: ```xml <image href="data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAjUAAAC9CAIAAADECxtTAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAAAJcEhZcwAAEnQAABJ0Ad5mH3gAABp8SURBVHhe7Zxbjlw3kkC9K6nWI6n20EsQ5K7f+ehNGLBRCxkDPXZ9CGiggQa0APV4ZgBPkEHyksHHfWRmJbPyHARuJ8lgRJCXjKgsVfuHPwEAAObD1qcfAAAAro2rR1qWEtoLAABwLahPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNCfQIAgBmhPgEAwIxQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNygfr0/CjT9/H4HOZek07cjdgWzSkCBwB4i0iOdU9tJLT3VPYXquun+zpmG9PL00MYUShRAACXQBKse2ojob2nY3O5kKXzxqhw5XxfFSgTz3oBgxXkvT88vYTGZXgFFwBwaSS/uqc2Etp7OuP65Gl8ybpqZqE+XRq3gxd+xa/gAgAujuRX99RGQnvPwJZsXutcM7dU0awVVMrTLnT7LvqCX8EFALwCcpHdUxsJ7T0Dde1ppPNa6YrJpQpmGDDFaRfp+/Tl3u8ruACA10EusntqI6G9Z2BTfWr8FvB6iX9LfYJDZFt7qeLxCi4A4LWQi+ye2kho7xnYVJ8aWtfLLtSny1Du60Ve7yu4AIBXRC6ye2ojob1nYFN94vvTW6fa1PMXj1dwAQCvi1xk99RGQnvPwKb6VCsNkostZtvS0PZZg/rUiNORrahVaZNCObged8ddxXoBbYcVQrBujm1oP4z1ZbQ9tuZ1YjvmYvsSKuzUlT0z8QVtMbIyD+DOkevintpIaO8ZqDNHnQMa2aWdKJJiHE4do3u+c1YVjYmlGrcKrbRXT/K0l2lMVGFbelYslYGHp+cq1sCmDU1aWbyru9XAOAtzst7SSlt9THfO+hIMaWr1XoyPQBxeRvOAt747gLtE7oh7aiOhvWcgv4uKvZG1RvOid/LHMruZHQ7MquJZDXjDirrYuUI5vR220M6GQ5qBJTt2tBGakGnlEeQVtRFZPuzpRb8olu5NcKe4OLgER6bSfi92Xm85aUp7kwHAI3fEPbWR0N4zYPKK0L7YC83U0EkpQm7CXvZDs6qYBlaVVQWHKlVJdG1yOVwOdpJon0ZghX0bnY2tVCjdD4aEat3t2Eu13L0NvZ5/xMWeJRQxmNE8ul7UdjN1rNpiAFiQO+Ke2kho7xmweWVMO6WURsyNLpJKMXZsVh2xTSH7FXINOzhIkYLZEDPc2a4uq5GvuC/nm8nl4DjwTuiV1uJiMBTY5uL4EopBM7HwvYyV9pruqmUAwIJcG/fURkJ7z4C9om2aqSRSJh6rWo4u1/3YLKGK2KaQ0xSGg2VYgonbjlvPK1jf9baPNFacm+HS9trCAsZ9plQZqJa+xcXxJZRD1nQ5mqxW71rIp8r4zjcIcF/IlXFPbSS09wzUV3T3jSxNjFPDMnpsllBFbAM+TWE4aBJkZdqO22WtYH3X0wcBnBRbNbkXeaa4761scXHCEkr/1nQ5MY1W/iJV8ADQQm6Le2ojob1noMoru+9mbWFEtH5slmM1E56mMBzcm0Ct5xWs7zqDD1TWJg9jrwYbrjtUG6ZUS9/g4vgSOkF0SNMqgxnbdwDgbpGb4p7aSGjvGagv9s6cai1svNbHZjmqiG3ApymMZ9uEZuI2wyduZWtXuuGtpv+hwursFikWiaEyUK193cUJSzDbsiV6pTJZsPcFAtwZckvcUxsJ7T0DVb47Oalum39slqOK2E49TWFtthkvR4vB7UkyYn03LHTDOyG5C6uzSzL1oFgZsJu+wcWqRl/BbkvlfUS1pzm7LAHcGXJF3FMbCe09A/Xd3HshuxlzyLFZjtWZpymsB1Zq9OZW0zZgfTeKRF9lbfKwgqwXj0ShmmwMrXs2uDi+BDtz9/ZXBhL9nQC4e+SGuKc2Etp7Bup7ufdmV0ljk4Vjsxyrqeg0hdXZnlwrKix9R1Oa9d2wY1WW6OyOriX3Yrx6HZ0llHrZ1lQGqm3b4GIYotAfr4y3X9uY2ojngCWAO0EuiHtqI6G9Z6BKx/uvY22ik9xEcxk5NkuoJtqAT1NYnb3QyGe9NWzD+q6tjTK4GbOTjfFyuFpKcyFGK9+ZykC1bVtcHF9C9draKxCy42ROlme7IQCQC+Ke2kho7xmor2M/H/eobfRyz7AQbJrVmGcDPk1hdbZD8+jZE5f1XTsYJnAzaAIvjRvTZmo+LkPBklXKbVQGqm074mLHEqr3Vq3R4ewvRsvWQmmrpQEADrkg7qmNhPaeTpU0mrd6lUZyKA15BXvRzzTrvAqrs6PGkX1awfquXIzTtxkvZ5dTe4taCLOlP9mpdJKZeqSxPVtcnHUJjtyCV8inqcHKUOmpMQwAitwQ99RGQntPp3WnD13IZnLIqfOVcGRWNcfGe5rC2uxs/NBGjah870vQjlwln9zrj5SmcxY3fZ3uf2i9F0JJuZKjS2jtn8FMSxbtVi6u2o4AwCN3xD21kdDeU+nd52OJd5AdBrd876yGfhnuqkIjTy4K9ewyikG4hgObaI1r2o+GirgHO5rpRa2sqx9Wc2mleksl+Gjsah3jBheeo0sYvp/GluXa2XDsbswAgAy5Ju6pjYT2HmRwhw2jTNCmSlOb7vi2WcO4H55+X1MQk40sGpDh4eCWGGo2rT5hjevsdu8a9WK2zCt8Nd9+rmEsFi67R2fdReTYEoRqYm+exBIisFM2egK4b+SuuKc2EtoLr06dM9fYUeOPVSIAgOsgaco9tZHQXrgGtoqssb1AUZ8A4JaQNOWe2khoL1yNHUVqe5WhPgHALSFpyj21kdBeeH2WGjL6XpRXGuoTALxNJE25pzYS2guvTaog66UjqfL7PQB4m0iack9tJLQXXpnlryM216c9NYb6BAC3hKQp99RGQnvhtcn+fG/8tSgo7qsw1CcAuCUkTbmnNhLaC1cgLyLNGhVr2O7qYsuTsP13gwAAr40kKffURkJ74XoM/o9Q+4vKwJiHMgUAEyLZyT21kdBeAACAa0F9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8AADAj1CcAAJgR6hMAAMwI9QkAAGaE+gQAADNCfQIAgBmhPgEAwIxQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZoT4BAMCMUJ8AAGBGqE8T8fwoG//w9BKa2zk88Q54eXqQzXl8Ds1d7NhXXgHA2ZFL5Z7aSGjvLePSxVmyhc87CZfmyp7kRPOg54jjZHXv5MMT74Sj9Sm9z/V93aEKANuRS+We2kho702SVY9zZYuQfrIUF3oqB67/2E/qHh/71qhfnh6T5q6JFyKPZ8x2zWvj3/O2fd2hevfczgGAKyN5zT21kdDemyOUjcjebCFTwidLnf+1DNpa9Px4Uobak+MkgEVzguRYxDNku+bVoT5dggsegP4VhptEX+gbqU8BrR37s4XOCo0Sb7Kwp7WwLFAnlqcdOc5oXj05Ho58bnYEe1PruiaX2yh/falPbwp9odQnh0z59u2bPEM7R23Wv+LLC1RVnmIgnqKULSOPzy9PT2EoXd04bL+fKYVZr78yMcTq6W+Kn5sbqf3nnhdDdTy2M9ppavbMBrrBL7PyPaxIexPw01w7Gq48LmRzlzD8YlIzzs7d5JH5wXz60GGh5yYXO6bWKtcNHY8qSkeY4Ye8cm7ZLC+bv3C2qIrOpJqbzzZHu2VqGG+GFpHx7uWFm0VfKPXJIVP+5ZEPoWshXZbUFPIecVs49FEU9zK7joumG4la3ocz6zsKAxkvL1lBCl2DiZWD1r74KY6Hx8cH1VbdxY5vl4a02YiniMA3/FhDs2/W0w1eGqW7ZU5GsJdmLcuUvfJ9RaQVfn4WbKldjJaqopgZzePTkDKbFZWG91rMEJXYLkJaNNWI41He6NOLG3n/l7+ETunTOTpBNNRcYazkPFE1DoAbtZvzezP+wlmBqPVvLtww+kKpTw6Z8s+IfA69Eb0x4SZJw/3Qbnv8J8UPpggKVXPXzPen8rJ21mDG+hPlc2FiyRUVaiQby636z8UCdZdDV+6y2eqsamxWRotQs+C7e1hROjRxV6MldrQMtxhdGu5TviRpFw6yVXSwGoUjQdrt+H1rcV4uVWmplPN7oZ0pqmqs8Fc4acVfIyrh0rauLdw0+kKpTw6Z8g+PfAhdOdk1k4/+f5fbFHsa6DxH1OgF6DVTZ9kqMWPdif5jTTPU2p/Odsr1WJlKWuOOxX90WWiOzS6TC9TS9pfsNDOt0mcrggU76ttp84rR0Hiq/u3fD9R4K3ERgbblZCEOt/+Zc/GTzHjzqaWUprXZ8WspR6O/vVEVVpbhgqDbir+JqI1uLtws+kKpTw6Z8vXrV3mGtkUvk9wX+ZDfIPGz9ORoIC6MNDWSXcwlTN+ZmmWrxIx1J7qPHRMVtT/f46OO6wwDnmW0Nbe39kJzbNZ9MoMFqukZqTkn2XgRgGlZ7Khvp7dYjPqGUppzAz0HukmJZDmMaNsn/szX82Om5+htdW4kUQQ9XFGDc0RV+HCNvr9W/D1Ec3h54SbRF0p9cuis0GjiLbt/eEqWtSf9o81CedXLK5rQ7qTXmNNZgxnrTvQfN97w2t/S4z8ZQ7l+ObfR6qenvln/aTV4tZH5s7hXlA0WAZiWxY6WERWjsVEZLOdsRg/W8pNPZj83VrrzrcWZt2FclxNMdGaw5gxRFWO+YSJcaMU/QJSF0IA3gb7Qe6hPektG510Gw6cuajo3HJxZq+UNLa9o+aNmdgnLScZEQct8c6K3boxUP+oqtT83OXT4we5oOdcY0qnRZTk4NjsIvruHNXmcQiMA42LBjvp28lOMLg3/KYtmsIoRajD7WcjbMZaKECrXrW1pzWivqIlqnBJVOTjcnFb8Y0Q/fII3gb7QN1Wf9Do48qPd7t2NuzLldXKGzQUT1F3sD5dYPD8//fX30obTDCH5SWmkbJUsY/4vA0YTg+/QdmMjmyaPZJuldgqzi6HF5cvTX//mZ8bBfO3SlWu6v2kYmo2TQ08WfO1+idTgbSyjSwB1y1KNZrZCaA7Xkav6z5nL7irGmMCrtqCeorXgR1T8Vrcm2CWVKmawyalRLT7yAxC13WgKoGHa2IY3jrxr99RGQntvDj27hniU46A573up/xm4819ryYN5fA630Ht31zS2hRBQ1pPpO5oRR4WHp7+tTixD0b4ar+V+VRk06yRQ7K8xtMQjk/prt5qOkdle8O09rMh3w/nL22ubbOd6lmjC32+WfwwtiGrWjmbbq1jBnjUxUs0tDceY4x+XBuKscknPpcp4NxZOicoPxmZjUxfPzfgF7S8jgDeLvGv31EZCe+Gu8Defiw8As0B9ggD1CQCmgvoEAeoTAEwF9QmE4rf91CgAmAFJR+6pjYT2AgAAXAvqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9Qnx7//5/8QBEEmkZCY7h7qk8McDgRBkCtKSEx3D/XJYQ4HgiDIFSUkpruH+uQwhwNBEOSiIjlWMJ1JQmK6e7QSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+UAQBLmc+PJEfVpBKxH1yZ4PBEGQy4kvT9SnFbQSUZ/s+ZhBfvoo7+H957//r+m/unz/45cPLrQff/1jutiacnMBzynTHsgDcvUjIc4F05kkJKa7R7bIPbWR0N6b5PnRv3fHw9NL6FzDHI4k3//47fP7YM3z6ad4mr///Cn0Cec+5eHyeNOzpYPvf//xXQjtNtL9zQU8ocx8IA/IDEdC/ZvOJCEx3T2yRe6pjYT23hwvTw/+pWc8PoexIeZwGNFS9O7Lb7bfVa8zX9dfv3xKBif+/uTL9tzpPt/JEwPOTd2bTHUgz/girn6GXXaiPq0hW+Se2kho740h1Sl9Y1q+RW36EqVnQrTT+chFf9p6hfrkf0pdDP76xV2gK6aDnsxfn8xOnhKwMXVXMtWBPO+LmLM+pZ6QmO4e2RD31EZCe2+M56e8FKUKteUbVDocgn7O5dXqk7n/1KfDYrbulICnfQuvIFMdyPN6n7A+5T0hMd09siHuqY2E9t408Xd9+74/ffv2TZ7aTLKxPvmf7wIffv7t85dfFs30+26hcyX8L08Szmy6kHFo+dcvkS02RdLvZBb7H5fARAZ2ekPNu91T9quQDdHlWO9JwqhQraVnOS0tzg3vot7JPODwWYnWwpTKdW1KOvMXnTqb0lv7cEXuLUe/WfCO8gC0wihW510XmtHX4KUnqdfuV+E+xKEjB1KkHfky3Zldmt5U+0V0PPa2PZfmGS68+FlbTsuxMFxPlmrkc558QmK6e2RD3FMbCe29acL3p21/I5GOyL888kF7VPT8jetT/lnEneCUjNw/X4WbHI5yeStc/x+//Zrdf+3U5rv37oiHK73PZsoC7z98fK+pylyVgZ3RUHW3m8r/+d//le72h4+fZAPdtlRxiuQLX3KE1+xYXpYm+yOW0/40d9IEHJTz+F1P8JKkbcrHEPN+SF7azCUNudFy7Wt79enDx+yNv//0wZ+rtMBgfxiG7mEelZuelt9/s0maa9fm4QOpsiHyaCe+Ndmcxos47cgl42mocO2MhysfX8Si6XtyzSNh6Gj6bDJPSEx3j2yIe2ojob23TChP2/48YqlP/4zIZ+0U0WOnhzUXf8TDhdFDnHTckP/+lOuo+GtQ9CRppYPyJvsjvt9mGZjrcDdqYGfsIg8mNvvK/qqnRFaLzXQ2f3Ut+8+N/ZGmzWXlqOtxUWUK4rQTYW4q2Mk0Q/AxQxmp1z5eUf7GzVC+hNUwQqLMFMTyli01YrYxD6+KZ5PN1ci9i7iKzEUcCjbHHlePnIgxbpvlBvZOyylhuCGfZOQZkk6WdkJiuntkQ9xTGwntvV30l3t7/75cpvzDk06Jih7WcX0S8UdTKM5rOOgVKQfl0koH8axn9+cUm7HHKQ/sjF20L3NFUPa3tBmbStOa7vbYcm9/zFA9uvTE3PHTx2KLcim8ZLElGeT3eu3bVxSbjWS9JYwiUcrcj5telpFWPJnNMp6a2uZq5L0lx6Hofehx9ciJGOO5eEeeVJ9UuTotp4ShmumzyTwhMd09siHuqY2E9t4qWp02fnXypFPy9evXdESSNC+V63endrm9IsvJLi5SoTOQ/AaaZpUODtqMPe7aDOyMXZi7vaK8JVlkOik81z+03NsfM1SPqiQd9yN8K0+pFF58nOYk5AEbqde+fUWx2apPG8LwjkJKFf2kPA7ASCueuBWHDuRq5L0lx6Hofeix3vZamkdCf8SU8PLdU0ne89NyShjOU5Zq5HOefEJiuntkQ9xTGwntvU38b/a2f3XypCMi6Odc6sMa+t2vJsJdykXvWzjNfu74qiTJb6BpVungoM28Z2Bn7MLc7RXlDclCRFODkpTHlnv7Y4bq0dgZfq8oyiZj5lJ48fGYk2B85VKvffuKYrP7/Wk1DL+lMl0mLqd0HICRVjxxK6p4tthcjby3ZKM29nigPtlmFWfztJwShgwJvZ6QmO4e2RD31EZCe28RV53K4hT/7Dz+QV/re1U6H/rBiDm7S7+cv3iC3fHNT3M8muFYl3NltHlq8xtomkU6OMGmiEtba3bGLqq7PVTekCzyO5/L2HJvf8xQPZrEqwmNHzKSNLx09jP1JKnXvn1FsdmqT9vC0Oz57n2xt+MAjLTiKbcixLPV5mrkvSXHoeT91CNnjIe9KgtPq44KWbE/IQxvyiac1BMS090jG+Ke2kho780RS1CJFqRlrFGg0vnoiR61/CCaq6sntTzf8Zrp3KjsJpYHOkmyKdM/hz9LTReyvE77bC43TZvpzgzsjIbKYFaU/dAgWcScVbIhDL+W0N/LZbqTdcAqIb+UaciINaXxFOGFGGpprn3jikTiFyDXbB+AtTByC0kGARgxa8/Da8ezweY4ch3VHfPBK24JJpiBx9xIT2z8WpBiM7j++Mv3n39cYmudlr1h+FW4ROFntX8gFgmJ6e6RLXJPbSS097ZoF6elHK1/fxpLOMGJeArD6B+/fNa/Ig0USaGY28+G4QII73/8j/DJ8eHn2O+ICWubTb3VHz66W+FZcoHKwE5zaAnSkWWWlrLexsBKkBb7w6wSjWRb3diffCezP0YX7PIHfxmhkpsqclnAZv8kg7Wvrujdl19c9oysH4BOGC5Ftra9GUAtlziQIoPIQ9nwuP8ToXxO2z5+EdHjliPXPMP5xGWBpYXmadkVhvZTnzYiW+Se2kho7/1gDsdbEn8fVlLw1aX+76rJnf/Q+o3fecXlqfLnDATpyXlPiy9P1KcVtBJRn+z5eDMyf31yP+NX117q0+fh72fOIrI56VsagozlvKfFlyfq0wpaiahP9ny8GZm/PvkIf8h/TX/przXx3wyE7m/nEETlQqdFLZrOJCEx3T2yRe6pjYT23g/mcLwNyX+VL8z8RSFLAcply0Z0R3FC1uVCp8XbpD6tIFvkntpIaO/9YA4HgiDIFSUkpruH+uQwhwNBEOSKEhLT3UN9cpjDgSAIckUJienuoT45zOFAEAS5ooTEdPdQnwAAYEaoTwAAMCPUJwAAmBHqEwAAzAj1CQAAZoT6BAAAM0J9AgCAGaE+AQDAjFCfAABgRqhPAAAwI9QnAACYEeoTAADMCPUJAABmhPoEAAAzQn0CAIAZ6dYnAACA6+LqkZYlAACAqaA+AQDAjFCfAABgPv788/8BAVd3+F5c4/IAAAAASUVORK5CYII="/> ``` Publishing to XHTML works with DITA OT 1.8. But publishing with DITA OT 2.x fails: >BUILD FAILED D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line: D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.base\build_preprocess.xml:438: java.net.MalformedURLException: unknown protocol: data at java.net.URL.<init>(Unknown Source) at java.net.URL.<init>(Unknown Source) at org.apache.tools.ant.types.resources.URLResource.getURL(URLResource.java:153) at org.apache.tools.ant.types.resources.URLResource.isExists(URLResource.java:226) at org.apache.tools.ant.types.resources.URLResource.isExists(URLResource.java:204) at org.apache.tools.ant.taskdefs.Copy.execute(Copy.java:519) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) at org.apache.tools.ant.Task.perform(Task.java:348) at org.apache.tools.ant.Target.execute(Target.java:435) at org.apache.tools.ant.Target.performTasks(Target.java:456) at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1393) at org.apache.tools.ant.helper.SingleCheckExecutor.executeTargets(SingleCheckExecutor.java:38) at org.apache.tools.ant.Project.executeTargets(Project.java:1248) at org.apache.tools.ant.taskdefs.Ant.execute(Ant.java:441) at org.apache.tools.ant.taskdefs.CallTarget.execute(CallTarget.java:105) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106) Such data referenced image resources should not be copied anywhere because they are already there.
process
can no longer publish to xhtml image with data protocol in my dita topic i refer to an embedded image like xml publishing to xhtml works with dita ot but publishing with dita ot x fails build failed d projects exml frameworks dita dita x build xml the following error occurred while executing this line d projects exml frameworks dita dita x plugins org dita base build preprocess xml java net malformedurlexception unknown protocol data at java net url unknown source at java net url unknown source at org apache tools ant types resources urlresource geturl urlresource java at org apache tools ant types resources urlresource isexists urlresource java at org apache tools ant types resources urlresource isexists urlresource java at org apache tools ant taskdefs copy execute copy java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache tools ant dispatch dispatchutils execute dispatchutils java at org apache tools ant task perform task java at org apache tools ant target execute target java at org apache tools ant target performtasks target java at org apache tools ant project executesortedtargets project java at org apache tools ant helper singlecheckexecutor executetargets singlecheckexecutor java at org apache tools ant project executetargets project java at org apache tools ant taskdefs ant execute ant java at org apache tools ant taskdefs calltarget execute calltarget java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org apache tools ant dispatch dispatchutils execute dispatchutils java such data referenced image resources should not be copied anywhere because they are already there
1
18,016
24,032,773,087
IssuesEvent
2022-09-15 16:18:26
googleapis/java-beyondcorp-appconnections
https://api.github.com/repos/googleapis/java-beyondcorp-appconnections
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'beyondcorp-appconnections' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'beyondcorp-appconnections' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname beyondcorp appconnections invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
18,419
10,962,706,858
IssuesEvent
2019-11-27 17:52:15
microsoft/BotFramework-Composer
https://api.github.com/repos/microsoft/BotFramework-Composer
closed
Initialize Property for metadata usage in QNA maker
Bot Services customer-replied-to customer-reported
Is there any way currently to initialize the property as the metadata from QNA to retain context. Without composer I would use the Post json response of qna to retain context. This holds good for Follow-up prompts as well. Any suggestions? Other way I tried is by using HTTP request rather than QNA to have post based logic, but it had some issues. I will continue to explore that.
1.0
Initialize Property for metadata usage in QNA maker - Is there any way currently to initialize the property as the metadata from QNA to retain context. Without composer I would use the Post json response of qna to retain context. This holds good for Follow-up prompts as well. Any suggestions? Other way I tried is by using HTTP request rather than QNA to have post based logic, but it had some issues. I will continue to explore that.
non_process
initialize property for metadata usage in qna maker is there any way currently to initialize the property as the metadata from qna to retain context without composer i would use the post json response of qna to retain context this holds good for follow up prompts as well any suggestions other way i tried is by using http request rather than qna to have post based logic but it had some issues i will continue to explore that
0
16,197
20,681,978,884
IssuesEvent
2022-03-10 14:41:16
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
GDAL Dissolve with temporal layers misses geometries
Feedback Processing Bug
### What is the bug or the crash? When doing a Dissolve using the GDAL library, the process missess some geometries. It is reproduced with a temporal layer, but if you save it in a Shapefile, it works well. ### Steps to reproduce the issue [Data.zip](https://github.com/qgis/QGIS/files/8222662/Data.zip) 1) We do an intersect between the SeccCensales Layer and the Vias Ciclistas layer, in order to get the 'Vias Ciclistas' with the SECCION attribute. Result is a temporary layer ![image](https://user-images.githubusercontent.com/101328861/157647430-3dce978b-5321-4d71-8daa-e5744438eb19.png) 2) We do a dissolve with the GDAL, setting the 'SECCION' attribute as the dissolve-attribute. The ![image](https://user-images.githubusercontent.com/101328861/157647620-bf7e40b8-6ee5-4472-bf1e-79922b65bd71.png) 3)Result missing some geometries (red is the result of step1, green is the dissolved result illustrating missing geometries) ![image](https://user-images.githubusercontent.com/101328861/157648047-ffb28bf2-e4d4-40c5-b164-905c4f34887f.png) ### Versions 3.16.14 version ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
1.0
GDAL Dissolve with temporal layers misses geometries - ### What is the bug or the crash? When doing a Dissolve using the GDAL library, the process missess some geometries. It is reproduced with a temporal layer, but if you save it in a Shapefile, it works well. ### Steps to reproduce the issue [Data.zip](https://github.com/qgis/QGIS/files/8222662/Data.zip) 1) We do an intersect between the SeccCensales Layer and the Vias Ciclistas layer, in order to get the 'Vias Ciclistas' with the SECCION attribute. Result is a temporary layer ![image](https://user-images.githubusercontent.com/101328861/157647430-3dce978b-5321-4d71-8daa-e5744438eb19.png) 2) We do a dissolve with the GDAL, setting the 'SECCION' attribute as the dissolve-attribute. The ![image](https://user-images.githubusercontent.com/101328861/157647620-bf7e40b8-6ee5-4472-bf1e-79922b65bd71.png) 3)Result missing some geometries (red is the result of step1, green is the dissolved result illustrating missing geometries) ![image](https://user-images.githubusercontent.com/101328861/157648047-ffb28bf2-e4d4-40c5-b164-905c4f34887f.png) ### Versions 3.16.14 version ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
process
gdal dissolve with temporal layers misses geometries what is the bug or the crash when doing a dissolve using the gdal library the process missess some geometries it is reproduced with a temporal layer but if you save it in a shapefile it works well steps to reproduce the issue we do an intersect between the secccensales layer and the vias ciclistas layer in order to get the vias ciclistas with the seccion attribute result is a temporary layer we do a dissolve with the gdal setting the seccion attribute as the dissolve attribute the result missing some geometries red is the result of green is the dissolved result illustrating missing geometries versions version supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
74,029
14,169,751,921
IssuesEvent
2020-11-12 13:40:02
mentaLwz/gitblogOfMental
https://api.github.com/repos/mentaLwz/gitblogOfMental
opened
922. 按奇偶排序数组 II
Leetcode2020
```go func sortArrayByParityII(A []int) []int { l := len(A) e := 0 o := 1 a := make([]int, l) for i := 0; i< l; i++ { if A[i] %2 == 0 { a[e] = A[i] e += 2 } else { a[o] = A[i] o += 2 } } return a } ``` ```rust impl Solution { pub fn sort_array_by_parity_ii(a: Vec<i32>) -> Vec<i32> { let mut array = vec![0;a.len()]; let mut o = 1; let mut e = 0; for i in a { if i % 2 == 0 { array[e] = i; e += 2; } else { array[o] = i; o += 2; } } array } } ``` ```python class Solution: def sortArrayByParityII(self, A: List[int]) -> List[int]: even = [] odd = [] for i in range(len(A)): if A[i] % 2 == 0: even.append(A[i]) else: odd.append(A[i]) # for i in range(len(A)): # if i % 2 == 0: # A[i] = even.pop() # else: # A[i] = odd.pop() # ou = [i for i in A if i % 2] # ji = [i for i in A if not i % 2] # return [i for n in zip(ji, ou) for i in n] return [i for n in zip(even, odd) for i in n] ```
1.0
922. 按奇偶排序数组 II - ```go func sortArrayByParityII(A []int) []int { l := len(A) e := 0 o := 1 a := make([]int, l) for i := 0; i< l; i++ { if A[i] %2 == 0 { a[e] = A[i] e += 2 } else { a[o] = A[i] o += 2 } } return a } ``` ```rust impl Solution { pub fn sort_array_by_parity_ii(a: Vec<i32>) -> Vec<i32> { let mut array = vec![0;a.len()]; let mut o = 1; let mut e = 0; for i in a { if i % 2 == 0 { array[e] = i; e += 2; } else { array[o] = i; o += 2; } } array } } ``` ```python class Solution: def sortArrayByParityII(self, A: List[int]) -> List[int]: even = [] odd = [] for i in range(len(A)): if A[i] % 2 == 0: even.append(A[i]) else: odd.append(A[i]) # for i in range(len(A)): # if i % 2 == 0: # A[i] = even.pop() # else: # A[i] = odd.pop() # ou = [i for i in A if i % 2] # ji = [i for i in A if not i % 2] # return [i for n in zip(ji, ou) for i in n] return [i for n in zip(even, odd) for i in n] ```
non_process
按奇偶排序数组 ii go func sortarraybyparityii a int int l len a e o a make int l for i i l i if a a a e else a a o return a rust impl solution pub fn sort array by parity ii a vec vec let mut array vec let mut o let mut e for i in a if i array i e else array i o array python class solution def sortarraybyparityii self a list list even odd for i in range len a if a even append a else odd append a for i in range len a if i a even pop else a odd pop ou ji return return
0
4,852
7,742,888,913
IssuesEvent
2018-05-29 11:00:10
ethereumjs/ethereumjs-client
https://api.github.com/repos/ethereumjs/ethereumjs-client
opened
Early on VM tests
Block Processing / VM External
#### Description Thanks to the work of @vpulim the ``ethereumjs-blockchain`` library is now compatible with Geth chain DBs starting with the [v3.x](https://github.com/ethereumjs/ethereumjs-blockchain/releases/tag/v3.0.0) release series. This can (and should 😛) be used to run the VM on a post-Byzantium synced Geth chain DB and process actual mainnet transactions to see how early on how the VM behave on real-world conditions and improve/move forward from there. To get to a completely mainnet-compliant VM will probably take some time.
1.0
Early on VM tests - #### Description Thanks to the work of @vpulim the ``ethereumjs-blockchain`` library is now compatible with Geth chain DBs starting with the [v3.x](https://github.com/ethereumjs/ethereumjs-blockchain/releases/tag/v3.0.0) release series. This can (and should 😛) be used to run the VM on a post-Byzantium synced Geth chain DB and process actual mainnet transactions to see how early on how the VM behave on real-world conditions and improve/move forward from there. To get to a completely mainnet-compliant VM will probably take some time.
process
early on vm tests description thanks to the work of vpulim the ethereumjs blockchain library is now compatible with geth chain dbs starting with the release series this can and should 😛 be used to run the vm on a post byzantium synced geth chain db and process actual mainnet transactions to see how early on how the vm behave on real world conditions and improve move forward from there to get to a completely mainnet compliant vm will probably take some time
1
97,605
16,236,396,570
IssuesEvent
2021-05-07 01:38:05
michaeldotson/auth-app
https://api.github.com/repos/michaeldotson/auth-app
opened
CVE-2020-8164 (High) detected in actionpack-5.2.2.gem
security vulnerability
## CVE-2020-8164 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionpack-5.2.2.gem</b></p></summary> <p>Web apps on Rails. Simple, battle-tested conventions for building and testing MVC web applications. Works with any Rack-compatible server.</p> <p>Library home page: <a href="https://rubygems.org/gems/actionpack-5.2.2.gem">https://rubygems.org/gems/actionpack-5.2.2.gem</a></p> <p>Path to dependency file: /auth-app/Gemfile.lock</p> <p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/actionpack-5.2.2.gem</p> <p> Dependency Hierarchy: - sass-rails-5.0.7.gem (Root Library) - sprockets-rails-3.2.1.gem - :x: **actionpack-5.2.2.gem** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A deserialization of untrusted data vulnerability exists in rails < 5.2.4.3, rails < 6.0.3.1 which can allow an attacker to supply information can be inadvertently leaked fromStrong Parameters. <p>Publish Date: 2020-06-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8164>CVE-2020-8164</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-8727-m6gj-mc37">https://github.com/advisories/GHSA-8727-m6gj-mc37</a></p> <p>Release Date: 2020-05-31</p> <p>Fix Resolution: 5.2.4.3,6.0.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8164 (High) detected in actionpack-5.2.2.gem - ## CVE-2020-8164 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionpack-5.2.2.gem</b></p></summary> <p>Web apps on Rails. Simple, battle-tested conventions for building and testing MVC web applications. Works with any Rack-compatible server.</p> <p>Library home page: <a href="https://rubygems.org/gems/actionpack-5.2.2.gem">https://rubygems.org/gems/actionpack-5.2.2.gem</a></p> <p>Path to dependency file: /auth-app/Gemfile.lock</p> <p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/actionpack-5.2.2.gem</p> <p> Dependency Hierarchy: - sass-rails-5.0.7.gem (Root Library) - sprockets-rails-3.2.1.gem - :x: **actionpack-5.2.2.gem** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A deserialization of untrusted data vulnerability exists in rails < 5.2.4.3, rails < 6.0.3.1 which can allow an attacker to supply information can be inadvertently leaked fromStrong Parameters. <p>Publish Date: 2020-06-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8164>CVE-2020-8164</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-8727-m6gj-mc37">https://github.com/advisories/GHSA-8727-m6gj-mc37</a></p> <p>Release Date: 2020-05-31</p> <p>Fix Resolution: 5.2.4.3,6.0.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in actionpack gem cve high severity vulnerability vulnerable library actionpack gem web apps on rails simple battle tested conventions for building and testing mvc web applications works with any rack compatible server library home page a href path to dependency file auth app gemfile lock path to vulnerable library var lib gems cache actionpack gem dependency hierarchy sass rails gem root library sprockets rails gem x actionpack gem vulnerable library vulnerability details a deserialization of untrusted data vulnerability exists in rails rails which can allow an attacker to supply information can be inadvertently leaked fromstrong parameters publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
128,306
17,475,859,579
IssuesEvent
2021-08-08 05:28:19
reconness/reconness-frontend
https://api.github.com/repos/reconness/reconness-frontend
closed
Delete pipeline
story points: 3 discussion design
We need an option to delete a pipeline from the Pipeline list (in Mosaic and List mode) and from a Pipeline Details page. When the user clicks the Delete option, a confirmation popup will appear. Confirmation popup will be similar to but the reference will be for a Pipeline ![image](https://user-images.githubusercontent.com/4720000/119244042-1e893700-bb32-11eb-90dd-10f670767489.png)
1.0
Delete pipeline - We need an option to delete a pipeline from the Pipeline list (in Mosaic and List mode) and from a Pipeline Details page. When the user clicks the Delete option, a confirmation popup will appear. Confirmation popup will be similar to but the reference will be for a Pipeline ![image](https://user-images.githubusercontent.com/4720000/119244042-1e893700-bb32-11eb-90dd-10f670767489.png)
non_process
delete pipeline we need an option to delete a pipeline from the pipeline list in mosaic and list mode and from a pipeline details page when the user clicks the delete option a confirmation popup will appear confirmation popup will be similar to but the reference will be for a pipeline
0
6,191
9,104,337,581
IssuesEvent
2019-02-20 17:54:47
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
ASDs added into ABM with IPs that are not served by ASD server
process_wontfix
``` nuvolat@NY2SRV0021:~$ alba dev-extract-config --config ... { "ips": [ "172.17.7.33" ], "asd_id": "EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v", "node_id": "LWcjrcPQVop59pQq15NKSe53QhypOR1E", "capacity": 3998830551040, "home": "/mnt/alba-asd/8EytFWg0OLYUYM9d/EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v", "log_level": "info", "rocksdb_block_cache_size": 976276990, "port": 8607, "transport": "tcp" } ```` So the asd only listens on `172.17.7.33` but the ABM tells us ``` 0 : { kind = (Asd ((["172.17.23.33"; "172.17.7.33"], 8607, false, false), "EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v")); decommissioned = false; node_id = "LWcjrcPQVop59pQq15NKSe53QhypOR1E"; other = "{ \"id\" : \"EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v\", \"node_id\" : \"LWcjrcPQVop59pQq15NKSe53QhypOR1E\", \"port\" : 8607, \"used_bytes\" : \"3722378166560\", \"total_bytes\" : \"3998830551040\", \"version\" : \"AsdV1\", \"network_interfaces\":[{\"ipv4_addr\": \"172.17.7.33\"} ]}"; total = 3998830551040L; used = 3723897345607L; ``` The `172.17.23.33` ip address will never be usable. Alba works around this, but constantly nags with `Connection refused`. Something went wrong during installation/configuration
1.0
ASDs added into ABM with IPs that are not served by ASD server - ``` nuvolat@NY2SRV0021:~$ alba dev-extract-config --config ... { "ips": [ "172.17.7.33" ], "asd_id": "EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v", "node_id": "LWcjrcPQVop59pQq15NKSe53QhypOR1E", "capacity": 3998830551040, "home": "/mnt/alba-asd/8EytFWg0OLYUYM9d/EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v", "log_level": "info", "rocksdb_block_cache_size": 976276990, "port": 8607, "transport": "tcp" } ```` So the asd only listens on `172.17.7.33` but the ABM tells us ``` 0 : { kind = (Asd ((["172.17.23.33"; "172.17.7.33"], 8607, false, false), "EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v")); decommissioned = false; node_id = "LWcjrcPQVop59pQq15NKSe53QhypOR1E"; other = "{ \"id\" : \"EdyVBJQRAqhZb31wlp5XqDU6vfMhL30v\", \"node_id\" : \"LWcjrcPQVop59pQq15NKSe53QhypOR1E\", \"port\" : 8607, \"used_bytes\" : \"3722378166560\", \"total_bytes\" : \"3998830551040\", \"version\" : \"AsdV1\", \"network_interfaces\":[{\"ipv4_addr\": \"172.17.7.33\"} ]}"; total = 3998830551040L; used = 3723897345607L; ``` The `172.17.23.33` ip address will never be usable. Alba works around this, but constantly nags with `Connection refused`. Something went wrong during installation/configuration
process
asds added into abm with ips that are not served by asd server nuvolat alba dev extract config config ips asd id node id capacity home mnt alba asd log level info rocksdb block cache size port transport tcp so the asd only listens on but the abm tells us kind asd false false decommissioned false node id other id node id port used bytes total bytes version network interfaces total used the ip address will never be usable alba works around this but constantly nags with connection refused something went wrong during installation configuration
1
133,729
18,948,595,403
IssuesEvent
2021-11-18 12:59:02
Lambda-IT/parlamentsdienste-components
https://api.github.com/repos/Lambda-IT/parlamentsdienste-components
closed
PD-Datepicker Error
bug design
If the PD-Datepicker is set to not have user_input the error state is not correctly displayed (no red border) Not urgent
1.0
PD-Datepicker Error - If the PD-Datepicker is set to not have user_input the error state is not correctly displayed (no red border) Not urgent
non_process
pd datepicker error if the pd datepicker is set to not have user input the error state is not correctly displayed no red border not urgent
0
18,980
24,968,680,959
IssuesEvent
2022-11-01 21:58:56
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Update OS packages in image
enhancement security process
### Problem The container images frequently contains out of date packages from the base image that may contain vulnerabilities. We should keep them up to date even if there's not an updated base image to consume. ### Solution * Update base image for all mirror node components * Add an extra command to update all OS packages so they at least are up to date as of each tag ``` RUN apt-get update \ && DEBIAN_FRONTEND=noninteractive apt-get install -y --no-install-recommends && rm -rf /var/lib/apt/lists/* ``` or ``` RUN apk update && apk upgrade && rm -rf /var/cache/apk/* ``` ### Alternatives _No response_
1.0
Update OS packages in image - ### Problem The container images frequently contains out of date packages from the base image that may contain vulnerabilities. We should keep them up to date even if there's not an updated base image to consume. ### Solution * Update base image for all mirror node components * Add an extra command to update all OS packages so they at least are up to date as of each tag ``` RUN apt-get update \ && DEBIAN_FRONTEND=noninteractive apt-get install -y --no-install-recommends && rm -rf /var/lib/apt/lists/* ``` or ``` RUN apk update && apk upgrade && rm -rf /var/cache/apk/* ``` ### Alternatives _No response_
process
update os packages in image problem the container images frequently contains out of date packages from the base image that may contain vulnerabilities we should keep them up to date even if there s not an updated base image to consume solution update base image for all mirror node components add an extra command to update all os packages so they at least are up to date as of each tag run apt get update debian frontend noninteractive apt get install y no install recommends rm rf var lib apt lists or run apk update apk upgrade rm rf var cache apk alternatives no response
1
73,401
8,871,796,987
IssuesEvent
2019-01-11 13:45:49
bitpay/copay
https://api.github.com/repos/bitpay/copay
closed
Improve delete wallet view design
Design Needed
_From @bitjson on October 18, 2016 22:45_ Needs love: ![screen shot 2016-10-18 at 6 44 41 pm](https://cloud.githubusercontent.com/assets/904007/19499175/010fd39a-9563-11e6-8b39-18bae774de12.png) _Copied from original issue: bitpay/bitpay-wallet#562_
1.0
Improve delete wallet view design - _From @bitjson on October 18, 2016 22:45_ Needs love: ![screen shot 2016-10-18 at 6 44 41 pm](https://cloud.githubusercontent.com/assets/904007/19499175/010fd39a-9563-11e6-8b39-18bae774de12.png) _Copied from original issue: bitpay/bitpay-wallet#562_
non_process
improve delete wallet view design from bitjson on october needs love copied from original issue bitpay bitpay wallet
0
176,217
28,045,023,192
IssuesEvent
2023-03-28 21:54:12
MozillaFoundation/foundation.mozilla.org
https://api.github.com/repos/MozillaFoundation/foundation.mozilla.org
closed
Roadmap and Prioritize Site Improvement Recommendations
design
following up issue #6303 The recommendations from the IA refresh has been labeled and categorized on a spreadsheet https://docs.google.com/spreadsheets/d/14XlcxPYT5qJFPnMUsPmDk512lMzTuLx0OJeSwGbmAQM/edit?usp=sharing These will need be prioritized so 'complexity for implementation' and 'impact' need to be scoped. The action items can then be roadmapped to plan when/where it fits in with the rest of the foundation site work.
1.0
Roadmap and Prioritize Site Improvement Recommendations - following up issue #6303 The recommendations from the IA refresh has been labeled and categorized on a spreadsheet https://docs.google.com/spreadsheets/d/14XlcxPYT5qJFPnMUsPmDk512lMzTuLx0OJeSwGbmAQM/edit?usp=sharing These will need be prioritized so 'complexity for implementation' and 'impact' need to be scoped. The action items can then be roadmapped to plan when/where it fits in with the rest of the foundation site work.
non_process
roadmap and prioritize site improvement recommendations following up issue the recommendations from the ia refresh has been labeled and categorized on a spreadsheet these will need be prioritized so complexity for implementation and impact need to be scoped the action items can then be roadmapped to plan when where it fits in with the rest of the foundation site work
0
150,577
11,967,347,977
IssuesEvent
2020-04-06 06:27:53
ubtue/DatenProbleme
https://api.github.com/repos/ubtue/DatenProbleme
closed
ISSN 1573-0697 Journal of business ethics Mix aus Online First und Standardartikel
Zotero_AUTO_RSS blocked ready for testing
Die Daten enthalten sowohl Artikel die einem Heft zugeordnet sind, als auch solche ohne Heftzuordnung
1.0
ISSN 1573-0697 Journal of business ethics Mix aus Online First und Standardartikel - Die Daten enthalten sowohl Artikel die einem Heft zugeordnet sind, als auch solche ohne Heftzuordnung
non_process
issn journal of business ethics mix aus online first und standardartikel die daten enthalten sowohl artikel die einem heft zugeordnet sind als auch solche ohne heftzuordnung
0
71,688
15,207,902,600
IssuesEvent
2021-02-17 01:17:40
billmcchesney1/hadoop
https://api.github.com/repos/billmcchesney1/hadoop
opened
CVE-2020-36189 (Medium) detected in jackson-databind-2.9.10.1.jar
security vulnerability
## CVE-2020-36189 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib/jackson-databind-2.9.10.1.jar,hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/target/lib/jackson-databind-2.9.10.1.jar,hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client/target/lib/jackson-databind-2.9.10.1.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189>CVE-2020-36189</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/commit/33d96c13fe18a2dad01b19ce195548c9acea9da4">https://github.com/FasterXML/jackson-databind/commit/33d96c13fe18a2dad01b19ce195548c9acea9da4</a></p> <p>Release Date: 2020-12-26</p> <p>Fix Resolution: Replace or update the following files: SubTypeValidator.java, VERSION-2.x</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.1","isMinimumFixVersionAvailable":false}],"baseBranches":["trunk"],"vulnerabilityIdentifier":"CVE-2020-36189","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
True
CVE-2020-36189 (Medium) detected in jackson-databind-2.9.10.1.jar - ## CVE-2020-36189 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib/jackson-databind-2.9.10.1.jar,hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/target/lib/jackson-databind-2.9.10.1.jar,hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client/target/lib/jackson-databind-2.9.10.1.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189>CVE-2020-36189</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/commit/33d96c13fe18a2dad01b19ce195548c9acea9da4">https://github.com/FasterXML/jackson-databind/commit/33d96c13fe18a2dad01b19ce195548c9acea9da4</a></p> <p>Release Date: 2020-12-26</p> <p>Fix Resolution: Replace or update the following files: SubTypeValidator.java, VERSION-2.x</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.1","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.1","isMinimumFixVersionAvailable":false}],"baseBranches":["trunk"],"vulnerabilityIdentifier":"CVE-2020-36189","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice documentstore target lib jackson databind jar hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice hbase hadoop yarn server timelineservice hbase common target lib jackson databind jar hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice hbase hadoop yarn server timelineservice hbase client target lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch trunk vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db drivermanagerconnectionsource publish date url a href cvss score details base score metrics not available suggested fix type change files origin a href release date fix resolution replace or update the following files subtypevalidator java version x isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db drivermanagerconnectionsource vulnerabilityurl
0
8,415
11,580,353,798
IssuesEvent
2020-02-21 19:57:11
usgpo/bill-status
https://api.github.com/repos/usgpo/bill-status
closed
Incorrect Senate Homeland Subcommittee Name
file reprocessed
In the bulk data source on govinfo.gov, some Senate bills have the incorrect subcommittee name for the Homeland Security and Governmental Affairs Subcommittee on Regulatory Affairs and Federal Management: https://www.govinfo.gov/bulkdata/BILLSTATUS/116/s/BILLSTATUS-116s1120.xml ``` <subcommittees> <item> <systemCode>ssga19</systemCode> <name>and Federal Management Subcommittee</name> <activities> <item> <name>Hearings by</name> <date>2019-05-22T18:31:34Z</date> </item> </activities> </item> </subcommittees> ``` The subcommittee name has been truncated to remove "Regulatory Affairs", and it's incorrect in the subcommittee listing as well as the bill actions. The text of the action has the correct full name of the subcommittee: ``` <item> <actionDate>2019-05-22</actionDate> <committees> <item> <systemCode>ssga19</systemCode> <name>and Federal Management Subcommittee</name> </item> </committees> <links/> <sourceSystem> <code>0</code> <name>Senate</name> </sourceSystem> <text> Committee on Homeland Security and Governmental Affairs Subcommittee on Regulatory Affairs and Federal Management. Hearings held. </text> <type>Committee</type> </item> ``` This is affecting the following bills in the 116th Congress: S. 1419 S. 83 S. 1120 S. 1339 S. 78 S. 1409 S. 1420
1.0
Incorrect Senate Homeland Subcommittee Name - In the bulk data source on govinfo.gov, some Senate bills have the incorrect subcommittee name for the Homeland Security and Governmental Affairs Subcommittee on Regulatory Affairs and Federal Management: https://www.govinfo.gov/bulkdata/BILLSTATUS/116/s/BILLSTATUS-116s1120.xml ``` <subcommittees> <item> <systemCode>ssga19</systemCode> <name>and Federal Management Subcommittee</name> <activities> <item> <name>Hearings by</name> <date>2019-05-22T18:31:34Z</date> </item> </activities> </item> </subcommittees> ``` The subcommittee name has been truncated to remove "Regulatory Affairs", and it's incorrect in the subcommittee listing as well as the bill actions. The text of the action has the correct full name of the subcommittee: ``` <item> <actionDate>2019-05-22</actionDate> <committees> <item> <systemCode>ssga19</systemCode> <name>and Federal Management Subcommittee</name> </item> </committees> <links/> <sourceSystem> <code>0</code> <name>Senate</name> </sourceSystem> <text> Committee on Homeland Security and Governmental Affairs Subcommittee on Regulatory Affairs and Federal Management. Hearings held. </text> <type>Committee</type> </item> ``` This is affecting the following bills in the 116th Congress: S. 1419 S. 83 S. 1120 S. 1339 S. 78 S. 1409 S. 1420
process
incorrect senate homeland subcommittee name in the bulk data source on govinfo gov some senate bills have the incorrect subcommittee name for the homeland security and governmental affairs subcommittee on regulatory affairs and federal management and federal management subcommittee hearings by the subcommittee name has been truncated to remove regulatory affairs and it s incorrect in the subcommittee listing as well as the bill actions the text of the action has the correct full name of the subcommittee and federal management subcommittee senate committee on homeland security and governmental affairs subcommittee on regulatory affairs and federal management hearings held committee this is affecting the following bills in the congress s s s s s s s
1
511,356
14,858,727,212
IssuesEvent
2021-01-18 17:14:31
weaveworks/eksctl
https://api.github.com/repos/weaveworks/eksctl
closed
Ability to specify egress/ingress rules for cluster shared security group
area/nodegroup kind/feature priority/backlog stale
**Why do you want this feature?** See background in https://github.com/weaveworks/eksctl/issues/1773 Managing security groups outside of `eksctl` just to customize egress/ingress adds inordinate complexity **What feature/behavior/change do you want?** > Ideally that would be accomplished through a feature where I can specify those [egress/ingress] rules in Cluster configuration file just like IAM policies.
1.0
Ability to specify egress/ingress rules for cluster shared security group - **Why do you want this feature?** See background in https://github.com/weaveworks/eksctl/issues/1773 Managing security groups outside of `eksctl` just to customize egress/ingress adds inordinate complexity **What feature/behavior/change do you want?** > Ideally that would be accomplished through a feature where I can specify those [egress/ingress] rules in Cluster configuration file just like IAM policies.
non_process
ability to specify egress ingress rules for cluster shared security group why do you want this feature see background in managing security groups outside of eksctl just to customize egress ingress adds inordinate complexity what feature behavior change do you want ideally that would be accomplished through a feature where i can specify those rules in cluster configuration file just like iam policies
0
101,047
11,211,974,826
IssuesEvent
2020-01-06 16:32:47
project-koku/koku
https://api.github.com/repos/project-koku/koku
opened
Update nise documentation
developer productivity documentation
## User Story As a user of nise (especially now that its on pypi) I want proper documentation so that I can use the tool. ## Assumptions - We can keep our readme (or some form of it) - nise does more than when it was first created, we can create a docs folder and add more structured documentation there - We can cover randomized data - We can cover static YAML file usage - We can cover how to really use nise data when docker compose so that data is loaded for a provider - We can cover how to really use nise data when OpenShift so that data is loaded for a provider ## Acceptance Criteria - [ ] Documentation is usable by a newcomer to the project and they can successfully generate data and load the data for a provider
1.0
Update nise documentation - ## User Story As a user of nise (especially now that its on pypi) I want proper documentation so that I can use the tool. ## Assumptions - We can keep our readme (or some form of it) - nise does more than when it was first created, we can create a docs folder and add more structured documentation there - We can cover randomized data - We can cover static YAML file usage - We can cover how to really use nise data when docker compose so that data is loaded for a provider - We can cover how to really use nise data when OpenShift so that data is loaded for a provider ## Acceptance Criteria - [ ] Documentation is usable by a newcomer to the project and they can successfully generate data and load the data for a provider
non_process
update nise documentation user story as a user of nise especially now that its on pypi i want proper documentation so that i can use the tool assumptions we can keep our readme or some form of it nise does more than when it was first created we can create a docs folder and add more structured documentation there we can cover randomized data we can cover static yaml file usage we can cover how to really use nise data when docker compose so that data is loaded for a provider we can cover how to really use nise data when openshift so that data is loaded for a provider acceptance criteria documentation is usable by a newcomer to the project and they can successfully generate data and load the data for a provider
0
153,339
13,503,385,993
IssuesEvent
2020-09-13 13:19:39
geek-engineer-future/podcast
https://api.github.com/repos/geek-engineer-future/podcast
closed
[2020-09-18] Recording Document
documentation
## テーマ hoge ## 内容 hoge ## appendix hoge # --- - 最近の気になるトピックをコメントに書きましょう!(技術トピックの場合は「タイトル」+「URL」+「概要」も一緒に書くこと) - 今週話せそうなテーマがある方はコメントに書きましょう!
1.0
[2020-09-18] Recording Document - ## テーマ hoge ## 内容 hoge ## appendix hoge # --- - 最近の気になるトピックをコメントに書きましょう!(技術トピックの場合は「タイトル」+「URL」+「概要」も一緒に書くこと) - 今週話せそうなテーマがある方はコメントに書きましょう!
non_process
recording document テーマ hoge 内容 hoge appendix hoge 最近の気になるトピックをコメントに書きましょう!(技術トピックの場合は「タイトル」 「url」 「概要」も一緒に書くこと) 今週話せそうなテーマがある方はコメントに書きましょう!
0
43,499
9,449,984,941
IssuesEvent
2019-04-16 04:30:01
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
Filing issues with multiple batched code locations ("bookmarks")
code-nav feature-request
As a user who is reviewing code for mistakes, I want to be able to "bookmark" multiple locations of code and then create a batch issue (eg on Jira) with all of the locations. I want anyone viewing the code on Sourcegraph to be able to see the issue I filed, and I want anyone viewing the issue on Jira to be able to go to the location on Sourcegraph.
1.0
Filing issues with multiple batched code locations ("bookmarks") - As a user who is reviewing code for mistakes, I want to be able to "bookmark" multiple locations of code and then create a batch issue (eg on Jira) with all of the locations. I want anyone viewing the code on Sourcegraph to be able to see the issue I filed, and I want anyone viewing the issue on Jira to be able to go to the location on Sourcegraph.
non_process
filing issues with multiple batched code locations bookmarks as a user who is reviewing code for mistakes i want to be able to bookmark multiple locations of code and then create a batch issue eg on jira with all of the locations i want anyone viewing the code on sourcegraph to be able to see the issue i filed and i want anyone viewing the issue on jira to be able to go to the location on sourcegraph
0
899
2,594,288,805
IssuesEvent
2015-02-20 01:31:09
BALL-Project/ball
https://api.github.com/repos/BALL-Project/ball
closed
BALLView does not report any progress when exporting VRML file
C: VIEW P: minor R: fixed T: defect
**Reported by odin on 9 Jul 39586499 03:33 UTC** Should write out at least its done into the log window. Now, the only way to know the export is finished is to watch the file size growing.
1.0
BALLView does not report any progress when exporting VRML file - **Reported by odin on 9 Jul 39586499 03:33 UTC** Should write out at least its done into the log window. Now, the only way to know the export is finished is to watch the file size growing.
non_process
ballview does not report any progress when exporting vrml file reported by odin on jul utc should write out at least its done into the log window now the only way to know the export is finished is to watch the file size growing
0
1,753
4,445,924,236
IssuesEvent
2016-08-20 10:26:01
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Issue passing file descriptors in OS X
child_process net os x
<!-- Thank you for reporting an issue. Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: `v7.0.0-pre` * **Platform**: `OS X 10.11.5` * **Subsystem**: child_process <!-- Enter your issue details below this comment. --> The `test-cluster-net-send.js` test is sometimes failing when running the test suite in my `OS X` with the following output: ``` === release test-cluster-net-send === Path: parallel/test-cluster-net-send [32787] master [32789] worker assert.js:90 throw new assert.AssertionError({ ^ AssertionError: false == true at process.<anonymous> (/Users/sgimeno/node/node/test/parallel/test-cluster-net-send.js:29:12) at process.g (events.js:286:16) at emitOne (events.js:101:20) at process.emit (events.js:188:7) Command: out/Release/node /Users/sgimeno/node/node/test/parallel/test-cluster-net-send.js ``` After investigating the issue it *looks* like that the error happens only when the `fd` that passes the worker to the master is closed before it is received in the master process. The following patch, that closes the `fd` only after receiving the `NODE_HANDLE_ACK` message fixes the issue for me. ```diff diff --git a/lib/internal/child_process.js b/lib/internal/child_process.js index 789c29e..44a245e 100644 --- a/lib/internal/child_process.js +++ b/lib/internal/child_process.js @@ -96,8 +96,8 @@ const handleConversion = { postSend: function(handle, options) { // Close the Socket handle after sending it - if (handle && !options.keepOpen) - handle.close(); + //if (handle && !options.keepOpen) + // global_handle = handle; }, got: function(message, handle, emit) { @@ -465,6 +465,11 @@ function setupChannel(target, channel) { target.on('internalMessage', function(message, handle) { // Once acknowledged - continue sending handles. if (message.cmd === 'NODE_HANDLE_ACK') { + if (target._pending_handle) { + target._pending_handle.close(); + target._pending_handle = null; + } + assert(Array.isArray(target._handleQueue)); var queue = target._handleQueue; target._handleQueue = null; @@ -615,8 +620,13 @@ function setupChannel(target, channel) { req.oncomplete = function() { if (this.async === true) control.unref(); - if (obj && obj.postSend) + if (obj && obj.postSend) { obj.postSend(handle, options); + if (handle && !options.keepOpen) { + assert(!target._pending_handle); + target._pending_handle = handle; + } + } ``` This seems strange to me as my understanding was that closing the file descriptor after sending it was safe (at least in my Linux box I have not been able to reproduce the same issue). Thoughts?
1.0
Issue passing file descriptors in OS X - <!-- Thank you for reporting an issue. Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: `v7.0.0-pre` * **Platform**: `OS X 10.11.5` * **Subsystem**: child_process <!-- Enter your issue details below this comment. --> The `test-cluster-net-send.js` test is sometimes failing when running the test suite in my `OS X` with the following output: ``` === release test-cluster-net-send === Path: parallel/test-cluster-net-send [32787] master [32789] worker assert.js:90 throw new assert.AssertionError({ ^ AssertionError: false == true at process.<anonymous> (/Users/sgimeno/node/node/test/parallel/test-cluster-net-send.js:29:12) at process.g (events.js:286:16) at emitOne (events.js:101:20) at process.emit (events.js:188:7) Command: out/Release/node /Users/sgimeno/node/node/test/parallel/test-cluster-net-send.js ``` After investigating the issue it *looks* like that the error happens only when the `fd` that passes the worker to the master is closed before it is received in the master process. The following patch, that closes the `fd` only after receiving the `NODE_HANDLE_ACK` message fixes the issue for me. ```diff diff --git a/lib/internal/child_process.js b/lib/internal/child_process.js index 789c29e..44a245e 100644 --- a/lib/internal/child_process.js +++ b/lib/internal/child_process.js @@ -96,8 +96,8 @@ const handleConversion = { postSend: function(handle, options) { // Close the Socket handle after sending it - if (handle && !options.keepOpen) - handle.close(); + //if (handle && !options.keepOpen) + // global_handle = handle; }, got: function(message, handle, emit) { @@ -465,6 +465,11 @@ function setupChannel(target, channel) { target.on('internalMessage', function(message, handle) { // Once acknowledged - continue sending handles. if (message.cmd === 'NODE_HANDLE_ACK') { + if (target._pending_handle) { + target._pending_handle.close(); + target._pending_handle = null; + } + assert(Array.isArray(target._handleQueue)); var queue = target._handleQueue; target._handleQueue = null; @@ -615,8 +620,13 @@ function setupChannel(target, channel) { req.oncomplete = function() { if (this.async === true) control.unref(); - if (obj && obj.postSend) + if (obj && obj.postSend) { obj.postSend(handle, options); + if (handle && !options.keepOpen) { + assert(!target._pending_handle); + target._pending_handle = handle; + } + } ``` This seems strange to me as my understanding was that closing the file descriptor after sending it was safe (at least in my Linux box I have not been able to reproduce the same issue). Thoughts?
process
issue passing file descriptors in os x thank you for reporting an issue please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version pre platform os x subsystem child process the test cluster net send js test is sometimes failing when running the test suite in my os x with the following output release test cluster net send path parallel test cluster net send master worker assert js throw new assert assertionerror assertionerror false true at process users sgimeno node node test parallel test cluster net send js at process g events js at emitone events js at process emit events js command out release node users sgimeno node node test parallel test cluster net send js after investigating the issue it looks like that the error happens only when the fd that passes the worker to the master is closed before it is received in the master process the following patch that closes the fd only after receiving the node handle ack message fixes the issue for me diff diff git a lib internal child process js b lib internal child process js index a lib internal child process js b lib internal child process js const handleconversion postsend function handle options close the socket handle after sending it if handle options keepopen handle close if handle options keepopen global handle handle got function message handle emit function setupchannel target channel target on internalmessage function message handle once acknowledged continue sending handles if message cmd node handle ack if target pending handle target pending handle close target pending handle null assert array isarray target handlequeue var queue target handlequeue target handlequeue null function setupchannel target channel req oncomplete function if this async true control unref if obj obj postsend if obj obj postsend obj postsend handle options if handle options keepopen assert target pending handle target pending handle handle this seems strange to me as my understanding was that closing the file descriptor after sending it was safe at least in my linux box i have not been able to reproduce the same issue thoughts
1
816,250
30,595,163,827
IssuesEvent
2023-07-21 21:06:06
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
Auto switching for interactive components
enhancement area-blazor Priority:1 feature-full-stack-web-ui
In scope: * Use Server by default while loading WebAssembly runtime in the background * If WebAssembly files are cached, use WebAssembly mode Out of scope: * Customizing how the auto mode makes its decision (#48756) * Shutting down Server circuits eagerly once WebAssembly files are downloaded (instead, just let them shut down naturally as the user navigates around)
1.0
Auto switching for interactive components - In scope: * Use Server by default while loading WebAssembly runtime in the background * If WebAssembly files are cached, use WebAssembly mode Out of scope: * Customizing how the auto mode makes its decision (#48756) * Shutting down Server circuits eagerly once WebAssembly files are downloaded (instead, just let them shut down naturally as the user navigates around)
non_process
auto switching for interactive components in scope use server by default while loading webassembly runtime in the background if webassembly files are cached use webassembly mode out of scope customizing how the auto mode makes its decision shutting down server circuits eagerly once webassembly files are downloaded instead just let them shut down naturally as the user navigates around
0
21,562
29,922,573,555
IssuesEvent
2023-06-22 00:38:08
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] Scrum Master na Coodesh
SALVADOR PJ BIG DATA PHP JAVA SCRUM AGILE MOBILE REQUISITOS REMOTO PROCESSOS GITHUB KANBAN CI UMA ANALYTICS ENGENHARIA DE SOFTWARE Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/scrum-master-144141746?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A Loginfo está em busca de Scrum Master para compor seu time!</p> <p>Como especialistas, escolhemos entregar ao mercado uma solução totalmente mobile, capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes. Somos movidos a tecnologia e temos fome de resultados. Nosso principal objetivo é transformar o mercado entregando logística aprimorada, apoiando a logística e o comércio exterior com a otimização de processos, redução de custos e ganho de produtividade no recebimento, armazenagem e expedição.</p> <p>Responsabilidades:</p> <ul> <li>Coordenar e orientar a equipe de desenvolvimento conforme as prioridades identificadas, tomando decisões e acompanhando a realização dos trabalhos para que sejam executados nos prazos e formas previstos;</li> <li>Atuar como Scrum Master em times ágeis dentro do framework safe, promover workshops de agilidade e team Building.</li> </ul> ## Loginfo Tecnologia da Informação LTDA: <p>Desde 2014, ano de nossa fundação, escolhemos conectar processos operacionais e comunicação. Decidimos tornar cada vez mais digital, ágil e intuitivo o mercado dos setores logísticos, portuários e de armazéns gerais. Somos inovadores em tudo que nos propomos a fazer. Como especialistas, escolhemos entregar ao mercado uma solução totalmente mobile, capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes. Somos movidos a tecnologia e temos fome de resultados. Nosso principal objetivo é transformar o mercado entregando logística aprimorada, apoiando a logística e o comércio exterior com a otimização de processos, redução de custos e ganho de produtividade no recebimento, armazenagem e expedição.</p></p> ## Habilidades: - PHP - Agile - Kanban ## Local: 100% Remoto ## Requisitos: - Experiência como Scrum Master; - Graduação Completa em Sistemas de Informação, Engenharia de Software ou Ciências da Computação; - Certificação em Scrum Master. ## Diferenciais: - Cursando Pós-graduação/MBA em Desenvolvimento em Soluções Corporativas JAVA, Gestão de TI ou Big Data Analytics. ## Benefícios: - Horários flexíveis; - Gympass; - Alura; - Seguro de Vida. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Scrum Master na Loginfo Tecnologia da Informação LTDA](https://coodesh.com/jobs/scrum-master-144141746?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime PJ #### Categoria Gestão em TI
1.0
[Remoto] Scrum Master na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/scrum-master-144141746?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A Loginfo está em busca de Scrum Master para compor seu time!</p> <p>Como especialistas, escolhemos entregar ao mercado uma solução totalmente mobile, capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes. Somos movidos a tecnologia e temos fome de resultados. Nosso principal objetivo é transformar o mercado entregando logística aprimorada, apoiando a logística e o comércio exterior com a otimização de processos, redução de custos e ganho de produtividade no recebimento, armazenagem e expedição.</p> <p>Responsabilidades:</p> <ul> <li>Coordenar e orientar a equipe de desenvolvimento conforme as prioridades identificadas, tomando decisões e acompanhando a realização dos trabalhos para que sejam executados nos prazos e formas previstos;</li> <li>Atuar como Scrum Master em times ágeis dentro do framework safe, promover workshops de agilidade e team Building.</li> </ul> ## Loginfo Tecnologia da Informação LTDA: <p>Desde 2014, ano de nossa fundação, escolhemos conectar processos operacionais e comunicação. Decidimos tornar cada vez mais digital, ágil e intuitivo o mercado dos setores logísticos, portuários e de armazéns gerais. Somos inovadores em tudo que nos propomos a fazer. Como especialistas, escolhemos entregar ao mercado uma solução totalmente mobile, capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes. Somos movidos a tecnologia e temos fome de resultados. Nosso principal objetivo é transformar o mercado entregando logística aprimorada, apoiando a logística e o comércio exterior com a otimização de processos, redução de custos e ganho de produtividade no recebimento, armazenagem e expedição.</p></p> ## Habilidades: - PHP - Agile - Kanban ## Local: 100% Remoto ## Requisitos: - Experiência como Scrum Master; - Graduação Completa em Sistemas de Informação, Engenharia de Software ou Ciências da Computação; - Certificação em Scrum Master. ## Diferenciais: - Cursando Pós-graduação/MBA em Desenvolvimento em Soluções Corporativas JAVA, Gestão de TI ou Big Data Analytics. ## Benefícios: - Horários flexíveis; - Gympass; - Alura; - Seguro de Vida. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Scrum Master na Loginfo Tecnologia da Informação LTDA](https://coodesh.com/jobs/scrum-master-144141746?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime PJ #### Categoria Gestão em TI
process
scrum master na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a loginfo está em busca de scrum master para compor seu time como especialistas escolhemos entregar ao mercado uma solução totalmente mobile capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes somos movidos a tecnologia e temos fome de resultados nosso principal objetivo é transformar o mercado entregando logística aprimorada apoiando a logística e o comércio exterior com a otimização de processos redução de custos e ganho de produtividade no recebimento armazenagem e expedição responsabilidades coordenar e orientar a equipe de desenvolvimento conforme as prioridades identificadas tomando decisões e acompanhando a realização dos trabalhos para que sejam executados nos prazos e formas previstos atuar como scrum master em times ágeis dentro do framework safe promover workshops de agilidade e team building loginfo tecnologia da informação ltda desde ano de nossa fundação escolhemos conectar processos operacionais e comunicação decidimos tornar cada vez mais digital ágil e intuitivo o mercado dos setores logísticos portuários e de armazéns gerais somos inovadores em tudo que nos propomos a fazer como especialistas escolhemos entregar ao mercado uma solução totalmente mobile capaz de cobrir de ponta a ponta a intralogística e o dia a dia de nossos clientes somos movidos a tecnologia e temos fome de resultados nosso principal objetivo é transformar o mercado entregando logística aprimorada apoiando a logística e o comércio exterior com a otimização de processos redução de custos e ganho de produtividade no recebimento armazenagem e expedição habilidades php agile kanban local remoto requisitos experiência como scrum master graduação completa em sistemas de informação engenharia de software ou ciências da computação certificação em scrum master diferenciais cursando pós graduação mba em desenvolvimento em soluções corporativas java gestão de ti ou big data analytics benefícios horários flexíveis gympass alura seguro de vida como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime pj categoria gestão em ti
1
167,584
26,517,867,958
IssuesEvent
2023-01-18 22:32:32
vegaprotocol/frontend-monorepo
https://api.github.com/repos/vegaprotocol/frontend-monorepo
opened
Replace dropdown icon with custom version
Trading ux-and-visual-design chore common
## The Chore Currently using a blueprint version of the icon: ![Screenshot 2023-01-18 at 14 28 53](https://user-images.githubusercontent.com/6803987/213309209-15da9cd9-71e1-4f65-b0b1-a2f51b1ec793.jpg) vega.xyz uses a different icon for dropdowns: ![Screenshot 2023-01-18 at 14 30 45](https://user-images.githubusercontent.com/6803987/213309481-9071b3d0-035c-4e1a-8d5d-8b1f3becd666.jpg) ## Tasks - [ ] Consider design options (use vega.xyz pattern, or add a new icon for the apps design system) - [ ] Add to design system - [ ] Dev
1.0
Replace dropdown icon with custom version - ## The Chore Currently using a blueprint version of the icon: ![Screenshot 2023-01-18 at 14 28 53](https://user-images.githubusercontent.com/6803987/213309209-15da9cd9-71e1-4f65-b0b1-a2f51b1ec793.jpg) vega.xyz uses a different icon for dropdowns: ![Screenshot 2023-01-18 at 14 30 45](https://user-images.githubusercontent.com/6803987/213309481-9071b3d0-035c-4e1a-8d5d-8b1f3becd666.jpg) ## Tasks - [ ] Consider design options (use vega.xyz pattern, or add a new icon for the apps design system) - [ ] Add to design system - [ ] Dev
non_process
replace dropdown icon with custom version the chore currently using a blueprint version of the icon vega xyz uses a different icon for dropdowns tasks consider design options use vega xyz pattern or add a new icon for the apps design system add to design system dev
0
17,013
22,386,217,717
IssuesEvent
2022-06-17 00:51:38
figlesias221/ProyectoDevOps_Grupo3_IglesiasPerezMolinoloJuan
https://api.github.com/repos/figlesias221/ProyectoDevOps_Grupo3_IglesiasPerezMolinoloJuan
closed
Review FrontEnd Alta de Puntos de carga
task process
Esfuerzo en HS-P: Estimado: 1 Real: 1 (@mperezjodal ), 1 (@andrujuanoo )
1.0
Review FrontEnd Alta de Puntos de carga - Esfuerzo en HS-P: Estimado: 1 Real: 1 (@mperezjodal ), 1 (@andrujuanoo )
process
review frontend alta de puntos de carga esfuerzo en hs p estimado real mperezjodal andrujuanoo
1
74,769
20,366,507,576
IssuesEvent
2022-02-21 06:33:23
pandres95/ndi.js
https://api.github.com/repos/pandres95/ndi.js
closed
Error 403 while installing module
bug build
An error `403` is encountered whilst attempting to install the package from https://ndijs.s3.us-east-2.amazonaws.com/ndi/v1.0.5/Release/linux-x64.tar.gz If you're wondering, I'm using WSL and tried Node 16.14.0 and Node 17.5.0
1.0
Error 403 while installing module - An error `403` is encountered whilst attempting to install the package from https://ndijs.s3.us-east-2.amazonaws.com/ndi/v1.0.5/Release/linux-x64.tar.gz If you're wondering, I'm using WSL and tried Node 16.14.0 and Node 17.5.0
non_process
error while installing module an error is encountered whilst attempting to install the package from if you re wondering i m using wsl and tried node and node
0
33,669
7,743,208,769
IssuesEvent
2018-05-29 12:07:13
guirisan/arrelaires
https://api.github.com/repos/guirisan/arrelaires
closed
Enviar correu a admins al registrar nova usuària o col·laboració
code things
Afegides en `app/config/mail.php` una clau `admins` amb les adreces de les admins ### REGISTRE Creem mail `NewUserNotification` i la vista `/resources/views/emails/admin/new-user-notification.blade.php En RegistersUsers@register afegim `\Mail::to(config('mail.admins'))->send(new NewUserNotification($user));` ### COL·LABORACIÓ Creem mail `NewCollaborationNotification` i la vista `/resources/views/emails/admin/new-collaboration-notification.blade.php En FormController@store_step_one afegim `\Mail::to(config('mail.admins'))->send(new NewCollaborationNotification($persona));`
1.0
Enviar correu a admins al registrar nova usuària o col·laboració - Afegides en `app/config/mail.php` una clau `admins` amb les adreces de les admins ### REGISTRE Creem mail `NewUserNotification` i la vista `/resources/views/emails/admin/new-user-notification.blade.php En RegistersUsers@register afegim `\Mail::to(config('mail.admins'))->send(new NewUserNotification($user));` ### COL·LABORACIÓ Creem mail `NewCollaborationNotification` i la vista `/resources/views/emails/admin/new-collaboration-notification.blade.php En FormController@store_step_one afegim `\Mail::to(config('mail.admins'))->send(new NewCollaborationNotification($persona));`
non_process
enviar correu a admins al registrar nova usuària o col·laboració afegides en app config mail php una clau admins amb les adreces de les admins registre creem mail newusernotification i la vista resources views emails admin new user notification blade php en registersusers register afegim mail to config mail admins send new newusernotification user col·laboració creem mail newcollaborationnotification i la vista resources views emails admin new collaboration notification blade php en formcontroller store step one afegim mail to config mail admins send new newcollaborationnotification persona
0
140,372
5,400,755,063
IssuesEvent
2017-02-27 22:52:56
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
[k8s.io] Empty [Feature:Empty] does nothing {Kubernetes e2e suite}
kind/flake priority/P2
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-kubemark-5-gce/1946/ Failed: [k8s.io] Empty [Feature:Empty] does nothing {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:143 Dec 9 12:38:16.486: Couldn't delete ns: "e2e-tests-empty-cgxh2": client: etcd cluster is unavailable or misconfigured (&errors.StatusError{ErrStatus:v1.Status{TypeMeta:v1.TypeMeta{Kind:"Status", APIVersion:"v1"}, ListMeta:v1.ListMeta{SelfLink:"", ResourceVersion:""}, Status:"Failure", Message:"client: etcd cluster is unavailable or misconfigured", Reason:"", Details:(*v1.StatusDetails)(nil), Code:500}}) /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:354 ```
1.0
[k8s.io] Empty [Feature:Empty] does nothing {Kubernetes e2e suite} - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-kubemark-5-gce/1946/ Failed: [k8s.io] Empty [Feature:Empty] does nothing {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:143 Dec 9 12:38:16.486: Couldn't delete ns: "e2e-tests-empty-cgxh2": client: etcd cluster is unavailable or misconfigured (&errors.StatusError{ErrStatus:v1.Status{TypeMeta:v1.TypeMeta{Kind:"Status", APIVersion:"v1"}, ListMeta:v1.ListMeta{SelfLink:"", ResourceVersion:""}, Status:"Failure", Message:"client: etcd cluster is unavailable or misconfigured", Reason:"", Details:(*v1.StatusDetails)(nil), Code:500}}) /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:354 ```
non_process
empty does nothing kubernetes suite failed empty does nothing kubernetes suite go src io kubernetes output dockerized go src io kubernetes test framework framework go dec couldn t delete ns tests empty client etcd cluster is unavailable or misconfigured errors statuserror errstatus status typemeta typemeta kind status apiversion listmeta listmeta selflink resourceversion status failure message client etcd cluster is unavailable or misconfigured reason details statusdetails nil code go src io kubernetes output dockerized go src io kubernetes test framework framework go
0
21,757
30,276,364,266
IssuesEvent
2023-07-07 20:04:57
gsoft-inc/ov-igloo-ui
https://api.github.com/repos/gsoft-inc/ov-igloo-ui
closed
[Bug]: ActionMenu with disablePortal has a weird behaviour
bug in process
### Contact Details _No response_ ### What happened? When I play with the component, it sometimes render at a random place (the top and left css properties are not good) : ![démo](https://github.com/gsoft-inc/ov-igloo-ui/assets/25105785/3862009f-00fb-4397-a4c5-fd1c7e22fa4b) Also, as you can hardly see at the beginning of the gif, the hover state is not always triggered (no translation + no cursor). It seems to work well on storybook, so maybe it is related to the previous bug, or maybe it is something on my side, and if you have any ideas, I'm all ears! Edit : it seems to be related to the previous bug since I can see the top and left properties changed. Edit 2: When testing on my side, it seems that if we change the position: fixed to position: absolute, it works! So maybe it is the fix to do here in the design system! ![image](https://github.com/gsoft-inc/ov-igloo-ui/assets/25105785/960c2673-509c-4008-92a3-a3c230fe7fee) (+ adding on the .ids-action-menu the property position: relative) ### Component ActionMenu ### Component Version 1.1.2 ### Which browsers are you seeing the problem on? _No response_ ### Mobile Device _No response_ ### Relevant log output _No response_
1.0
[Bug]: ActionMenu with disablePortal has a weird behaviour - ### Contact Details _No response_ ### What happened? When I play with the component, it sometimes render at a random place (the top and left css properties are not good) : ![démo](https://github.com/gsoft-inc/ov-igloo-ui/assets/25105785/3862009f-00fb-4397-a4c5-fd1c7e22fa4b) Also, as you can hardly see at the beginning of the gif, the hover state is not always triggered (no translation + no cursor). It seems to work well on storybook, so maybe it is related to the previous bug, or maybe it is something on my side, and if you have any ideas, I'm all ears! Edit : it seems to be related to the previous bug since I can see the top and left properties changed. Edit 2: When testing on my side, it seems that if we change the position: fixed to position: absolute, it works! So maybe it is the fix to do here in the design system! ![image](https://github.com/gsoft-inc/ov-igloo-ui/assets/25105785/960c2673-509c-4008-92a3-a3c230fe7fee) (+ adding on the .ids-action-menu the property position: relative) ### Component ActionMenu ### Component Version 1.1.2 ### Which browsers are you seeing the problem on? _No response_ ### Mobile Device _No response_ ### Relevant log output _No response_
process
actionmenu with disableportal has a weird behaviour contact details no response what happened when i play with the component it sometimes render at a random place the top and left css properties are not good also as you can hardly see at the beginning of the gif the hover state is not always triggered no translation no cursor it seems to work well on storybook so maybe it is related to the previous bug or maybe it is something on my side and if you have any ideas i m all ears edit it seems to be related to the previous bug since i can see the top and left properties changed edit when testing on my side it seems that if we change the position fixed to position absolute it works so maybe it is the fix to do here in the design system adding on the ids action menu the property position relative component actionmenu component version which browsers are you seeing the problem on no response mobile device no response relevant log output no response
1
51,259
6,506,412,627
IssuesEvent
2017-08-24 08:56:18
thememachine/eleven
https://api.github.com/repos/thememachine/eleven
opened
edit tabs on the frontned
Design FEAT
# The issue is: Tabs on the frontend is taking up a lot of space + they are fugly to look at idea: a floating edit icon on the left or right would probably solve this issue - with fold out elements n stuff # Screenshots: ![image](https://user-images.githubusercontent.com/65756/29658294-a3867eda-88ba-11e7-9199-d1d1fdb03a50.png) # Where (url) - /admin/... # Modules Installed: (just the important ones)
1.0
edit tabs on the frontned - # The issue is: Tabs on the frontend is taking up a lot of space + they are fugly to look at idea: a floating edit icon on the left or right would probably solve this issue - with fold out elements n stuff # Screenshots: ![image](https://user-images.githubusercontent.com/65756/29658294-a3867eda-88ba-11e7-9199-d1d1fdb03a50.png) # Where (url) - /admin/... # Modules Installed: (just the important ones)
non_process
edit tabs on the frontned the issue is tabs on the frontend is taking up a lot of space they are fugly to look at idea a floating edit icon on the left or right would probably solve this issue with fold out elements n stuff screenshots where url admin modules installed just the important ones
0
20,896
3,644,910,503
IssuesEvent
2016-02-15 12:13:20
coder-molok/foowd_alpha2
https://api.github.com/repos/coder-molok/foowd_alpha2
opened
Correzioni minori form registrazione
design ELGG
Compare un flag "nome inserito non corretto" quando si seleziona il campo "Nome utente", e non va più via anche se il nome è valido ![nome registrazione](https://cloud.githubusercontent.com/assets/13355694/13048368/b03d1714-d3e5-11e5-8e7b-3093c634e0b0.png) Il campo "sito internet" accetta solo domini. Serve che prenda anche indirizzi di pagine perchè potrebbe essere che i produttori vogliano linkare la propria pagina facebook. ![nome registrazione](https://cloud.githubusercontent.com/assets/13355694/13048392/dd2b3986-d3e5-11e5-993f-46dee5d881ea.png)
1.0
Correzioni minori form registrazione - Compare un flag "nome inserito non corretto" quando si seleziona il campo "Nome utente", e non va più via anche se il nome è valido ![nome registrazione](https://cloud.githubusercontent.com/assets/13355694/13048368/b03d1714-d3e5-11e5-8e7b-3093c634e0b0.png) Il campo "sito internet" accetta solo domini. Serve che prenda anche indirizzi di pagine perchè potrebbe essere che i produttori vogliano linkare la propria pagina facebook. ![nome registrazione](https://cloud.githubusercontent.com/assets/13355694/13048392/dd2b3986-d3e5-11e5-993f-46dee5d881ea.png)
non_process
correzioni minori form registrazione compare un flag nome inserito non corretto quando si seleziona il campo nome utente e non va più via anche se il nome è valido il campo sito internet accetta solo domini serve che prenda anche indirizzi di pagine perchè potrebbe essere che i produttori vogliano linkare la propria pagina facebook
0
690,497
23,661,863,445
IssuesEvent
2022-08-26 16:21:30
TheYellowArchitect/doubledamnation
https://api.github.com/repos/TheYellowArchitect/doubledamnation
opened
Secret Ending 1 Alternative Intro Voiceline Swap
low priority
There are five endings in the game. This is the second hardest. The camera of the intro is flipped/inverted, to signify the difference, but imo, all voicelines should be swapped ("Abandoned by the gods" should be spoken by P2/Mage, not P1/Warrior)
1.0
Secret Ending 1 Alternative Intro Voiceline Swap - There are five endings in the game. This is the second hardest. The camera of the intro is flipped/inverted, to signify the difference, but imo, all voicelines should be swapped ("Abandoned by the gods" should be spoken by P2/Mage, not P1/Warrior)
non_process
secret ending alternative intro voiceline swap there are five endings in the game this is the second hardest the camera of the intro is flipped inverted to signify the difference but imo all voicelines should be swapped abandoned by the gods should be spoken by mage not warrior
0
690,608
23,665,571,490
IssuesEvent
2022-08-26 20:28:39
googleapis/doc-pipeline
https://api.github.com/repos/googleapis/doc-pipeline
closed
generate: docfx-python-assuredworkloads-0.2.0.tar.gz failed
type: bug priority: p1 flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: d542fc9c171ed5d1f13eca605ad9f516637aaef6 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/feb651ce-9f7d-40ce-acc8-a682802ce480), [Sponge](http://sponge2/feb651ce-9f7d-40ce-acc8-a682802ce480) status: failed
1.0
generate: docfx-python-assuredworkloads-0.2.0.tar.gz failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: d542fc9c171ed5d1f13eca605ad9f516637aaef6 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/feb651ce-9f7d-40ce-acc8-a682802ce480), [Sponge](http://sponge2/feb651ce-9f7d-40ce-acc8-a682802ce480) status: failed
non_process
generate docfx python assuredworkloads tar gz failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed
0
2,051
4,861,287,471
IssuesEvent
2016-11-14 08:09:28
triplea-game/triplea
https://api.github.com/repos/triplea-game/triplea
opened
Non-Compatible Changes - Next release branch?
Discussion Process
I'm concerned/thinking about the next PR that breaks serialization/RMI and would require a major release. - I would like for us to ask PR submitters to test version compatibility. This means logging in to the lobby and seeing if a game can launch and chat works (this tests java RMI for the most part). Ideally we'd automate or make this no longer required.. Some effort needs to be put into the docs so we have a single /clear place to put these instructions too. I'll see if I can help with that in the next week. - For changes that we do want, that must break compatibility, I suggest we merge them to a single well known branch. We then merge that to master when we are ready to break version compatiblility. We'll need/want to merge master into this 'next-release' branch frequently, (each each PR merge ideally).
1.0
Non-Compatible Changes - Next release branch? - I'm concerned/thinking about the next PR that breaks serialization/RMI and would require a major release. - I would like for us to ask PR submitters to test version compatibility. This means logging in to the lobby and seeing if a game can launch and chat works (this tests java RMI for the most part). Ideally we'd automate or make this no longer required.. Some effort needs to be put into the docs so we have a single /clear place to put these instructions too. I'll see if I can help with that in the next week. - For changes that we do want, that must break compatibility, I suggest we merge them to a single well known branch. We then merge that to master when we are ready to break version compatiblility. We'll need/want to merge master into this 'next-release' branch frequently, (each each PR merge ideally).
process
non compatible changes next release branch i m concerned thinking about the next pr that breaks serialization rmi and would require a major release i would like for us to ask pr submitters to test version compatibility this means logging in to the lobby and seeing if a game can launch and chat works this tests java rmi for the most part ideally we d automate or make this no longer required some effort needs to be put into the docs so we have a single clear place to put these instructions too i ll see if i can help with that in the next week for changes that we do want that must break compatibility i suggest we merge them to a single well known branch we then merge that to master when we are ready to break version compatiblility we ll need want to merge master into this next release branch frequently each each pr merge ideally
1
14,328
17,362,458,021
IssuesEvent
2021-07-29 23:17:53
googleapis/google-auth-library-java
https://api.github.com/repos/googleapis/google-auth-library-java
closed
Setup end-to-end integration tests
type: process
We should be testing this library in different execution environments (GCE, GAE, etc)
1.0
Setup end-to-end integration tests - We should be testing this library in different execution environments (GCE, GAE, etc)
process
setup end to end integration tests we should be testing this library in different execution environments gce gae etc
1
8,977
12,093,585,419
IssuesEvent
2020-04-19 20:14:21
Pretronic/PretronicLibraries
https://api.github.com/repos/Pretronic/PretronicLibraries
closed
Copy utility
In processing global-utility
Create a copy utility based on reflections. - [x] Copy normal classes - [x] Deep Copy option - [x] Annotations for extra handlers - [x] CopyAble and DeepCopyAble interface (Self copy method) - [ ] Adapters for special objects - [ ] Default Adapter for List - [ ] Default Adapter for Map - [ ] Default Adapter for Collection - [ ] Default Adapter for Set
1.0
Copy utility - Create a copy utility based on reflections. - [x] Copy normal classes - [x] Deep Copy option - [x] Annotations for extra handlers - [x] CopyAble and DeepCopyAble interface (Self copy method) - [ ] Adapters for special objects - [ ] Default Adapter for List - [ ] Default Adapter for Map - [ ] Default Adapter for Collection - [ ] Default Adapter for Set
process
copy utility create a copy utility based on reflections copy normal classes deep copy option annotations for extra handlers copyable and deepcopyable interface self copy method adapters for special objects default adapter for list default adapter for map default adapter for collection default adapter for set
1
570,245
17,023,071,492
IssuesEvent
2021-07-03 00:15:28
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
ability to not show landsat in the applet
Component: applet Priority: minor Resolution: wontfix Type: enhancement
**[Submitted to the original trac issue database at 3.19pm, Thursday, 10th November 2005]** simplest would be to request the WMS layers seperately and show/hide them?
1.0
ability to not show landsat in the applet - **[Submitted to the original trac issue database at 3.19pm, Thursday, 10th November 2005]** simplest would be to request the WMS layers seperately and show/hide them?
non_process
ability to not show landsat in the applet simplest would be to request the wms layers seperately and show hide them
0
21,102
28,056,452,302
IssuesEvent
2023-03-29 09:41:12
camunda/issues
https://api.github.com/repos/camunda/issues
opened
Support OIDC for Elasticsearch in Self-Managed
component:distribution component:operate component:optimize component:tasklist component:zeebe component:zeebe-process-automation public kind:epic potential:8.3
### Value Proposition Statement Secure connections to Elasticsearch using OpenIDConnect in Self-Managed ### User Problem Today connection between Webapps & Zeebe Elastic Exporter can only use basic authentication. Nowadays organizations have often policies that forbid using Basic Authentication and that rely on token-based authentication mechanisms. They expect to be able to use SAML and/or OpenIDConnect. Currently this is not supported and prevents adoption of our Platform for some customers. ### User Stories I can use OpenIDConnect for connecting to Elasticsearch in Zeebe Elastic Exporter. I can use OpenIDConnect for connecting to Elasticsearch in Operate. I can use OpenIDConnect for connecting to Elasticsearch in Tasklist. I can use OpenIDConnect for connecting to Elasticsearch in Optimize. ### Implementation Notes From user perspective the best would be if this is just a configuration and not something I have to implement.
1.0
Support OIDC for Elasticsearch in Self-Managed - ### Value Proposition Statement Secure connections to Elasticsearch using OpenIDConnect in Self-Managed ### User Problem Today connection between Webapps & Zeebe Elastic Exporter can only use basic authentication. Nowadays organizations have often policies that forbid using Basic Authentication and that rely on token-based authentication mechanisms. They expect to be able to use SAML and/or OpenIDConnect. Currently this is not supported and prevents adoption of our Platform for some customers. ### User Stories I can use OpenIDConnect for connecting to Elasticsearch in Zeebe Elastic Exporter. I can use OpenIDConnect for connecting to Elasticsearch in Operate. I can use OpenIDConnect for connecting to Elasticsearch in Tasklist. I can use OpenIDConnect for connecting to Elasticsearch in Optimize. ### Implementation Notes From user perspective the best would be if this is just a configuration and not something I have to implement.
process
support oidc for elasticsearch in self managed value proposition statement secure connections to elasticsearch using openidconnect in self managed user problem today connection between webapps zeebe elastic exporter can only use basic authentication nowadays organizations have often policies that forbid using basic authentication and that rely on token based authentication mechanisms they expect to be able to use saml and or openidconnect currently this is not supported and prevents adoption of our platform for some customers user stories i can use openidconnect for connecting to elasticsearch in zeebe elastic exporter i can use openidconnect for connecting to elasticsearch in operate i can use openidconnect for connecting to elasticsearch in tasklist i can use openidconnect for connecting to elasticsearch in optimize implementation notes from user perspective the best would be if this is just a configuration and not something i have to implement
1
296,076
22,287,757,962
IssuesEvent
2022-06-11 22:46:10
CR6Community/CR-6-touchscreen
https://api.github.com/repos/CR6Community/CR-6-touchscreen
closed
Strange screen is displayed on Touch Screen firmware flashing.
documentation
I just ordered an additional touch screen for the CR-6 SE and it was delivered. While flashing the new CR-6 SE's touch screen using the 61F_RC_290422_v1 released by CR6Community/CR-6-touchscreen, near completion, a white noise-like screen appears. ![Flashing3](https://user-images.githubusercontent.com/96027590/167519966-102e1db4-7a0e-4ded-b7ad-90b4544b3786.jpg) It looks like the flashing is almost done right before the white noise display. ![Flashing1](https://user-images.githubusercontent.com/96027590/167519988-3b1d0622-ff83-48cf-82b3-bf2dc231bf62.JPG) ![Flashing2](https://user-images.githubusercontent.com/96027590/167519995-3125169b-2880-4b3f-8e69-072a9afd001d.JPG) The new touch screen uses the same PCB as the old LCD. However, flashing completion is indicated differently. However, if I wait a few tens of seconds and turn on the CR-6 SE, it operates normally (displaying icons, selecting menus, etc.). First, to check the kernel version, I checked the kernel version by flashing using an empty DWIN_SET, and at this time, the normal completion screen was displayed. After flashing using the contents of DWIN_SET, a strange completion screen is displayed. ![Flashing End](https://user-images.githubusercontent.com/96027590/167520048-bb61e388-3d2b-42a4-8ea0-3b7c47d87ef6.jpg) On the touch screen I had it shows the normal completion screen. It is strange because it is a symptom that appears on a new touch screen that uses the same PCB.
1.0
Strange screen is displayed on Touch Screen firmware flashing. - I just ordered an additional touch screen for the CR-6 SE and it was delivered. While flashing the new CR-6 SE's touch screen using the 61F_RC_290422_v1 released by CR6Community/CR-6-touchscreen, near completion, a white noise-like screen appears. ![Flashing3](https://user-images.githubusercontent.com/96027590/167519966-102e1db4-7a0e-4ded-b7ad-90b4544b3786.jpg) It looks like the flashing is almost done right before the white noise display. ![Flashing1](https://user-images.githubusercontent.com/96027590/167519988-3b1d0622-ff83-48cf-82b3-bf2dc231bf62.JPG) ![Flashing2](https://user-images.githubusercontent.com/96027590/167519995-3125169b-2880-4b3f-8e69-072a9afd001d.JPG) The new touch screen uses the same PCB as the old LCD. However, flashing completion is indicated differently. However, if I wait a few tens of seconds and turn on the CR-6 SE, it operates normally (displaying icons, selecting menus, etc.). First, to check the kernel version, I checked the kernel version by flashing using an empty DWIN_SET, and at this time, the normal completion screen was displayed. After flashing using the contents of DWIN_SET, a strange completion screen is displayed. ![Flashing End](https://user-images.githubusercontent.com/96027590/167520048-bb61e388-3d2b-42a4-8ea0-3b7c47d87ef6.jpg) On the touch screen I had it shows the normal completion screen. It is strange because it is a symptom that appears on a new touch screen that uses the same PCB.
non_process
strange screen is displayed on touch screen firmware flashing i just ordered an additional touch screen for the cr se and it was delivered while flashing the new cr se s touch screen using the rc released by cr touchscreen near completion a white noise like screen appears it looks like the flashing is almost done right before the white noise display the new touch screen uses the same pcb as the old lcd however flashing completion is indicated differently however if i wait a few tens of seconds and turn on the cr se it operates normally displaying icons selecting menus etc first to check the kernel version i checked the kernel version by flashing using an empty dwin set and at this time the normal completion screen was displayed after flashing using the contents of dwin set a strange completion screen is displayed on the touch screen i had it shows the normal completion screen it is strange because it is a symptom that appears on a new touch screen that uses the same pcb
0
8,594
11,758,834,040
IssuesEvent
2020-03-13 16:07:00
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
Support current Intel x86/x64 manuals (again)
Feature: Processor/x86 Type: Enhancement
Version 9.1 of Ghidra has switched to Intel's 325383-60US manual for the instruction set reference. Thank you. This is still available on Intel's website. However they appear to have issued 325383-70US in May of this year. It is what you get if you start here: https://software.intel.com/en-us/articles/intel-sdm. The two manuals are off by 1 page on the last opcode so it's not a big deal. However, someone might want to consider updating the index file sometime.
1.0
Support current Intel x86/x64 manuals (again) - Version 9.1 of Ghidra has switched to Intel's 325383-60US manual for the instruction set reference. Thank you. This is still available on Intel's website. However they appear to have issued 325383-70US in May of this year. It is what you get if you start here: https://software.intel.com/en-us/articles/intel-sdm. The two manuals are off by 1 page on the last opcode so it's not a big deal. However, someone might want to consider updating the index file sometime.
process
support current intel manuals again version of ghidra has switched to intel s manual for the instruction set reference thank you this is still available on intel s website however they appear to have issued in may of this year it is what you get if you start here the two manuals are off by page on the last opcode so it s not a big deal however someone might want to consider updating the index file sometime
1
8,682
11,811,459,158
IssuesEvent
2020-03-19 18:14:41
googleapis/java-mediatranslation
https://api.github.com/repos/googleapis/java-mediatranslation
opened
Switch samples/snippets/pom.xml to use libraries-bom
type: process
We cannot suggest using the libraries-bom here until this library is included in the libraries-bom
1.0
Switch samples/snippets/pom.xml to use libraries-bom - We cannot suggest using the libraries-bom here until this library is included in the libraries-bom
process
switch samples snippets pom xml to use libraries bom we cannot suggest using the libraries bom here until this library is included in the libraries bom
1
348,434
10,442,372,910
IssuesEvent
2019-09-18 12:57:40
getkirby/kirby
https://api.github.com/repos/getkirby/kirby
closed
Json::encode escapes unicode entities
priority: low 🐌 type: enhancement ✨
**To Reproduce** Steps to reproduce the behavior: 1. Put this in a template: ```php echo \Kirby\Data\Json::encode('здравей'); ``` 2. Echoed string is: ``` \u0437\u0434\u0440\u0430\u0432\u0435\u0439 ``` **Expected behavior** The encoded string should be `здравей` **Kirby Version** 3.2.3 **Additional context** This can be fixed by simply adding [a flag](https://www.php.net/manual/en/json.constants.php) in `Json::encode()` (tested): ```php json_encode($data, JSON_UNESCAPED_UNICODE); ```
1.0
Json::encode escapes unicode entities - **To Reproduce** Steps to reproduce the behavior: 1. Put this in a template: ```php echo \Kirby\Data\Json::encode('здравей'); ``` 2. Echoed string is: ``` \u0437\u0434\u0440\u0430\u0432\u0435\u0439 ``` **Expected behavior** The encoded string should be `здравей` **Kirby Version** 3.2.3 **Additional context** This can be fixed by simply adding [a flag](https://www.php.net/manual/en/json.constants.php) in `Json::encode()` (tested): ```php json_encode($data, JSON_UNESCAPED_UNICODE); ```
non_process
json encode escapes unicode entities to reproduce steps to reproduce the behavior put this in a template php echo kirby data json encode здравей echoed string is expected behavior the encoded string should be здравей kirby version additional context this can be fixed by simply adding in json encode tested php json encode data json unescaped unicode
0
11,113
13,957,681,439
IssuesEvent
2020-10-24 08:07:25
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
DE: request regarding XML schema validation of metadata records
DE - Germany Geoportal Harvesting process
Related to issue #3563 note-47 to note-50 Dear Daniele, We are planning to change our schema-validation-file for the harvest process (from apiso.xsd version 1.0.0 to apiso.xsd version 1.0.1). To be on the save site and don&#39;t &quot;loose&quot; any records during the harvest process for the INSPIRE Geoportal, it is important for us to know, which schema you are use for the INSPIRE Geoportal. Angelo Quaglia wrote us, that during the harvest process a validation or check for data sets (gmd) and services (srv) is against both schemata http://schemas.opengis.net/iso/19139/20060504/ and http://schemas.opengis.net/iso/19139/20070417/. Referring to this we have a couple of questions How long you will support both schemata http://schemas.opengis.net/iso/19139/20060504/ and http://schemas.opengis.net/iso/19139/20070417/? Are you planning to stop supporting the schema http://schemas.opengis.net/iso/19139/20060504/ in the (near) future? Thanks in advance and best regards, Anja (on behalf of Coordination Office SDI Germany)
1.0
DE: request regarding XML schema validation of metadata records - Related to issue #3563 note-47 to note-50 Dear Daniele, We are planning to change our schema-validation-file for the harvest process (from apiso.xsd version 1.0.0 to apiso.xsd version 1.0.1). To be on the save site and don&#39;t &quot;loose&quot; any records during the harvest process for the INSPIRE Geoportal, it is important for us to know, which schema you are use for the INSPIRE Geoportal. Angelo Quaglia wrote us, that during the harvest process a validation or check for data sets (gmd) and services (srv) is against both schemata http://schemas.opengis.net/iso/19139/20060504/ and http://schemas.opengis.net/iso/19139/20070417/. Referring to this we have a couple of questions How long you will support both schemata http://schemas.opengis.net/iso/19139/20060504/ and http://schemas.opengis.net/iso/19139/20070417/? Are you planning to stop supporting the schema http://schemas.opengis.net/iso/19139/20060504/ in the (near) future? Thanks in advance and best regards, Anja (on behalf of Coordination Office SDI Germany)
process
de request regarding xml schema validation of metadata records related to issue note to note dear daniele we are planning to change our schema validation file for the harvest process from apiso xsd version to apiso xsd version to be on the save site and don t quot loose quot any records during the harvest process for the inspire geoportal it is important for us to know which schema you are use for the inspire geoportal angelo quaglia wrote us that during the harvest process a validation or check for data sets gmd and services srv is against both schemata and referring to this we have a couple of questions how long you will support both schemata and are you planning to stop supporting the schema in the near future thanks in advance and best regards anja on behalf of coordination office sdi germany
1
344,787
10,349,640,108
IssuesEvent
2019-09-04 23:18:11
oslc-op/jira-migration-landfill
https://api.github.com/repos/oslc-op/jira-migration-landfill
closed
literal_value of the oslc_where syntax is not well-defined
Core: Query Priority: High Xtra: Jira
The spec is not clear on how to interpret the literals w/o the xsd data type. E.g. The terms boolean and decimal are short forms for typed literals. For example, true is a short form for "true"^xsd:booleancode>, 42 is a short form for "42"xsd:integer and 3.14159 is a short form for "3.14159"^xsd:decimal. does not specify how I am supposed to know whether 42 is an integer but 3.14 is a decimal (or a single-precision float?), let alone how I am supposed to ensure that ‘true‘ is a boolean True, not a "true" string literal. --- _Migrated from https://issues.oasis-open.org/browse/OSLCCORE-134 (opened by @berezovskyi; previously assigned to @oslc-bot)_
1.0
literal_value of the oslc_where syntax is not well-defined - The spec is not clear on how to interpret the literals w/o the xsd data type. E.g. The terms boolean and decimal are short forms for typed literals. For example, true is a short form for "true"^xsd:booleancode>, 42 is a short form for "42"xsd:integer and 3.14159 is a short form for "3.14159"^xsd:decimal. does not specify how I am supposed to know whether 42 is an integer but 3.14 is a decimal (or a single-precision float?), let alone how I am supposed to ensure that ‘true‘ is a boolean True, not a "true" string literal. --- _Migrated from https://issues.oasis-open.org/browse/OSLCCORE-134 (opened by @berezovskyi; previously assigned to @oslc-bot)_
non_process
literal value of the oslc where syntax is not well defined the spec is not clear on how to interpret the literals w o the xsd data type e g the terms boolean and decimal are short forms for typed literals for example true is a short form for true xsd booleancode is a short form for xsd integer and is a short form for xsd decimal does not specify how i am supposed to know whether is an integer but is a decimal or a single precision float let alone how i am supposed to ensure that ‘true‘ is a boolean true not a true string literal migrated from opened by berezovskyi previously assigned to oslc bot
0
11,592
14,447,380,996
IssuesEvent
2020-12-08 03:37:45
A01731346/5a
https://api.github.com/repos/A01731346/5a
closed
fill_size_estimating_template
process-dashboard
- Llenado de template de estimación de líneas de código en process dashboard - Correr el PROBE Wizard
1.0
fill_size_estimating_template - - Llenado de template de estimación de líneas de código en process dashboard - Correr el PROBE Wizard
process
fill size estimating template llenado de template de estimación de líneas de código en process dashboard correr el probe wizard
1
301,800
26,101,935,283
IssuesEvent
2022-12-27 08:18:30
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
opened
Release 4.4.0 - Alpha 2 - E2E UX tests - Wazuh Indexer
type/test/manual release test/4.4.0
The following issue aims to run the specified test for the current release candidate, report the results, and open new issues for any encountered errors. ## Test information | | | |-------------------------|--------------------------------------------| | **Test name** | Wazuh Indexer | | **Category** | Installation | | **Deployment option** | Single Indexer and agent, Multi Server: Step by step | | **Main release issue** | https://github.com/wazuh/wazuh/issues/15749 | | **Release candidate #** | Alpha 2 | | **Previous issue** | https://github.com/wazuh/wazuh/issues/15534 | ## Environment | **Component** | **OS** | **Installation** | **Type** | |-|-|-|-| | Wazuh dashboard | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-dashboard/step-by-step.html) | - | | Wazuh indexer | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-indexer/step-by-step.html) | Single node | | Wazuh server | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-server/step-by-step.html) | Multi node | | Wazuh agent | Amazon Linux 2 | [Installing Wazuh agents](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-agent/index.html) | - | ## Test description Best effort to test Wazuh indexer package. Think critically and at least review/test: - Wazuh indexer package specs - Indexer package size - Indexer package metadata (description) - Indexer package digital signature - Installed files location, size and permissions - Installation footprint (check that no unnecessary files are modified/broken in the file system. For example that operating system files do keep their right owner/pemissions and that the installer did not break the system.) - Installed Wazuh indexer service - Wazuh indexer logs when installed - Wazuh indexer templates and indices created - Wazuh indexer configuration (e.g. replicas are expected to be zero by default, how many shards per index,...) Try to compare and find anomalies with the previous Wazuh indexer version using appropiate E2E UX issue. Write down and report as much information as possible to allow comparison between versions using this issue. - Wazuh indexer cluster node communication and configuration - Wazuh indexer cluster status - Wazuh indexer packages uninstallation procedure ## Test report procedure All test results must have one of the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed result. | | :yellow_circle: | There is at least one expected failure or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. An extended report of the test results must be attached as a ZIP or TXT file. Please attach any documents, screenshots, or tables to the issue update with the results. This report can be used by the auditors to dig deeper into any possible failures and details. ## Conclusions | **Status** | **Test** | **Failure type** | **Notes** | |----------------|-------------|---------------------|----------------| | ⚫ | [Environment installation]() | | | | ⚫ | [Wazuh indexer package information]() | | | | ⚫ | [Installed files location, size and permissions]() | | | | ⚫ | [Installation footprint]() | | | | ⚫ | [Wazuh indexer service]() | | | | ⚫ | [Wazuh indexer installation logs]() | | Found some warnings that don't affect the performance | | ⚫ | [Wazuh indexer indices, templates, and shards]() | | | | ⚫ | [Wazuh indexer cluster status]() | | | | ⚫ | [Uninstall procedure]() | | | | ⚫ | [E2E dataflow]() | | | ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. - [ ] @wazuh/cicd - [ ] @davidjiglesias
2.0
Release 4.4.0 - Alpha 2 - E2E UX tests - Wazuh Indexer - The following issue aims to run the specified test for the current release candidate, report the results, and open new issues for any encountered errors. ## Test information | | | |-------------------------|--------------------------------------------| | **Test name** | Wazuh Indexer | | **Category** | Installation | | **Deployment option** | Single Indexer and agent, Multi Server: Step by step | | **Main release issue** | https://github.com/wazuh/wazuh/issues/15749 | | **Release candidate #** | Alpha 2 | | **Previous issue** | https://github.com/wazuh/wazuh/issues/15534 | ## Environment | **Component** | **OS** | **Installation** | **Type** | |-|-|-|-| | Wazuh dashboard | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-dashboard/step-by-step.html) | - | | Wazuh indexer | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-indexer/step-by-step.html) | Single node | | Wazuh server | Amazon Linux 2 | [Step by step](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-server/step-by-step.html) | Multi node | | Wazuh agent | Amazon Linux 2 | [Installing Wazuh agents](https://documentation-dev.wazuh.com/current/installation-guide/wazuh-agent/index.html) | - | ## Test description Best effort to test Wazuh indexer package. Think critically and at least review/test: - Wazuh indexer package specs - Indexer package size - Indexer package metadata (description) - Indexer package digital signature - Installed files location, size and permissions - Installation footprint (check that no unnecessary files are modified/broken in the file system. For example that operating system files do keep their right owner/pemissions and that the installer did not break the system.) - Installed Wazuh indexer service - Wazuh indexer logs when installed - Wazuh indexer templates and indices created - Wazuh indexer configuration (e.g. replicas are expected to be zero by default, how many shards per index,...) Try to compare and find anomalies with the previous Wazuh indexer version using appropiate E2E UX issue. Write down and report as much information as possible to allow comparison between versions using this issue. - Wazuh indexer cluster node communication and configuration - Wazuh indexer cluster status - Wazuh indexer packages uninstallation procedure ## Test report procedure All test results must have one of the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed result. | | :yellow_circle: | There is at least one expected failure or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. An extended report of the test results must be attached as a ZIP or TXT file. Please attach any documents, screenshots, or tables to the issue update with the results. This report can be used by the auditors to dig deeper into any possible failures and details. ## Conclusions | **Status** | **Test** | **Failure type** | **Notes** | |----------------|-------------|---------------------|----------------| | ⚫ | [Environment installation]() | | | | ⚫ | [Wazuh indexer package information]() | | | | ⚫ | [Installed files location, size and permissions]() | | | | ⚫ | [Installation footprint]() | | | | ⚫ | [Wazuh indexer service]() | | | | ⚫ | [Wazuh indexer installation logs]() | | Found some warnings that don't affect the performance | | ⚫ | [Wazuh indexer indices, templates, and shards]() | | | | ⚫ | [Wazuh indexer cluster status]() | | | | ⚫ | [Uninstall procedure]() | | | | ⚫ | [E2E dataflow]() | | | ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. - [ ] @wazuh/cicd - [ ] @davidjiglesias
non_process
release alpha ux tests wazuh indexer the following issue aims to run the specified test for the current release candidate report the results and open new issues for any encountered errors test information test name wazuh indexer category installation deployment option single indexer and agent multi server step by step main release issue release candidate alpha previous issue environment component os installation type wazuh dashboard amazon linux wazuh indexer amazon linux single node wazuh server amazon linux multi node wazuh agent amazon linux test description best effort to test wazuh indexer package think critically and at least review test wazuh indexer package specs indexer package size indexer package metadata description indexer package digital signature installed files location size and permissions installation footprint check that no unnecessary files are modified broken in the file system for example that operating system files do keep their right owner pemissions and that the installer did not break the system installed wazuh indexer service wazuh indexer logs when installed wazuh indexer templates and indices created wazuh indexer configuration e g replicas are expected to be zero by default how many shards per index try to compare and find anomalies with the previous wazuh indexer version using appropiate ux issue write down and report as much information as possible to allow comparison between versions using this issue wazuh indexer cluster node communication and configuration wazuh indexer cluster status wazuh indexer packages uninstallation procedure test report procedure all test results must have one of the following statuses green circle all checks passed red circle there is at least one failed result yellow circle there is at least one expected failure or skipped test and no failures any failing test must be properly addressed with a new issue detailing the error and the possible cause an extended report of the test results must be attached as a zip or txt file please attach any documents screenshots or tables to the issue update with the results this report can be used by the auditors to dig deeper into any possible failures and details conclusions status test failure type notes ⚫ ⚫ ⚫ ⚫ ⚫ ⚫ found some warnings that don t affect the performance ⚫ ⚫ ⚫ ⚫   auditors validation the definition of done for this one is the validation of the conclusions and the test results from all auditors wazuh cicd davidjiglesias
0
13,867
16,623,137,571
IssuesEvent
2021-06-03 05:57:28
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Refactor incorrectly converts from int64 to int8
Bug Processing
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When refactoring **_non-fid_** fields within a Geopackage attribute table, the output table's _**fid**_ field is also incorrectly converted from **Integer64** to **Integer8**, without the user requesting. This error only happens when the table contains records; empty tables correctly output the _**fid**_ as **Integer64**. This problem occurs with layer attribute tables as well as standalone attribute tables. **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> 1. Refactor a Geopackage table 2. Review the output table's properties 3. Note that the output table's _**fid**_ field is now **Integer8**, when it should be **Integer64** I am attaching screenshots of input and output table properties. Input table: ![input_table_properties](https://user-images.githubusercontent.com/12805900/120008221-98387f00-bf8f-11eb-90b8-92876d6fbe57.jpg) Refactored table: ![refactored_table_properties](https://user-images.githubusercontent.com/12805900/120008457-d0d85880-bf8f-11eb-8db8-fefa336279c6.jpg) **QGIS and OS versions** <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> <!--StartFragment--> <meta http-equiv="Content-Type" content="text/html; charset=utf-8"><style type="text/css"> p, li { white-space: pre-wrap; } </style> QGIS version | 3.18.3-Zürich | QGIS code revision | 735cc85be9 -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4 Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3 Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0 PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020 OS Version | Windows 10 (10.0) Active python plugins | batch_hillshader-master; changeDataSource; DissolveWithStats; ee_plugin; GlobeBuilder; GroupStats; LAStools; latlontools; mapswipetool_plugin; mmqgis; MultiDistanceBuffer; NNJoin; PointConnector; pointstopaths; qdraw; qfieldsync; Qgis2threejs; qgis2web; QGISSortAndNumber-master; qgis_gee_data_catalog; QRectangleCreator; quick_map_services; refFunctions; SemiAutomaticClassificationPlugin; Serval; shapetools; SpreadsheetLayers; ViewshedAnalysis; db_manager; MetaSearch; processing <!--EndFragment--> **Additional context** I first reported this problem in Stackexchange at: https://gis.stackexchange.com/questions/397984/qgis-refactor-fields-curiosity <!-- Add any other context about the problem here. -->
1.0
Refactor incorrectly converts from int64 to int8 - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When refactoring **_non-fid_** fields within a Geopackage attribute table, the output table's _**fid**_ field is also incorrectly converted from **Integer64** to **Integer8**, without the user requesting. This error only happens when the table contains records; empty tables correctly output the _**fid**_ as **Integer64**. This problem occurs with layer attribute tables as well as standalone attribute tables. **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> 1. Refactor a Geopackage table 2. Review the output table's properties 3. Note that the output table's _**fid**_ field is now **Integer8**, when it should be **Integer64** I am attaching screenshots of input and output table properties. Input table: ![input_table_properties](https://user-images.githubusercontent.com/12805900/120008221-98387f00-bf8f-11eb-90b8-92876d6fbe57.jpg) Refactored table: ![refactored_table_properties](https://user-images.githubusercontent.com/12805900/120008457-d0d85880-bf8f-11eb-8db8-fefa336279c6.jpg) **QGIS and OS versions** <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> <!--StartFragment--> <meta http-equiv="Content-Type" content="text/html; charset=utf-8"><style type="text/css"> p, li { white-space: pre-wrap; } </style> QGIS version | 3.18.3-Zürich | QGIS code revision | 735cc85be9 -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4 Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3 Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0 PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020 OS Version | Windows 10 (10.0) Active python plugins | batch_hillshader-master; changeDataSource; DissolveWithStats; ee_plugin; GlobeBuilder; GroupStats; LAStools; latlontools; mapswipetool_plugin; mmqgis; MultiDistanceBuffer; NNJoin; PointConnector; pointstopaths; qdraw; qfieldsync; Qgis2threejs; qgis2web; QGISSortAndNumber-master; qgis_gee_data_catalog; QRectangleCreator; quick_map_services; refFunctions; SemiAutomaticClassificationPlugin; Serval; shapetools; SpreadsheetLayers; ViewshedAnalysis; db_manager; MetaSearch; processing <!--EndFragment--> **Additional context** I first reported this problem in Stackexchange at: https://gis.stackexchange.com/questions/397984/qgis-refactor-fields-curiosity <!-- Add any other context about the problem here. -->
process
refactor incorrectly converts from to bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug when refactoring non fid fields within a geopackage attribute table the output table s fid field is also incorrectly converted from to without the user requesting this error only happens when the table contains records empty tables correctly output the fid as this problem occurs with layer attribute tables as well as standalone attribute tables how to reproduce refactor a geopackage table review the output table s properties note that the output table s fid field is now when it should be i am attaching screenshots of input and output table properties input table refactored table qgis and os versions about click in the table ctrl a and then ctrl c finally paste here p li white space pre wrap qgis version zürich qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins batch hillshader master changedatasource dissolvewithstats ee plugin globebuilder groupstats lastools latlontools mapswipetool plugin mmqgis multidistancebuffer nnjoin pointconnector pointstopaths qdraw qfieldsync qgissortandnumber master qgis gee data catalog qrectanglecreator quick map services reffunctions semiautomaticclassificationplugin serval shapetools spreadsheetlayers viewshedanalysis db manager metasearch processing additional context i first reported this problem in stackexchange at
1
278,349
21,075,277,652
IssuesEvent
2022-04-02 03:39:20
Shopify/shopify-cli
https://api.github.com/repos/Shopify/shopify-cli
closed
Update the blog regarding CLI
area:documentation no-issue-activity
Per feedback provided on #762 > Finding the install instructions was a bit of a maze > The blog leads you to shopify.dev which doesn't actually mention CLI but refers to "tools" down the page - OK, click on that - there is a link to the CLI source on GitHub, with a documentation link in the readme — click — then click again on install instructions - finally. > > This was unnecessarily hard. Ideally, a landing page on shopify.dev that explains how to install on various platforms, linked directly from the blog post.
1.0
Update the blog regarding CLI - Per feedback provided on #762 > Finding the install instructions was a bit of a maze > The blog leads you to shopify.dev which doesn't actually mention CLI but refers to "tools" down the page - OK, click on that - there is a link to the CLI source on GitHub, with a documentation link in the readme — click — then click again on install instructions - finally. > > This was unnecessarily hard. Ideally, a landing page on shopify.dev that explains how to install on various platforms, linked directly from the blog post.
non_process
update the blog regarding cli per feedback provided on finding the install instructions was a bit of a maze the blog leads you to shopify dev which doesn t actually mention cli but refers to tools down the page ok click on that there is a link to the cli source on github with a documentation link in the readme — click — then click again on install instructions finally this was unnecessarily hard ideally a landing page on shopify dev that explains how to install on various platforms linked directly from the blog post
0
231,417
7,632,155,584
IssuesEvent
2018-05-05 11:56:28
pzahemszky/sudoku
https://api.github.com/repos/pzahemszky/sudoku
opened
Separate primary and secondary peers
enhancement good first issue low priority
There could be a simple two-level hierarchy between peers: those operations that are more likely to be successful should be placed in front of the secondary operations in the queue of `remove_rearrange`. In particular, for the triple `(dig, row, col)` and corresponding `row_slice` and `col_slice` objects the below operations should be prioritised before any of the other ones. - `Operation('digcol', dig, c)`, if `c` is within `col_slice` - `Operation('digrow', dig, r)`, if `r` is within `row_slice` - `Operation('digbox', dig, b)`, if `b` intersects with `row` or `col`
1.0
Separate primary and secondary peers - There could be a simple two-level hierarchy between peers: those operations that are more likely to be successful should be placed in front of the secondary operations in the queue of `remove_rearrange`. In particular, for the triple `(dig, row, col)` and corresponding `row_slice` and `col_slice` objects the below operations should be prioritised before any of the other ones. - `Operation('digcol', dig, c)`, if `c` is within `col_slice` - `Operation('digrow', dig, r)`, if `r` is within `row_slice` - `Operation('digbox', dig, b)`, if `b` intersects with `row` or `col`
non_process
separate primary and secondary peers there could be a simple two level hierarchy between peers those operations that are more likely to be successful should be placed in front of the secondary operations in the queue of remove rearrange in particular for the triple dig row col and corresponding row slice and col slice objects the below operations should be prioritised before any of the other ones operation digcol dig c if c is within col slice operation digrow dig r if r is within row slice operation digbox dig b if b intersects with row or col
0
242,465
18,545,121,060
IssuesEvent
2021-10-21 20:59:20
ReznikovRoman/airbnb-clone
https://api.github.com/repos/ReznikovRoman/airbnb-clone
opened
[FEATURE] Improve documentation
documentation feature cleanup/optimization
**Description** - Change README.md file: makefile.env file is required (used by pre-commit) - Add docker-compose.yml file: populate database with fake data, run tests, etc. - Add guidelines.md file: specify project style guide: - Code style - Naming conventions - Project structure (models, views, services, etc.) **Possible solution** - **Additional context** -
1.0
[FEATURE] Improve documentation - **Description** - Change README.md file: makefile.env file is required (used by pre-commit) - Add docker-compose.yml file: populate database with fake data, run tests, etc. - Add guidelines.md file: specify project style guide: - Code style - Naming conventions - Project structure (models, views, services, etc.) **Possible solution** - **Additional context** -
non_process
improve documentation description change readme md file makefile env file is required used by pre commit add docker compose yml file populate database with fake data run tests etc add guidelines md file specify project style guide code style naming conventions project structure models views services etc possible solution additional context
0
22,510
31,562,740,230
IssuesEvent
2023-09-03 12:57:05
nextflow-io/nextflow
https://api.github.com/repos/nextflow-io/nextflow
closed
make inputs read-only
lang/processes good first issue
I've run into several hard-to-trace pipeline bugs caused by tasks inadvertently modifying input files that were staged in as symlinks or hardlinks. It would be good if Nextflow could make such inputs read-only before task execution, and restore their mode afterwards.
1.0
make inputs read-only - I've run into several hard-to-trace pipeline bugs caused by tasks inadvertently modifying input files that were staged in as symlinks or hardlinks. It would be good if Nextflow could make such inputs read-only before task execution, and restore their mode afterwards.
process
make inputs read only i ve run into several hard to trace pipeline bugs caused by tasks inadvertently modifying input files that were staged in as symlinks or hardlinks it would be good if nextflow could make such inputs read only before task execution and restore their mode afterwards
1
1,410
3,971,742,637
IssuesEvent
2016-05-04 13:10:47
openvstorage/openvstorage-health-check
https://api.github.com/repos/openvstorage/openvstorage-health-check
closed
Exception in halted volumes when volume is detached/unreachable
priority_critical process_duplicate type_bug
``` [INFO] Checking vPool 'env1newvpool': Traceback (most recent call last): File "<string>", line 1, in <module> File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda> __call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw) File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__ return self.run(*args, **kwargs) File "healthcheck.py", line 85, in check_attended return HealthCheckController.execute_check() File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda> __call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw) File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__ return self.run(*args, **kwargs) File "healthcheck.py", line 132, in execute_check HealthCheckController.check_openvstorage() File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda> __call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw) File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__ return self.run(*args, **kwargs) File "healthcheck.py", line 185, in check_openvstorage ovs.check_for_halted_volumes() File "/opt/OpenvStorage/ovs/extensions/healthcheck/openvstorage/openvstoragecluster_health_check.py", line 947, in check_for_halted_volumes if int(self.utility.convert_xml_to_json(voldrv_client.info_volume(volume))["boost_serialization"] volumedriver.storagerouter.storagerouterclient.MaxRedirectsExceededException: volumeInfo ```
1.0
Exception in halted volumes when volume is detached/unreachable - ``` [INFO] Checking vPool 'env1newvpool': Traceback (most recent call last): File "<string>", line 1, in <module> File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda> __call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw) File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__ return self.run(*args, **kwargs) File "healthcheck.py", line 85, in check_attended return HealthCheckController.execute_check() File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda> __call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw) File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__ return self.run(*args, **kwargs) File "healthcheck.py", line 132, in execute_check HealthCheckController.check_openvstorage() File "/usr/lib/python2.7/dist-packages/celery/local.py", line 167, in <lambda> __call__ = lambda x, *a, **kw: x._get_current_object()(*a, **kw) File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__ return self.run(*args, **kwargs) File "healthcheck.py", line 185, in check_openvstorage ovs.check_for_halted_volumes() File "/opt/OpenvStorage/ovs/extensions/healthcheck/openvstorage/openvstoragecluster_health_check.py", line 947, in check_for_halted_volumes if int(self.utility.convert_xml_to_json(voldrv_client.info_volume(volume))["boost_serialization"] volumedriver.storagerouter.storagerouterclient.MaxRedirectsExceededException: volumeInfo ```
process
exception in halted volumes when volume is detached unreachable checking vpool traceback most recent call last file line in file usr lib dist packages celery local py line in call lambda x a kw x get current object a kw file usr lib dist packages celery app task py line in call return self run args kwargs file healthcheck py line in check attended return healthcheckcontroller execute check file usr lib dist packages celery local py line in call lambda x a kw x get current object a kw file usr lib dist packages celery app task py line in call return self run args kwargs file healthcheck py line in execute check healthcheckcontroller check openvstorage file usr lib dist packages celery local py line in call lambda x a kw x get current object a kw file usr lib dist packages celery app task py line in call return self run args kwargs file healthcheck py line in check openvstorage ovs check for halted volumes file opt openvstorage ovs extensions healthcheck openvstorage openvstoragecluster health check py line in check for halted volumes if int self utility convert xml to json voldrv client info volume volume volumedriver storagerouter storagerouterclient maxredirectsexceededexception volumeinfo
1
13,730
5,435,976,267
IssuesEvent
2017-03-05 21:20:26
docker/docker
https://api.github.com/repos/docker/docker
closed
Possible docker 1.12.5 problem with Centos 7.3 upgrade
area/builder area/distribution area/runtime status/more-info-needed version/1.12
**Description** After running yum upgrade (to Centos 7.3) the Centos distribution docker 1.12.5 is unable to complete a docker build due to an oci runtime error apparently due to a missing /var/lib/docker/devicemapper/mnt/..../rootfs directory. Reverting back to a VM snapshot from ~ two weeks ago when docker was working normally after the snapshot was made and, with or without removing all containers / images / volumes, repeating the yum upgrade yields the same result upon trying docker build. **Steps to reproduce the issue:** 1. Have a Dockerfile containing: """ From alpine COPY . / RUN /docker-build/build CMD ["/bin/sleep","infinity"] """ 2. Have a docker-build/build file containing: """ #!/bin/sh set -ex apk update apk add bash apk add coreutils apk add go apk add git apk add make apk add xz """ 3. Attempt to build the image with docker build -t golang-build:latest . **Describe the results you received:** Docker pulled the base alpine normally, says it copied the build script into the image but threw the oci runtime error while executing the build script: """ $ docker build -t golang-build:latest . Sending build context to Docker daemon 4.608 kB Step 1 : FROM alpine Trying to pull repository docker.io/library/alpine ... latest: Pulling from docker.io/library/alpine 627beaf3eaaf: Pull complete Digest: sha256:58e1a1bb75db1b5a24a462dd5e2915277ea06438c3f105138f97eb53149673c4 ---> 4a415e366388 Step 2 : COPY . / ---> 641277111855 Removing intermediate container f49d4e90b877 Step 3 : RUN /docker-build/build ---> Running in 56e0411caa59 container_linux.go:247: starting container process caused "process_linux.go:359: container init caused \"rootfs_linux.go:54: mounting \\\"/var/lib/docker/containers/56e0411caa59d37303a221bd06cca92810250db517db7e06e92167cef383696f/secrets\\\" to rootfs \\\"/var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs\\\" at \\\"/run/secrets\\\" caused \\\"lstat /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs/run/secrets: not a directory\\\"\"" invalid header field value "oci runtime error: container_linux.go:247: starting container process caused \"process_linux.go:359: container init caused \\\"rootfs_linux.go:54: mounting \\\\\\\"/var/lib/docker/containers/56e0411caa59d37303a221bd06cca92810250db517db7e06e92167cef383696f/secrets\\\\\\\" to rootfs \\\\\\\"/var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs\\\\\\\" at \\\\\\\"/run/secrets\\\\\\\" caused \\\\\\\"lstat /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs/run/secrets: not a directory\\\\\\\"\\\"\"\n" """ The /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/ directory does not contain a rootfs directory $ docker ps -a CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 56e0411caa59 641277111855 "/bin/sh -c /docker-b" About a minute ago Created elegant_bohr $ docker images -a REPOSITORY TAG IMAGE ID CREATED SIZE `<none>` `<none>` 641277111855 2 minutes ago 3.984 MB docker.io/alpine latest 4a415e366388 43 hours ago 3.984 MB **Describe the results you expected:** Previously docker built the image and docker images -a would show docker.io/alpine and the build image with the specified name and tag would appear and there would be nothing in docker ps -a **Additional information you deem important (e.g. issue happens only occasionally):** Since the yum upgrade to Centos 7.3 I have been unable to do any docker builds due to the problem above. The name of the container appearing due to the docker build changes on every attempt. The os user is in the docker group to perform all docker operations as a non-privileged user. The closest existing issue I could find is https://github.com/docker/docker/issues/28109 but that appears to be a /sys/fs/cgroup problem which is not being reported here. **Output of `docker version`:** ``` Client: Version: 1.12.5 API version: 1.24 Package version: docker-common-1.12.5-14.el7.centos.x86_64 Go version: go1.7.4 Git commit: 047e51b/1.12.5 Built: Mon Jan 23 15:35:13 2017 OS/Arch: linux/amd64 Server: Version: 1.12.5 API version: 1.24 Package version: docker-common-1.12.5-14.el7.centos.x86_64 Go version: go1.7.4 Git commit: 047e51b/1.12.5 Built: Mon Jan 23 15:35:13 2017 OS/Arch: linux/amd64 ``` **Output of `docker info`:** ``` Containers: 1 Running: 0 Paused: 0 Stopped: 1 Images: 2 Server Version: 1.12.5 Storage Driver: devicemapper Pool Name: centos-docker--pool Pool Blocksize: 524.3 kB Base Device Size: 10.74 GB Backing Filesystem: xfs Data file: Metadata file: Data Space Used: 60.29 MB Data Space Total: 27.92 GB Data Space Available: 27.86 GB Metadata Space Used: 73.73 kB Metadata Space Total: 109.1 MB Metadata Space Available: 109 MB Thin Pool Minimum Free Space: 2.791 GB Udev Sync Supported: true Deferred Removal Enabled: true Deferred Deletion Enabled: false Deferred Deleted Device Count: 0 Library Version: 1.02.135-RHEL7 (2016-11-16) Logging Driver: journald Cgroup Driver: systemd Plugins: Volume: local Network: null bridge host overlay Swarm: inactive Runtimes: docker-runc runc Default Runtime: docker-runc Security Options: seccomp selinux Kernel Version: 3.10.0-514.10.2.el7.x86_64 Operating System: CentOS Linux 7 (Core) OSType: linux Architecture: x86_64 Number of Docker Hooks: 2 CPUs: 4 Total Memory: 3.702 GiB Name: localhost.localdomain ID: ITFX:BTFW:TKNH:XZ77:LXGN:2SD5:SGPE:ZC3B:COAB:PHMY:O2NR:Y3LZ Docker Root Dir: /var/lib/docker Debug Mode (client): false Debug Mode (server): false Registry: https://index.docker.io/v1/ Insecure Registries: 127.0.0.0/8 Registries: docker.io (secure) ``` **Additional environment details (AWS, VirtualBox, physical, etc.):** VirtualBox 5.0.32 running on Windows 7 64bit pro VM is Centos 7.3 $ rpm --query centos-release centos-release-7-3.1611.el7.centos.x86_64 Virtualbox guest additions are installed and appear to be working normally Disks are not full $ df Filesystem 1K-blocks Used Available Use% Mounted on /dev/mapper/centos-root 20961280 5822412 15138868 28% / devtmpfs 1925036 0 1925036 0% /dev tmpfs 1940864 80 1940784 1% /dev/shm tmpfs 1940864 8912 1931952 1% /run tmpfs 1940864 0 1940864 0% /sys/fs/cgroup /dev/sda1 508580 314972 193608 62% /boot /dev/mapper/centos-home 10475520 1051912 9423608 11% /home tmpfs 388176 4 388172 1% /run/user/992 VmShare 966214652 288862344 677352308 30% /media/sf_VmShare tmpfs 388176 24 388152 1% /run/user/1000 $ sudo lvdisplay --- Logical volume --- LV Path /dev/centos/root LV Name root VG Name centos LV UUID V3PdeT-eF3c-yd2l-rNNp-mSjE-KBIZ-7hlSdF LV Write Access read/write LV Creation host, time localhost.localdomain, 2017-02-10 22:41:57 +0000 LV Status available ` # open 1` LV Size 20.00 GiB Current LE 5120 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 8192 Block device 253:0 --- Logical volume --- LV Path /dev/centos/home LV Name home VG Name centos LV UUID cLdgQy-gRpY-vhiU-73hL-t1xP-2dyO-x1z4Fv LV Write Access read/write LV Creation host, time localhost.localdomain, 2017-02-10 22:41:59 +0000 LV Status available `# open 1` LV Size 10.00 GiB Current LE 2560 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 8192 Block device 253:2 --- Logical volume --- LV Path /dev/centos/swap LV Name swap VG Name centos LV UUID PqR0HU-TbW3-Ew4C-gET1-f7aP-OVRB-5xhx5H LV Write Access read/write LV Creation host, time localhost.localdomain, 2017-02-10 22:42:00 +0000 LV Status available `# open 2` LV Size 4.00 GiB Current LE 1024 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 8192 Block device 253:1 --- Logical volume --- LV Name docker-pool VG Name centos LV UUID TtSvhh-Yb8c-8yGo-dnHk-tWwm-4ujD-MejUJX LV Write Access read/write LV Creation host, time localhost.localdomain, 2017-02-10 23:02:07 +0000 LV Pool metadata docker-pool_tmeta LV Pool data docker-pool_tdata LV Status available `# open 0` LV Size 26.00 GiB Allocated pool data 0.22% Allocated metadata 0.07% Current LE 6656 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 8192 Block device 253:5
1.0
Possible docker 1.12.5 problem with Centos 7.3 upgrade - **Description** After running yum upgrade (to Centos 7.3) the Centos distribution docker 1.12.5 is unable to complete a docker build due to an oci runtime error apparently due to a missing /var/lib/docker/devicemapper/mnt/..../rootfs directory. Reverting back to a VM snapshot from ~ two weeks ago when docker was working normally after the snapshot was made and, with or without removing all containers / images / volumes, repeating the yum upgrade yields the same result upon trying docker build. **Steps to reproduce the issue:** 1. Have a Dockerfile containing: """ From alpine COPY . / RUN /docker-build/build CMD ["/bin/sleep","infinity"] """ 2. Have a docker-build/build file containing: """ #!/bin/sh set -ex apk update apk add bash apk add coreutils apk add go apk add git apk add make apk add xz """ 3. Attempt to build the image with docker build -t golang-build:latest . **Describe the results you received:** Docker pulled the base alpine normally, says it copied the build script into the image but threw the oci runtime error while executing the build script: """ $ docker build -t golang-build:latest . Sending build context to Docker daemon 4.608 kB Step 1 : FROM alpine Trying to pull repository docker.io/library/alpine ... latest: Pulling from docker.io/library/alpine 627beaf3eaaf: Pull complete Digest: sha256:58e1a1bb75db1b5a24a462dd5e2915277ea06438c3f105138f97eb53149673c4 ---> 4a415e366388 Step 2 : COPY . / ---> 641277111855 Removing intermediate container f49d4e90b877 Step 3 : RUN /docker-build/build ---> Running in 56e0411caa59 container_linux.go:247: starting container process caused "process_linux.go:359: container init caused \"rootfs_linux.go:54: mounting \\\"/var/lib/docker/containers/56e0411caa59d37303a221bd06cca92810250db517db7e06e92167cef383696f/secrets\\\" to rootfs \\\"/var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs\\\" at \\\"/run/secrets\\\" caused \\\"lstat /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs/run/secrets: not a directory\\\"\"" invalid header field value "oci runtime error: container_linux.go:247: starting container process caused \"process_linux.go:359: container init caused \\\"rootfs_linux.go:54: mounting \\\\\\\"/var/lib/docker/containers/56e0411caa59d37303a221bd06cca92810250db517db7e06e92167cef383696f/secrets\\\\\\\" to rootfs \\\\\\\"/var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs\\\\\\\" at \\\\\\\"/run/secrets\\\\\\\" caused \\\\\\\"lstat /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/rootfs/run/secrets: not a directory\\\\\\\"\\\"\"\n" """ The /var/lib/docker/devicemapper/mnt/a8aa9c7063f2d2e4d682096cb0eba25f783e43a14d25088c4ccbb4cd2451bc7d/ directory does not contain a rootfs directory $ docker ps -a CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 56e0411caa59 641277111855 "/bin/sh -c /docker-b" About a minute ago Created elegant_bohr $ docker images -a REPOSITORY TAG IMAGE ID CREATED SIZE `<none>` `<none>` 641277111855 2 minutes ago 3.984 MB docker.io/alpine latest 4a415e366388 43 hours ago 3.984 MB **Describe the results you expected:** Previously docker built the image and docker images -a would show docker.io/alpine and the build image with the specified name and tag would appear and there would be nothing in docker ps -a **Additional information you deem important (e.g. issue happens only occasionally):** Since the yum upgrade to Centos 7.3 I have been unable to do any docker builds due to the problem above. The name of the container appearing due to the docker build changes on every attempt. The os user is in the docker group to perform all docker operations as a non-privileged user. The closest existing issue I could find is https://github.com/docker/docker/issues/28109 but that appears to be a /sys/fs/cgroup problem which is not being reported here. **Output of `docker version`:** ``` Client: Version: 1.12.5 API version: 1.24 Package version: docker-common-1.12.5-14.el7.centos.x86_64 Go version: go1.7.4 Git commit: 047e51b/1.12.5 Built: Mon Jan 23 15:35:13 2017 OS/Arch: linux/amd64 Server: Version: 1.12.5 API version: 1.24 Package version: docker-common-1.12.5-14.el7.centos.x86_64 Go version: go1.7.4 Git commit: 047e51b/1.12.5 Built: Mon Jan 23 15:35:13 2017 OS/Arch: linux/amd64 ``` **Output of `docker info`:** ``` Containers: 1 Running: 0 Paused: 0 Stopped: 1 Images: 2 Server Version: 1.12.5 Storage Driver: devicemapper Pool Name: centos-docker--pool Pool Blocksize: 524.3 kB Base Device Size: 10.74 GB Backing Filesystem: xfs Data file: Metadata file: Data Space Used: 60.29 MB Data Space Total: 27.92 GB Data Space Available: 27.86 GB Metadata Space Used: 73.73 kB Metadata Space Total: 109.1 MB Metadata Space Available: 109 MB Thin Pool Minimum Free Space: 2.791 GB Udev Sync Supported: true Deferred Removal Enabled: true Deferred Deletion Enabled: false Deferred Deleted Device Count: 0 Library Version: 1.02.135-RHEL7 (2016-11-16) Logging Driver: journald Cgroup Driver: systemd Plugins: Volume: local Network: null bridge host overlay Swarm: inactive Runtimes: docker-runc runc Default Runtime: docker-runc Security Options: seccomp selinux Kernel Version: 3.10.0-514.10.2.el7.x86_64 Operating System: CentOS Linux 7 (Core) OSType: linux Architecture: x86_64 Number of Docker Hooks: 2 CPUs: 4 Total Memory: 3.702 GiB Name: localhost.localdomain ID: ITFX:BTFW:TKNH:XZ77:LXGN:2SD5:SGPE:ZC3B:COAB:PHMY:O2NR:Y3LZ Docker Root Dir: /var/lib/docker Debug Mode (client): false Debug Mode (server): false Registry: https://index.docker.io/v1/ Insecure Registries: 127.0.0.0/8 Registries: docker.io (secure) ``` **Additional environment details (AWS, VirtualBox, physical, etc.):** VirtualBox 5.0.32 running on Windows 7 64bit pro VM is Centos 7.3 $ rpm --query centos-release centos-release-7-3.1611.el7.centos.x86_64 Virtualbox guest additions are installed and appear to be working normally Disks are not full $ df Filesystem 1K-blocks Used Available Use% Mounted on /dev/mapper/centos-root 20961280 5822412 15138868 28% / devtmpfs 1925036 0 1925036 0% /dev tmpfs 1940864 80 1940784 1% /dev/shm tmpfs 1940864 8912 1931952 1% /run tmpfs 1940864 0 1940864 0% /sys/fs/cgroup /dev/sda1 508580 314972 193608 62% /boot /dev/mapper/centos-home 10475520 1051912 9423608 11% /home tmpfs 388176 4 388172 1% /run/user/992 VmShare 966214652 288862344 677352308 30% /media/sf_VmShare tmpfs 388176 24 388152 1% /run/user/1000 $ sudo lvdisplay --- Logical volume --- LV Path /dev/centos/root LV Name root VG Name centos LV UUID V3PdeT-eF3c-yd2l-rNNp-mSjE-KBIZ-7hlSdF LV Write Access read/write LV Creation host, time localhost.localdomain, 2017-02-10 22:41:57 +0000 LV Status available ` # open 1` LV Size 20.00 GiB Current LE 5120 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 8192 Block device 253:0 --- Logical volume --- LV Path /dev/centos/home LV Name home VG Name centos LV UUID cLdgQy-gRpY-vhiU-73hL-t1xP-2dyO-x1z4Fv LV Write Access read/write LV Creation host, time localhost.localdomain, 2017-02-10 22:41:59 +0000 LV Status available `# open 1` LV Size 10.00 GiB Current LE 2560 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 8192 Block device 253:2 --- Logical volume --- LV Path /dev/centos/swap LV Name swap VG Name centos LV UUID PqR0HU-TbW3-Ew4C-gET1-f7aP-OVRB-5xhx5H LV Write Access read/write LV Creation host, time localhost.localdomain, 2017-02-10 22:42:00 +0000 LV Status available `# open 2` LV Size 4.00 GiB Current LE 1024 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 8192 Block device 253:1 --- Logical volume --- LV Name docker-pool VG Name centos LV UUID TtSvhh-Yb8c-8yGo-dnHk-tWwm-4ujD-MejUJX LV Write Access read/write LV Creation host, time localhost.localdomain, 2017-02-10 23:02:07 +0000 LV Pool metadata docker-pool_tmeta LV Pool data docker-pool_tdata LV Status available `# open 0` LV Size 26.00 GiB Allocated pool data 0.22% Allocated metadata 0.07% Current LE 6656 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 8192 Block device 253:5
non_process
possible docker problem with centos upgrade description after running yum upgrade to centos the centos distribution docker is unable to complete a docker build due to an oci runtime error apparently due to a missing var lib docker devicemapper mnt rootfs directory reverting back to a vm snapshot from two weeks ago when docker was working normally after the snapshot was made and with or without removing all containers images volumes repeating the yum upgrade yields the same result upon trying docker build steps to reproduce the issue have a dockerfile containing from alpine copy run docker build build cmd have a docker build build file containing bin sh set ex apk update apk add bash apk add coreutils apk add go apk add git apk add make apk add xz attempt to build the image with docker build t golang build latest describe the results you received docker pulled the base alpine normally says it copied the build script into the image but threw the oci runtime error while executing the build script docker build t golang build latest sending build context to docker daemon kb step from alpine trying to pull repository docker io library alpine latest pulling from docker io library alpine pull complete digest step copy removing intermediate container step run docker build build running in container linux go starting container process caused process linux go container init caused rootfs linux go mounting var lib docker containers secrets to rootfs var lib docker devicemapper mnt rootfs at run secrets caused lstat var lib docker devicemapper mnt rootfs run secrets not a directory invalid header field value oci runtime error container linux go starting container process caused process linux go container init caused rootfs linux go mounting var lib docker containers secrets to rootfs var lib docker devicemapper mnt rootfs at run secrets caused lstat var lib docker devicemapper mnt rootfs run secrets not a directory n the var lib docker devicemapper mnt directory does not contain a rootfs directory docker ps a container id image command created status ports names bin sh c docker b about a minute ago created elegant bohr docker images a repository tag image id created size minutes ago mb docker io alpine latest hours ago mb describe the results you expected previously docker built the image and docker images a would show docker io alpine and the build image with the specified name and tag would appear and there would be nothing in docker ps a additional information you deem important e g issue happens only occasionally since the yum upgrade to centos i have been unable to do any docker builds due to the problem above the name of the container appearing due to the docker build changes on every attempt the os user is in the docker group to perform all docker operations as a non privileged user the closest existing issue i could find is but that appears to be a sys fs cgroup problem which is not being reported here output of docker version client version api version package version docker common centos go version git commit built mon jan os arch linux server version api version package version docker common centos go version git commit built mon jan os arch linux output of docker info containers running paused stopped images server version storage driver devicemapper pool name centos docker pool pool blocksize kb base device size gb backing filesystem xfs data file metadata file data space used mb data space total gb data space available gb metadata space used kb metadata space total mb metadata space available mb thin pool minimum free space gb udev sync supported true deferred removal enabled true deferred deletion enabled false deferred deleted device count library version logging driver journald cgroup driver systemd plugins volume local network null bridge host overlay swarm inactive runtimes docker runc runc default runtime docker runc security options seccomp selinux kernel version operating system centos linux core ostype linux architecture number of docker hooks cpus total memory gib name localhost localdomain id itfx btfw tknh lxgn sgpe coab phmy docker root dir var lib docker debug mode client false debug mode server false registry insecure registries registries docker io secure additional environment details aws virtualbox physical etc virtualbox running on windows pro vm is centos rpm query centos release centos release centos virtualbox guest additions are installed and appear to be working normally disks are not full df filesystem blocks used available use mounted on dev mapper centos root devtmpfs dev tmpfs dev shm tmpfs run tmpfs sys fs cgroup dev boot dev mapper centos home home tmpfs run user vmshare media sf vmshare tmpfs run user sudo lvdisplay logical volume lv path dev centos root lv name root vg name centos lv uuid rnnp msje kbiz lv write access read write lv creation host time localhost localdomain lv status available open lv size gib current le segments allocation inherit read ahead sectors auto currently set to block device logical volume lv path dev centos home lv name home vg name centos lv uuid cldgqy grpy vhiu lv write access read write lv creation host time localhost localdomain lv status available open lv size gib current le segments allocation inherit read ahead sectors auto currently set to block device logical volume lv path dev centos swap lv name swap vg name centos lv uuid ovrb lv write access read write lv creation host time localhost localdomain lv status available open lv size gib current le segments allocation inherit read ahead sectors auto currently set to block device logical volume lv name docker pool vg name centos lv uuid ttsvhh dnhk twwm mejujx lv write access read write lv creation host time localhost localdomain lv pool metadata docker pool tmeta lv pool data docker pool tdata lv status available open lv size gib allocated pool data allocated metadata current le segments allocation inherit read ahead sectors auto currently set to block device
0
1,533
4,119,268,774
IssuesEvent
2016-06-08 14:25:27
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
opened
NTR: epidermal growth factor receptor signaling pathway involved in heart process
BHF-UCL miRNA New term request RNA processes signaling
Dear Editors, I'd like to request a new term re: PMID:23069713 In this study inhibition of epidermal growth factor receptor (EGFR) was shown to promote cardiogenic differentiation of human Mesenchymal Stem Cells (hMSCs) and the transplantation of hMSCs, in which EGFR was inhibited, resulted in enhancement of heart functions. hMSCs transfected with microRNA-133a (miR-133a), which targets EGFR, expressed cardiac-specific markers and luciferase assays demonstrated that EGFR is silenced by miR-133a. The new requested term: 'epidermal growth factor receptor signaling pathway involved in heart process'. I am subsequently planning to request regulation terms to create the following annotation: hsa-miR-133a-3p GO term: negative regulation of epidermal growth factor receptor signaling pathway involved in heart process AE: *regulates(cardiocyte differentiation), occurs_in(mesenchymal stem cell) Thank you, Barbara GOC:BHF, GOC:BHF_miRNA and GOC:bc @rachhuntley @RLovering *When available I intend to change the ‘regulates’ relation to ‘positively regulates’ (The relation has been requested by @rachhuntley).
1.0
NTR: epidermal growth factor receptor signaling pathway involved in heart process - Dear Editors, I'd like to request a new term re: PMID:23069713 In this study inhibition of epidermal growth factor receptor (EGFR) was shown to promote cardiogenic differentiation of human Mesenchymal Stem Cells (hMSCs) and the transplantation of hMSCs, in which EGFR was inhibited, resulted in enhancement of heart functions. hMSCs transfected with microRNA-133a (miR-133a), which targets EGFR, expressed cardiac-specific markers and luciferase assays demonstrated that EGFR is silenced by miR-133a. The new requested term: 'epidermal growth factor receptor signaling pathway involved in heart process'. I am subsequently planning to request regulation terms to create the following annotation: hsa-miR-133a-3p GO term: negative regulation of epidermal growth factor receptor signaling pathway involved in heart process AE: *regulates(cardiocyte differentiation), occurs_in(mesenchymal stem cell) Thank you, Barbara GOC:BHF, GOC:BHF_miRNA and GOC:bc @rachhuntley @RLovering *When available I intend to change the ‘regulates’ relation to ‘positively regulates’ (The relation has been requested by @rachhuntley).
process
ntr epidermal growth factor receptor signaling pathway involved in heart process dear editors i d like to request a new term re pmid in this study inhibition of epidermal growth factor receptor egfr was shown to promote cardiogenic differentiation of human mesenchymal stem cells hmscs and the transplantation of hmscs in which egfr was inhibited resulted in enhancement of heart functions hmscs transfected with microrna mir which targets egfr expressed cardiac specific markers and luciferase assays demonstrated that egfr is silenced by mir the new requested term epidermal growth factor receptor signaling pathway involved in heart process i am subsequently planning to request regulation terms to create the following annotation hsa mir go term negative regulation of epidermal growth factor receptor signaling pathway involved in heart process ae regulates cardiocyte differentiation occurs in mesenchymal stem cell thank you barbara goc bhf goc bhf mirna and goc bc rachhuntley rlovering when available i intend to change the ‘regulates’ relation to ‘positively regulates’ the relation has been requested by rachhuntley
1
473,413
13,641,998,933
IssuesEvent
2020-09-25 14:55:16
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
create cli flags for install_gems
Category: Packages Community Priority: Low Status: Accepted Type: Backlog
--- Author Name: **Rogier Mars** (Rogier Mars) Original Redmine Issue: 4980, https://dev.opennebula.org/issues/4980 Original Date: 2017-01-12 --- Hi, Would it be possible to create flags to set the OS and make the script non-interactive? This would make it easier to run the script from configmanagement like ansible. Now we have to wrap it in expect, and that causes other issues. For example: ``` install_gems --os=centos --silent --force install_gems --os=ubuntu --silent --force ```
1.0
create cli flags for install_gems - --- Author Name: **Rogier Mars** (Rogier Mars) Original Redmine Issue: 4980, https://dev.opennebula.org/issues/4980 Original Date: 2017-01-12 --- Hi, Would it be possible to create flags to set the OS and make the script non-interactive? This would make it easier to run the script from configmanagement like ansible. Now we have to wrap it in expect, and that causes other issues. For example: ``` install_gems --os=centos --silent --force install_gems --os=ubuntu --silent --force ```
non_process
create cli flags for install gems author name rogier mars rogier mars original redmine issue original date hi would it be possible to create flags to set the os and make the script non interactive this would make it easier to run the script from configmanagement like ansible now we have to wrap it in expect and that causes other issues for example install gems os centos silent force install gems os ubuntu silent force
0
13,815
16,577,454,616
IssuesEvent
2021-05-31 07:19:30
paul-buerkner/brms
https://api.github.com/repos/paul-buerkner/brms
closed
loo moment_match crashes R if save_all_pars not specified
bug post-processing
As mentioned in #1126 loo with moment matching doesn't work without `save_all_pars=save_pars(all = TRUE)`. But it seems worse than just not having an appropriate warning message, as it can crash R if tried without the parameter. Example: With `save_all_pars` ```{r} library(brms) m <- brm(yield ~ N*P*K, npk, save_all_pars = save_pars(all = TRUE)) loo(m, moment_match = TRUE) ``` Yields: ```{r} Computed from 4000 by 24 log-likelihood matrix Estimate SE elpd_loo -81.2 3.1 p_loo 7.5 1.3 looic 162.4 6.3 ------ Monte Carlo SE of elpd_loo is 0.1. Pareto k diagnostic values: Count Pct. Min. n_eff (-Inf, 0.5] (good) 20 83.3% 377 (0.5, 0.7] (ok) 4 16.7% 538 (0.7, 1] (bad) 0 0.0% <NA> (1, Inf) (very bad) 0 0.0% <NA> All Pareto k estimates are ok (k < 0.7). See help('pareto-k-diagnostic') for details. ``` Whereas, ```{r} library(brms) m <- brm(yield ~ N*P*K, npk) loo(m, moment_match = TRUE) ``` Yields: ``` *** caught segfault *** address (nil), cause 'memory not mapped' Traceback: 1: .External(list(name = "CppMethod__invoke_notvoid", address = <pointer: 0x564515d9c940>, dll = list(name = "Rcpp", path = "/lib/R/library/Rcpp/libs/Rcpp.so", dynamicLookup = TRUE, handle = <pointer: 0x564515fcbf70>, info = <pointer: 0x564513e69b30>), numParameters = -1L), <\ pointer: 0x564524ed9590>, <pointer: 0x56451dc4f180>, .pointer, ...) 2: object@.MISC$stan_fit_instance$unconstrain_pars(pars) 3: .local(object, ...) 4: rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton)) 5: rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton)) 6: FUN(newX[, i], ...) 7: apply(pars, 1, FUN = function(theta) { rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton))}) 8: unconstrain_pars_stanfit(x$fit, pars = pars, ...) 9: unconstrain_pars(x, pars = pars, ...) 10: loo::loo_moment_match.default(x, loo = loo, post_draws = as.matrix, log_lik_i = .log_lik_i, unconstrain_pars = .unconstrain_pars, log_prob_upars = .log_prob_upars, log_lik_i_upars = .log_lik_i_upars, k_threshold = k_threshold, newdata = newdata, resp = resp, ...) 11: doTryCatch(return(expr), name, parentenv, handler) 12: tryCatchOne(expr, names, parentenv, handlers[[1L]]) 13: tryCatchList(expr, classes, parentenv, handlers) 14: tryCatch(expr, error = function(e) { call <- conditionCall(e) if (!is.null(call)) { if (identical(call[[1L]], quote(doTryCatch))) call <- sys.call(-4L) dcall <- deparse(call)[1L] prefix <- paste("Error in", dcall, ": ") LONG <- 75L sm <- strsplit(conditionMessage(e), "\n")[[1L]] \ w <- 14L + nchar(dcall, type = "w") + nchar(sm[1L], type = "w") if (is.na(w)) w <- 14L + nchar(dcall, type = "b") + nchar(sm[1L], type = "b") if (w > LONG) prefix <- paste0(prefix, "\n ") } else prefix <- "Error : " msg <- paste0(prefix, conditionMessage(e), "\n") \ .Internal(seterrmessage(msg[1L])) if (!silent && isTRUE(getOption("show.error.messages"))) { cat(msg, file = outFile) .Internal(printDeferredWarnings()) } invisible(structure(msg, class = "try-error", condition = e))}) 15: try(loo::loo_moment_match.default(x, loo = loo, post_draws = as.matrix, log_lik_i = .log_lik_i, unconstrain_pars = .unconstrain_pars, log_prob_upars = .log_prob_upars, log_lik_i_upars = .log_lik_i_upars, k_threshold = k_threshold, newdata = newdata, resp = resp, ...)) 16: loo_moment_match.brmsfit(x = .x1, loo = .x2, newdata = .x3, resp = .x4, k_threshold = .x5, check = .x6) 17: loo_moment_match(x = .x1, loo = .x2, newdata = .x3, resp = .x4, k_threshold = .x5, check = .x6) 18: eval(expr, envir, ...) 19: eval(expr, envir, ...) 20: eval2(call, envir = args, enclos = envir) 21: do_call("loo_moment_match", moment_match_args) 22: .loo(x = .x1, newdata = .x2, resp = .x3, model_name = .x4, pointwise = .x5, k_threshold = .x6, save_psis = .x7, moment_match = .x8, reloo = .x9, moment_match_args = .x10, reloo_args = .x11) 23: eval(expr, envir, ...) 24: eval(expr, envir, ...) 25: eval2(call, envir = args, enclos = envir) 26: do_call(paste0(".", criterion), args) 27: .fun(criterion = .x1, pointwise = .x2, resp = .x3, k_threshold = .x4, save_psis = .x5, moment_match = .x6, reloo = .x7, moment_match_args = .x8, reloo_args = .x9, x = .x10, model_name = .x11, use_stored = .x12) 28: eval(expr, envir, ...) 29: eval(expr, envir, ...) 30: eval2(call, envir = args, enclos = envir) 31: do_call(compute_loo, args) 32: .fun(models = .x1, criterion = .x2, pointwise = .x3, compare = .x4, resp = .x5, k_threshold = .x6, save_psis = .x7, moment_match = .x8, reloo = .x9, moment_match_args = .x10, reloo_args = .x11) 33: eval(expr, envir, ...) 34: eval(expr, envir, ...) 35: eval2(call, envir = args, enclos = envir) 36: do_call(compute_loolist, args) 37: loo.brmsfit(m, moment_match = TRUE) 38: loo(m, moment_match = TRUE) An irrecoverable exception occurred. R is aborting now ... ``` Session Info: ```{r} R version 4.0.5 (2021-03-31) Platform: x86_64-pc-linux-gnu (64-bit) Running under: Ubuntu 18.04.5 LTS Matrix products: default BLAS: /lib/R/lib/libRblas.so LAPACK: /lib/R/lib/libRlapack.so locale: [1] LC_CTYPE=en_GB.utf8 LC_NUMERIC=C [3] LC_TIME=en_DK.utf8 LC_COLLATE=en_GB.utf8 [5] LC_MONETARY=en_GB.utf8 LC_MESSAGES=en_GB.utf8 [7] LC_PAPER=fi_FI.utf8 LC_NAME=C [9] LC_ADDRESS=C LC_TELEPHONE=C [11] LC_MEASUREMENT=en_GB.utf8 LC_IDENTIFICATION=C attached base packages: [1] stats graphics grDevices utils datasets methods base other attached packages: [1] brms_2.15.5 Rcpp_1.0.6 loaded via a namespace (and not attached): [1] nlme_3.1-152 matrixStats_0.58.0 xts_0.12.1 [4] threejs_0.3.3 rstan_2.26.1 backports_1.2.1 [7] tools_4.0.5 utf8_1.2.1 R6_2.5.0 [10] DT_0.18 DBI_1.1.1 mgcv_1.8-35 [13] projpred_2.0.2 colorspace_2.0-1 tidyselect_1.1.1 [16] gridExtra_2.3 prettyunits_1.1.1 processx_3.5.2 [19] Brobdingnag_1.2-6 emmeans_1.6.0 curl_4.3.1 [22] compiler_4.0.5 cli_2.5.0 shinyjs_2.0.0 [25] colourpicker_1.1.0 scales_1.1.1 dygraphs_1.1.1.6 [28] mvtnorm_1.1-1 ggridges_0.5.3 callr_3.7.0 [31] stringr_1.4.0 digest_0.6.27 StanHeaders_2.26.1 [34] minqa_1.2.4 base64enc_0.1-3 pkgconfig_2.0.3 [37] htmltools_0.5.1.1 lme4_1.1-27 fastmap_1.1.0 [40] htmlwidgets_1.5.3 rlang_0.4.11 shiny_1.6.0 [43] generics_0.1.0 zoo_1.8-9 jsonlite_1.7.2 [46] crosstalk_1.1.1 gtools_3.8.2 dplyr_1.0.6 [49] inline_0.3.18 magrittr_2.0.1 loo_2.4.1 [52] bayesplot_1.8.0 Matrix_1.3-2 munsell_0.5.0 [55] fansi_0.4.2 abind_1.4-5 lifecycle_1.0.0 [58] stringi_1.6.2 MASS_7.3-53.1 pkgbuild_1.2.0 [61] plyr_1.8.6 grid_4.0.5 parallel_4.0.5 [64] promises_1.2.0.1 crayon_1.4.1 miniUI_0.1.1.1 [67] lattice_0.20-41 splines_4.0.5 ps_1.6.0 [70] pillar_1.6.1 igraph_1.2.6 boot_1.3-27 [73] estimability_1.3 markdown_1.1 shinystan_2.5.0 [76] codetools_0.2-18 reshape2_1.4.4 stats4_4.0.5 [79] rstantools_2.1.1 glue_1.4.2 V8_3.4.2 [82] RcppParallel_5.1.4 vctrs_0.3.8 nloptr_1.2.2.2 [85] httpuv_1.6.1 gtable_0.3.0 purrr_0.3.4 [88] assertthat_0.2.1 ggplot2_3.3.3 mime_0.10 [91] xtable_1.8-4 coda_0.19-4 later_1.2.0 [94] rsconnect_0.8.17 tibble_3.1.2 shinythemes_1.2.0 [97] gamm4_0.2-6 ellipsis_0.3.2 bridgesampling_1.1-2 ```
1.0
loo moment_match crashes R if save_all_pars not specified - As mentioned in #1126 loo with moment matching doesn't work without `save_all_pars=save_pars(all = TRUE)`. But it seems worse than just not having an appropriate warning message, as it can crash R if tried without the parameter. Example: With `save_all_pars` ```{r} library(brms) m <- brm(yield ~ N*P*K, npk, save_all_pars = save_pars(all = TRUE)) loo(m, moment_match = TRUE) ``` Yields: ```{r} Computed from 4000 by 24 log-likelihood matrix Estimate SE elpd_loo -81.2 3.1 p_loo 7.5 1.3 looic 162.4 6.3 ------ Monte Carlo SE of elpd_loo is 0.1. Pareto k diagnostic values: Count Pct. Min. n_eff (-Inf, 0.5] (good) 20 83.3% 377 (0.5, 0.7] (ok) 4 16.7% 538 (0.7, 1] (bad) 0 0.0% <NA> (1, Inf) (very bad) 0 0.0% <NA> All Pareto k estimates are ok (k < 0.7). See help('pareto-k-diagnostic') for details. ``` Whereas, ```{r} library(brms) m <- brm(yield ~ N*P*K, npk) loo(m, moment_match = TRUE) ``` Yields: ``` *** caught segfault *** address (nil), cause 'memory not mapped' Traceback: 1: .External(list(name = "CppMethod__invoke_notvoid", address = <pointer: 0x564515d9c940>, dll = list(name = "Rcpp", path = "/lib/R/library/Rcpp/libs/Rcpp.so", dynamicLookup = TRUE, handle = <pointer: 0x564515fcbf70>, info = <pointer: 0x564513e69b30>), numParameters = -1L), <\ pointer: 0x564524ed9590>, <pointer: 0x56451dc4f180>, .pointer, ...) 2: object@.MISC$stan_fit_instance$unconstrain_pars(pars) 3: .local(object, ...) 4: rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton)) 5: rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton)) 6: FUN(newX[, i], ...) 7: apply(pars, 1, FUN = function(theta) { rstan::unconstrain_pars(x, pars = .rstan_relist(theta, skeleton))}) 8: unconstrain_pars_stanfit(x$fit, pars = pars, ...) 9: unconstrain_pars(x, pars = pars, ...) 10: loo::loo_moment_match.default(x, loo = loo, post_draws = as.matrix, log_lik_i = .log_lik_i, unconstrain_pars = .unconstrain_pars, log_prob_upars = .log_prob_upars, log_lik_i_upars = .log_lik_i_upars, k_threshold = k_threshold, newdata = newdata, resp = resp, ...) 11: doTryCatch(return(expr), name, parentenv, handler) 12: tryCatchOne(expr, names, parentenv, handlers[[1L]]) 13: tryCatchList(expr, classes, parentenv, handlers) 14: tryCatch(expr, error = function(e) { call <- conditionCall(e) if (!is.null(call)) { if (identical(call[[1L]], quote(doTryCatch))) call <- sys.call(-4L) dcall <- deparse(call)[1L] prefix <- paste("Error in", dcall, ": ") LONG <- 75L sm <- strsplit(conditionMessage(e), "\n")[[1L]] \ w <- 14L + nchar(dcall, type = "w") + nchar(sm[1L], type = "w") if (is.na(w)) w <- 14L + nchar(dcall, type = "b") + nchar(sm[1L], type = "b") if (w > LONG) prefix <- paste0(prefix, "\n ") } else prefix <- "Error : " msg <- paste0(prefix, conditionMessage(e), "\n") \ .Internal(seterrmessage(msg[1L])) if (!silent && isTRUE(getOption("show.error.messages"))) { cat(msg, file = outFile) .Internal(printDeferredWarnings()) } invisible(structure(msg, class = "try-error", condition = e))}) 15: try(loo::loo_moment_match.default(x, loo = loo, post_draws = as.matrix, log_lik_i = .log_lik_i, unconstrain_pars = .unconstrain_pars, log_prob_upars = .log_prob_upars, log_lik_i_upars = .log_lik_i_upars, k_threshold = k_threshold, newdata = newdata, resp = resp, ...)) 16: loo_moment_match.brmsfit(x = .x1, loo = .x2, newdata = .x3, resp = .x4, k_threshold = .x5, check = .x6) 17: loo_moment_match(x = .x1, loo = .x2, newdata = .x3, resp = .x4, k_threshold = .x5, check = .x6) 18: eval(expr, envir, ...) 19: eval(expr, envir, ...) 20: eval2(call, envir = args, enclos = envir) 21: do_call("loo_moment_match", moment_match_args) 22: .loo(x = .x1, newdata = .x2, resp = .x3, model_name = .x4, pointwise = .x5, k_threshold = .x6, save_psis = .x7, moment_match = .x8, reloo = .x9, moment_match_args = .x10, reloo_args = .x11) 23: eval(expr, envir, ...) 24: eval(expr, envir, ...) 25: eval2(call, envir = args, enclos = envir) 26: do_call(paste0(".", criterion), args) 27: .fun(criterion = .x1, pointwise = .x2, resp = .x3, k_threshold = .x4, save_psis = .x5, moment_match = .x6, reloo = .x7, moment_match_args = .x8, reloo_args = .x9, x = .x10, model_name = .x11, use_stored = .x12) 28: eval(expr, envir, ...) 29: eval(expr, envir, ...) 30: eval2(call, envir = args, enclos = envir) 31: do_call(compute_loo, args) 32: .fun(models = .x1, criterion = .x2, pointwise = .x3, compare = .x4, resp = .x5, k_threshold = .x6, save_psis = .x7, moment_match = .x8, reloo = .x9, moment_match_args = .x10, reloo_args = .x11) 33: eval(expr, envir, ...) 34: eval(expr, envir, ...) 35: eval2(call, envir = args, enclos = envir) 36: do_call(compute_loolist, args) 37: loo.brmsfit(m, moment_match = TRUE) 38: loo(m, moment_match = TRUE) An irrecoverable exception occurred. R is aborting now ... ``` Session Info: ```{r} R version 4.0.5 (2021-03-31) Platform: x86_64-pc-linux-gnu (64-bit) Running under: Ubuntu 18.04.5 LTS Matrix products: default BLAS: /lib/R/lib/libRblas.so LAPACK: /lib/R/lib/libRlapack.so locale: [1] LC_CTYPE=en_GB.utf8 LC_NUMERIC=C [3] LC_TIME=en_DK.utf8 LC_COLLATE=en_GB.utf8 [5] LC_MONETARY=en_GB.utf8 LC_MESSAGES=en_GB.utf8 [7] LC_PAPER=fi_FI.utf8 LC_NAME=C [9] LC_ADDRESS=C LC_TELEPHONE=C [11] LC_MEASUREMENT=en_GB.utf8 LC_IDENTIFICATION=C attached base packages: [1] stats graphics grDevices utils datasets methods base other attached packages: [1] brms_2.15.5 Rcpp_1.0.6 loaded via a namespace (and not attached): [1] nlme_3.1-152 matrixStats_0.58.0 xts_0.12.1 [4] threejs_0.3.3 rstan_2.26.1 backports_1.2.1 [7] tools_4.0.5 utf8_1.2.1 R6_2.5.0 [10] DT_0.18 DBI_1.1.1 mgcv_1.8-35 [13] projpred_2.0.2 colorspace_2.0-1 tidyselect_1.1.1 [16] gridExtra_2.3 prettyunits_1.1.1 processx_3.5.2 [19] Brobdingnag_1.2-6 emmeans_1.6.0 curl_4.3.1 [22] compiler_4.0.5 cli_2.5.0 shinyjs_2.0.0 [25] colourpicker_1.1.0 scales_1.1.1 dygraphs_1.1.1.6 [28] mvtnorm_1.1-1 ggridges_0.5.3 callr_3.7.0 [31] stringr_1.4.0 digest_0.6.27 StanHeaders_2.26.1 [34] minqa_1.2.4 base64enc_0.1-3 pkgconfig_2.0.3 [37] htmltools_0.5.1.1 lme4_1.1-27 fastmap_1.1.0 [40] htmlwidgets_1.5.3 rlang_0.4.11 shiny_1.6.0 [43] generics_0.1.0 zoo_1.8-9 jsonlite_1.7.2 [46] crosstalk_1.1.1 gtools_3.8.2 dplyr_1.0.6 [49] inline_0.3.18 magrittr_2.0.1 loo_2.4.1 [52] bayesplot_1.8.0 Matrix_1.3-2 munsell_0.5.0 [55] fansi_0.4.2 abind_1.4-5 lifecycle_1.0.0 [58] stringi_1.6.2 MASS_7.3-53.1 pkgbuild_1.2.0 [61] plyr_1.8.6 grid_4.0.5 parallel_4.0.5 [64] promises_1.2.0.1 crayon_1.4.1 miniUI_0.1.1.1 [67] lattice_0.20-41 splines_4.0.5 ps_1.6.0 [70] pillar_1.6.1 igraph_1.2.6 boot_1.3-27 [73] estimability_1.3 markdown_1.1 shinystan_2.5.0 [76] codetools_0.2-18 reshape2_1.4.4 stats4_4.0.5 [79] rstantools_2.1.1 glue_1.4.2 V8_3.4.2 [82] RcppParallel_5.1.4 vctrs_0.3.8 nloptr_1.2.2.2 [85] httpuv_1.6.1 gtable_0.3.0 purrr_0.3.4 [88] assertthat_0.2.1 ggplot2_3.3.3 mime_0.10 [91] xtable_1.8-4 coda_0.19-4 later_1.2.0 [94] rsconnect_0.8.17 tibble_3.1.2 shinythemes_1.2.0 [97] gamm4_0.2-6 ellipsis_0.3.2 bridgesampling_1.1-2 ```
process
loo moment match crashes r if save all pars not specified as mentioned in loo with moment matching doesn t work without save all pars save pars all true but it seems worse than just not having an appropriate warning message as it can crash r if tried without the parameter example with save all pars r library brms m brm yield n p k npk save all pars save pars all true loo m moment match true yields r computed from by log likelihood matrix estimate se elpd loo p loo looic monte carlo se of elpd loo is pareto k diagnostic values count pct min n eff inf good ok bad inf very bad all pareto k estimates are ok k see help pareto k diagnostic for details whereas r library brms m brm yield n p k npk loo m moment match true yields caught segfault address nil cause memory not mapped traceback external list name cppmethod invoke notvoid address dll list name rcpp path lib r library rcpp libs rcpp so dynamiclookup true handle info numparameters pointer pointer object misc stan fit instance unconstrain pars pars local object rstan unconstrain pars x pars rstan relist theta skeleton rstan unconstrain pars x pars rstan relist theta skeleton fun newx apply pars fun function theta rstan unconstrain pars x pars rstan relist theta skeleton unconstrain pars stanfit x fit pars pars unconstrain pars x pars pars loo loo moment match default x loo loo post draws as matrix log lik i log lik i unconstrain pars unconstrain pars log prob upars log prob upars log lik i upars log lik i upars k threshold k threshold newdata newdata resp resp dotrycatch return expr name parentenv handler trycatchone expr names parentenv handlers trycatchlist expr classes parentenv handlers trycatch expr error function e call conditioncall e if is null call if identical call quote dotrycatch call sys call dcall deparse call prefix paste error in dcall long sm strsplit conditionmessage e n w long prefix prefix n else prefix error msg prefix conditionmessage e n internal seterrmessage msg if silent istrue getoption show error messages cat msg file outfile internal printdeferredwarnings invisible structure msg class try error condition e try loo loo moment match default x loo loo post draws as matrix log lik i log lik i unconstrain pars unconstrain pars log prob upars log prob upars log lik i upars log lik i upars k threshold k threshold newdata newdata resp resp loo moment match brmsfit x loo newdata resp k threshold check loo moment match x loo newdata resp k threshold check eval expr envir eval expr envir call envir args enclos envir do call loo moment match moment match args loo x newdata resp model name pointwise k threshold save psis moment match reloo moment match args reloo args eval expr envir eval expr envir call envir args enclos envir do call criterion args fun criterion pointwise resp k threshold save psis moment match reloo moment match args reloo args x model name use stored eval expr envir eval expr envir call envir args enclos envir do call compute loo args fun models criterion pointwise compare resp k threshold save psis moment match reloo moment match args reloo args eval expr envir eval expr envir call envir args enclos envir do call compute loolist args loo brmsfit m moment match true loo m moment match true an irrecoverable exception occurred r is aborting now session info r r version platform pc linux gnu bit running under ubuntu lts matrix products default blas lib r lib librblas so lapack lib r lib librlapack so locale lc ctype en gb lc numeric c lc time en dk lc collate en gb lc monetary en gb lc messages en gb lc paper fi fi lc name c lc address c lc telephone c lc measurement en gb lc identification c attached base packages stats graphics grdevices utils datasets methods base other attached packages brms rcpp loaded via a namespace and not attached nlme matrixstats xts threejs rstan backports tools dt dbi mgcv projpred colorspace tidyselect gridextra prettyunits processx brobdingnag emmeans curl compiler cli shinyjs colourpicker scales dygraphs mvtnorm ggridges callr stringr digest stanheaders minqa pkgconfig htmltools fastmap htmlwidgets rlang shiny generics zoo jsonlite crosstalk gtools dplyr inline magrittr loo bayesplot matrix munsell fansi abind lifecycle stringi mass pkgbuild plyr grid parallel promises crayon miniui lattice splines ps pillar igraph boot estimability markdown shinystan codetools rstantools glue rcppparallel vctrs nloptr httpuv gtable purrr assertthat mime xtable coda later rsconnect tibble shinythemes ellipsis bridgesampling
1
10,967
3,152,375,513
IssuesEvent
2015-09-16 13:35:36
galenframework/galen
https://api.github.com/repos/galenframework/galen
closed
Add COUNT spec for multiple objects
c1 enhancement p2 ready for test
When using multiple object identification, would be great to have a COUNT spec Example: ``` ===================================== menu-items-* css .menuitem ===================================== menu-items-* count: 5 ```
1.0
Add COUNT spec for multiple objects - When using multiple object identification, would be great to have a COUNT spec Example: ``` ===================================== menu-items-* css .menuitem ===================================== menu-items-* count: 5 ```
non_process
add count spec for multiple objects when using multiple object identification would be great to have a count spec example menu items css menuitem menu items count
0
75,229
9,829,284,390
IssuesEvent
2019-06-15 19:17:24
paul-buerkner/brms
https://api.github.com/repos/paul-buerkner/brms
closed
Number of multiple imputation in "missing values" vignette
documentation
In the vignette "Handle Missing Values with brms", section "Imputation before model fitting", there are m = 5 multiply imputed datasets used for mixing their posterior draws. However, according to the paper cited below, m = 5 is not enough for reliable posterior inferences. The authors recommend to choose a larger number of multiply imputed datasets. They don't recommend a specific number meant by "large", but m = 100 is doing well in their simulation study. In contrast, m = 20 is not enough. Xiang Zhou & Jerome P. Reiter (2010) A Note on Bayesian Inference After Multiple Imputation, The American Statistician, 64:2, 159-163, DOI: 10.1198/tast.2010.09109 I understand that m = 100 increases the computational burden heavily, but perhaps a remark should be added to the vignette that m = 5 might not be enough for reliable posterior inferences.
1.0
Number of multiple imputation in "missing values" vignette - In the vignette "Handle Missing Values with brms", section "Imputation before model fitting", there are m = 5 multiply imputed datasets used for mixing their posterior draws. However, according to the paper cited below, m = 5 is not enough for reliable posterior inferences. The authors recommend to choose a larger number of multiply imputed datasets. They don't recommend a specific number meant by "large", but m = 100 is doing well in their simulation study. In contrast, m = 20 is not enough. Xiang Zhou & Jerome P. Reiter (2010) A Note on Bayesian Inference After Multiple Imputation, The American Statistician, 64:2, 159-163, DOI: 10.1198/tast.2010.09109 I understand that m = 100 increases the computational burden heavily, but perhaps a remark should be added to the vignette that m = 5 might not be enough for reliable posterior inferences.
non_process
number of multiple imputation in missing values vignette in the vignette handle missing values with brms section imputation before model fitting there are m multiply imputed datasets used for mixing their posterior draws however according to the paper cited below m is not enough for reliable posterior inferences the authors recommend to choose a larger number of multiply imputed datasets they don t recommend a specific number meant by large but m is doing well in their simulation study in contrast m is not enough xiang zhou jerome p reiter a note on bayesian inference after multiple imputation the american statistician doi tast i understand that m increases the computational burden heavily but perhaps a remark should be added to the vignette that m might not be enough for reliable posterior inferences
0
332,059
10,083,740,195
IssuesEvent
2019-07-25 14:16:54
getkirby/kirby
https://api.github.com/repos/getkirby/kirby
closed
KirbyTag gets escaped twice when using escape()
missing: discussion 🗣 missing: information ❓ priority: minor 🔜
**Describe the bug** When using `escape()` on a KirbyText, KirbyTags get escaped twice. **To Reproduce** Given a field text with a textarea and the following content: ``` foo & bar (link: http://example.com text: foo & bar) ``` In my template/snippet I am using the following line: `<?= $page->text()->escape()->kirbytext() ?>` The output is the following: ``` <p>foo &amp; bar</p> <p><a href="http://example.com">foo &amp;amp; bar</a></p> ``` **Expected behavior** The KirbyTag Link shouldn't be escaped twice. **Kirby Version** 3.1.4
1.0
KirbyTag gets escaped twice when using escape() - **Describe the bug** When using `escape()` on a KirbyText, KirbyTags get escaped twice. **To Reproduce** Given a field text with a textarea and the following content: ``` foo & bar (link: http://example.com text: foo & bar) ``` In my template/snippet I am using the following line: `<?= $page->text()->escape()->kirbytext() ?>` The output is the following: ``` <p>foo &amp; bar</p> <p><a href="http://example.com">foo &amp;amp; bar</a></p> ``` **Expected behavior** The KirbyTag Link shouldn't be escaped twice. **Kirby Version** 3.1.4
non_process
kirbytag gets escaped twice when using escape describe the bug when using escape on a kirbytext kirbytags get escaped twice to reproduce given a field text with a textarea and the following content foo bar link text foo bar in my template snippet i am using the following line text escape kirbytext the output is the following foo amp bar expected behavior the kirbytag link shouldn t be escaped twice kirby version
0
10,423
13,215,849,266
IssuesEvent
2020-08-17 01:25:55
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
opened
Add ability to designate dependent/source data to be used in displays, computations
f - processing f - user-interface feature stage - planning type - enhancement
For example, if background subtraction is applied to a line plot, the user could process the original data (re-binning or smoothing, for example) and then designate the resulting data to be used anywhere the original data is used (computations, displays, maybe more). Other ideas are: if the designated replacement is deleted, it reverts to the original source; some specialized operations may have the option to auto-designate without explicitly doing so (UI needed).
1.0
Add ability to designate dependent/source data to be used in displays, computations - For example, if background subtraction is applied to a line plot, the user could process the original data (re-binning or smoothing, for example) and then designate the resulting data to be used anywhere the original data is used (computations, displays, maybe more). Other ideas are: if the designated replacement is deleted, it reverts to the original source; some specialized operations may have the option to auto-designate without explicitly doing so (UI needed).
process
add ability to designate dependent source data to be used in displays computations for example if background subtraction is applied to a line plot the user could process the original data re binning or smoothing for example and then designate the resulting data to be used anywhere the original data is used computations displays maybe more other ideas are if the designated replacement is deleted it reverts to the original source some specialized operations may have the option to auto designate without explicitly doing so ui needed
1
7,120
10,266,291,252
IssuesEvent
2019-08-22 21:02:10
automotive-edge-computing-consortium/AECC
https://api.github.com/repos/automotive-edge-computing-consortium/AECC
opened
Good place to keep the Issues list
priority:High status:Open type:Process
Looking for good place to store this issue list. Perhaps this is part of ticketing system. Currently this issue list doesn't include sensitive item. But it will have sensitive information such as launching new SIGs. So better to be stored in work space where limited people (chairs or sponsor members) have access right. This is the place...
1.0
Good place to keep the Issues list - Looking for good place to store this issue list. Perhaps this is part of ticketing system. Currently this issue list doesn't include sensitive item. But it will have sensitive information such as launching new SIGs. So better to be stored in work space where limited people (chairs or sponsor members) have access right. This is the place...
process
good place to keep the issues list looking for good place to store this issue list perhaps this is part of ticketing system currently this issue list doesn t include sensitive item but it will have sensitive information such as launching new sigs so better to be stored in work space where limited people chairs or sponsor members have access right this is the place
1
189,861
14,525,777,071
IssuesEvent
2020-12-14 13:24:10
fourMs/MGT-python
https://api.github.com/repos/fourMs/MGT-python
closed
Make sure the ffmpeg commands always use the -y flag
bug testing
Because if not and the destination file happens to exist already the process will just quit without doing anything.
1.0
Make sure the ffmpeg commands always use the -y flag - Because if not and the destination file happens to exist already the process will just quit without doing anything.
non_process
make sure the ffmpeg commands always use the y flag because if not and the destination file happens to exist already the process will just quit without doing anything
0
20,843
27,612,216,142
IssuesEvent
2023-03-09 16:43:24
influxdata/telegraf
https://api.github.com/repos/influxdata/telegraf
closed
processors.converter - convert time
help wanted feature request plugin/processor size/m
### Use Case The idea is to improve `processors.converter` and allow it to manage timestamps. Currently, there is no "easy" way to override/set the timestamp of a point by getting it from a tag/field, the only option is to use `processors.starlark`. This is a rare use case, but still, I think it's worth having it ### Expected behavior Being able to override/set the time of a point base on a field/tag value using `processors.converter`. As done for "generic" input plugins I just expect to specify the source tag/field and time format ### Actual behavior The only way to override/set the time of a point base on a field/tag value is to use `processors.starlark` ### Additional info _No response_
1.0
processors.converter - convert time - ### Use Case The idea is to improve `processors.converter` and allow it to manage timestamps. Currently, there is no "easy" way to override/set the timestamp of a point by getting it from a tag/field, the only option is to use `processors.starlark`. This is a rare use case, but still, I think it's worth having it ### Expected behavior Being able to override/set the time of a point base on a field/tag value using `processors.converter`. As done for "generic" input plugins I just expect to specify the source tag/field and time format ### Actual behavior The only way to override/set the time of a point base on a field/tag value is to use `processors.starlark` ### Additional info _No response_
process
processors converter convert time use case the idea is to improve processors converter and allow it to manage timestamps currently there is no easy way to override set the timestamp of a point by getting it from a tag field the only option is to use processors starlark this is a rare use case but still i think it s worth having it expected behavior being able to override set the time of a point base on a field tag value using processors converter as done for generic input plugins i just expect to specify the source tag field and time format actual behavior the only way to override set the time of a point base on a field tag value is to use processors starlark additional info no response
1
521,769
15,115,337,783
IssuesEvent
2021-02-09 04:12:47
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
opened
Add cumulative breach calculation logic to BreachManager
Docs: not needed Feature Module: vaccines Priority: high
## Is your feature request related to a problem? Please describe. Add cumulative breach calculation logic to mobile app ## Describe the solution you'd like - Add methods to BreachManager to calculate cumulative duration above/below configured thresholds - When consecutive breach calculation is performed, call above method to do cumulative breach calculation ## Implementation N/A ## Describe alternatives you've considered Not sure, CCE app has this calculation done in SQL but think we would want it to be DB agnostic? ## Additional context N/A
1.0
Add cumulative breach calculation logic to BreachManager - ## Is your feature request related to a problem? Please describe. Add cumulative breach calculation logic to mobile app ## Describe the solution you'd like - Add methods to BreachManager to calculate cumulative duration above/below configured thresholds - When consecutive breach calculation is performed, call above method to do cumulative breach calculation ## Implementation N/A ## Describe alternatives you've considered Not sure, CCE app has this calculation done in SQL but think we would want it to be DB agnostic? ## Additional context N/A
non_process
add cumulative breach calculation logic to breachmanager is your feature request related to a problem please describe add cumulative breach calculation logic to mobile app describe the solution you d like add methods to breachmanager to calculate cumulative duration above below configured thresholds when consecutive breach calculation is performed call above method to do cumulative breach calculation implementation n a describe alternatives you ve considered not sure cce app has this calculation done in sql but think we would want it to be db agnostic additional context n a
0
5,712
8,567,916,927
IssuesEvent
2018-11-10 16:33:43
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
Wrong phone number on website / Web Form does not deliver mail
status-inprocess type-bug website-general
<img width="467" alt="screen shot 2018-07-06 at 4 39 24 pm" src="https://user-images.githubusercontent.com/5417918/43462686-c5509e8c-94a4-11e8-9d51-71aafb574283.png">
1.0
Wrong phone number on website / Web Form does not deliver mail - <img width="467" alt="screen shot 2018-07-06 at 4 39 24 pm" src="https://user-images.githubusercontent.com/5417918/43462686-c5509e8c-94a4-11e8-9d51-71aafb574283.png">
process
wrong phone number on website web form does not deliver mail img width alt screen shot at pm src
1
33,204
4,818,098,226
IssuesEvent
2016-11-04 15:29:37
infiniteautomation/ma-core-public
https://api.github.com/repos/infiniteautomation/ma-core-public
closed
Log4j data source - ALL level doesn't work
Bug Ready for Testing
The ALL level fails to match any messages whether using Regex or not. This has been fixed, with validation added to saving a point with regex, a bug fix on points that were edited a few times, and some i18n issues.
1.0
Log4j data source - ALL level doesn't work - The ALL level fails to match any messages whether using Regex or not. This has been fixed, with validation added to saving a point with regex, a bug fix on points that were edited a few times, and some i18n issues.
non_process
data source all level doesn t work the all level fails to match any messages whether using regex or not this has been fixed with validation added to saving a point with regex a bug fix on points that were edited a few times and some issues
0
38,494
8,850,936,038
IssuesEvent
2019-01-08 14:34:56
STEllAR-GROUP/phylanx
https://api.github.com/repos/STEllAR-GROUP/phylanx
closed
PhySL parser silently converts the type to node_data<double>
category: PhySL submodule: backend type: defect
Using square brackets to indicate a vector of integers in PhySL produces a `node_data<double>` instead of `node_data<std::int64_t>`. ``` cout(repeat(2,hstack(3))) cout(repeat(2,[3])) ``` ```console [2, 2, 2] physl: exception caught: <unknown>: primitive_argument_type does not hold an integer value type (type held: 'phylanx::ir::node_data<double>'): HPX(bad_parameter) ``` ```scheme define(a,[3]) cout(repeat(2,a)) ``` ```console physl: exception caught: <unknown>: primitive_argument_type does not hold an integer value type (type held: 'phylanx::ir::node_data<double>'): HPX(bad_parameter) ```
1.0
PhySL parser silently converts the type to node_data<double> - Using square brackets to indicate a vector of integers in PhySL produces a `node_data<double>` instead of `node_data<std::int64_t>`. ``` cout(repeat(2,hstack(3))) cout(repeat(2,[3])) ``` ```console [2, 2, 2] physl: exception caught: <unknown>: primitive_argument_type does not hold an integer value type (type held: 'phylanx::ir::node_data<double>'): HPX(bad_parameter) ``` ```scheme define(a,[3]) cout(repeat(2,a)) ``` ```console physl: exception caught: <unknown>: primitive_argument_type does not hold an integer value type (type held: 'phylanx::ir::node_data<double>'): HPX(bad_parameter) ```
non_process
physl parser silently converts the type to node data using square brackets to indicate a vector of integers in physl produces a node data instead of node data cout repeat hstack cout repeat console physl exception caught primitive argument type does not hold an integer value type type held phylanx ir node data hpx bad parameter scheme define a cout repeat a console physl exception caught primitive argument type does not hold an integer value type type held phylanx ir node data hpx bad parameter
0
52,661
13,224,887,074
IssuesEvent
2020-08-17 20:02:56
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
h5booker should refuse to book if some framenames need sanitizing (Trac #118)
Migrated from Trac booking defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/118">https://code.icecube.wisc.edu/projects/icecube/ticket/118</a>, reported by troyand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2011-04-14T19:16:34", "_ts": "1302808594000000", "description": "", "reporter": "troy", "cc": "", "resolution": "wont or cant fix", "time": "2008-09-03T00:00:00", "component": "booking", "summary": "h5booker should refuse to book if some framenames need sanitizing", "priority": "major", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
h5booker should refuse to book if some framenames need sanitizing (Trac #118) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/118">https://code.icecube.wisc.edu/projects/icecube/ticket/118</a>, reported by troyand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2011-04-14T19:16:34", "_ts": "1302808594000000", "description": "", "reporter": "troy", "cc": "", "resolution": "wont or cant fix", "time": "2008-09-03T00:00:00", "component": "booking", "summary": "h5booker should refuse to book if some framenames need sanitizing", "priority": "major", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
non_process
should refuse to book if some framenames need sanitizing trac migrated from json status closed changetime ts description reporter troy cc resolution wont or cant fix time component booking summary should refuse to book if some framenames need sanitizing priority major keywords milestone owner troy type defect
0
20,487
27,144,858,344
IssuesEvent
2023-02-16 19:06:34
mehta-lab/microDL
https://api.github.com/repos/mehta-lab/microDL
closed
Housecleaning for unused code in gunpowder implementation
preprocessing
The following sections of code will be deleted in gunpowder implementation - Tensorflow networks module - Inference dataset - tensorflow dataset - resizing
1.0
Housecleaning for unused code in gunpowder implementation - The following sections of code will be deleted in gunpowder implementation - Tensorflow networks module - Inference dataset - tensorflow dataset - resizing
process
housecleaning for unused code in gunpowder implementation the following sections of code will be deleted in gunpowder implementation tensorflow networks module inference dataset tensorflow dataset resizing
1
17,143
22,690,449,054
IssuesEvent
2022-07-04 19:22:44
threefoldtech/grid_weblets
https://api.github.com/repos/threefoldtech/grid_weblets
closed
UX regarding the node selection
process_wontfix
It is very confusing for the user that when selecting the node to deploy anything, you have too many options in Weblets: ![image](https://user-images.githubusercontent.com/65153287/167369771-cdc5c4e0-9a4e-4191-919d-a9d5cb353efc.png) You can have a "Manual Selection" which currently doesn't work, You can select the farm and node, But actually you need to click the "Apply Filters and Suggested Nodes" button in order to search for available nodes from which you need to select to provision a resource
1.0
UX regarding the node selection - It is very confusing for the user that when selecting the node to deploy anything, you have too many options in Weblets: ![image](https://user-images.githubusercontent.com/65153287/167369771-cdc5c4e0-9a4e-4191-919d-a9d5cb353efc.png) You can have a "Manual Selection" which currently doesn't work, You can select the farm and node, But actually you need to click the "Apply Filters and Suggested Nodes" button in order to search for available nodes from which you need to select to provision a resource
process
ux regarding the node selection it is very confusing for the user that when selecting the node to deploy anything you have too many options in weblets you can have a manual selection which currently doesn t work you can select the farm and node but actually you need to click the apply filters and suggested nodes button in order to search for available nodes from which you need to select to provision a resource
1
870
3,329,827,475
IssuesEvent
2015-11-11 05:45:35
uavis/mlproj
https://api.github.com/repos/uavis/mlproj
closed
Automate Skull removal
MS Lesion Preprocessing
Input: .nrrd (labels is in .nrrd as well) .nrrd(.nhdr, .raw) -> .nii.gz -> .nrrd -> Matlab Tool: bet2 to remove the skull. Issue: bet2 only takes .nii.gz Right now, we’re converting the format with ITKSNAP: ``` $ /usr/work/vincent/itksnap-3.2.0-20141023-Linux-x86_64/bin/itksnap # The nrrd produced by itk has irrelevant chars. To remove it: $ sed -i '13,61d' UNC_train_Case01_T2_mask.nrrd % in place removal of line 13-61 ``` Strip the skull with BET2: ``` /usr/work/share/fsl/bin ``` may need to set up the environmental variable before starting it: ``` $ export FSLDIR=/usr/work/share/fsl $ . ${FSLDIR}/etc/fslconf/fsl.sh ``` ``` Example: vincent@richdale:/usr/data/medical_images/MSlesion08/tmp$ /usr/work/share/fsl/bin/bet2 UNC_train_Case01_T2_original.nii.gz UNC_train_Case01_T2_output -m Arguments: “UNC_train_Case01_T2_original.nii.gz” is the input scan, “UNC_train_Case01_T2_output” is the output file without the extension, -m : option to generate a mask (1 for brain tissue, 0 otherwise) Output: UNC_train_Case01_T2_output_mask.nii.gz (the mask generated by -m) UNC_train_Case01_T2_output.nii.gz ``` Open task: - [x] Find a converter (command line or matlab package) - [x] write script to automate the process
1.0
Automate Skull removal - Input: .nrrd (labels is in .nrrd as well) .nrrd(.nhdr, .raw) -> .nii.gz -> .nrrd -> Matlab Tool: bet2 to remove the skull. Issue: bet2 only takes .nii.gz Right now, we’re converting the format with ITKSNAP: ``` $ /usr/work/vincent/itksnap-3.2.0-20141023-Linux-x86_64/bin/itksnap # The nrrd produced by itk has irrelevant chars. To remove it: $ sed -i '13,61d' UNC_train_Case01_T2_mask.nrrd % in place removal of line 13-61 ``` Strip the skull with BET2: ``` /usr/work/share/fsl/bin ``` may need to set up the environmental variable before starting it: ``` $ export FSLDIR=/usr/work/share/fsl $ . ${FSLDIR}/etc/fslconf/fsl.sh ``` ``` Example: vincent@richdale:/usr/data/medical_images/MSlesion08/tmp$ /usr/work/share/fsl/bin/bet2 UNC_train_Case01_T2_original.nii.gz UNC_train_Case01_T2_output -m Arguments: “UNC_train_Case01_T2_original.nii.gz” is the input scan, “UNC_train_Case01_T2_output” is the output file without the extension, -m : option to generate a mask (1 for brain tissue, 0 otherwise) Output: UNC_train_Case01_T2_output_mask.nii.gz (the mask generated by -m) UNC_train_Case01_T2_output.nii.gz ``` Open task: - [x] Find a converter (command line or matlab package) - [x] write script to automate the process
process
automate skull removal input nrrd labels is in nrrd as well nrrd nhdr raw nii gz nrrd matlab tool to remove the skull issue only takes nii gz right now we’re converting the format with itksnap usr work vincent itksnap linux bin itksnap the nrrd produced by itk has irrelevant chars to remove it sed i unc train mask nrrd in place removal of line strip the skull with usr work share fsl bin may need to set up the environmental variable before starting it export fsldir usr work share fsl fsldir etc fslconf fsl sh example vincent richdale usr data medical images tmp usr work share fsl bin unc train original nii gz unc train output m arguments “unc train original nii gz” is the input scan “unc train output” is the output file without the extension m option to generate a mask for brain tissue otherwise output unc train output mask nii gz the mask generated by m unc train output nii gz open task find a converter command line or matlab package write script to automate the process
1
911
3,373,517,198
IssuesEvent
2015-11-24 08:13:49
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
reopened
Not all types of 'eval' function calls wrapped
AREA: client SYSTEM: resource processing TYPE: bug
We have to wrap: ```javascript var temp = eval; temp('some code'); ``` to ```javascript var temp = eval === window.eval ? script => eval(__proc$Script(script)) : eval; temp('some code'); ```
1.0
Not all types of 'eval' function calls wrapped - We have to wrap: ```javascript var temp = eval; temp('some code'); ``` to ```javascript var temp = eval === window.eval ? script => eval(__proc$Script(script)) : eval; temp('some code'); ```
process
not all types of eval function calls wrapped we have to wrap javascript var temp eval temp some code to javascript var temp eval window eval script eval proc script script eval temp some code
1
17,687
23,532,206,709
IssuesEvent
2022-08-19 16:26:55
googleapis/google-cloud-node
https://api.github.com/repos/googleapis/google-cloud-node
closed
Warning: a recent release failed
type: process
The following release PRs may have failed: * #3282 - The release job failed -- check the build log. * #3270 - The release job failed -- check the build log. * #3281 - The release job failed -- check the build log. * #3280 - The release job failed -- check the build log.
1.0
Warning: a recent release failed - The following release PRs may have failed: * #3282 - The release job failed -- check the build log. * #3270 - The release job failed -- check the build log. * #3281 - The release job failed -- check the build log. * #3280 - The release job failed -- check the build log.
process
warning a recent release failed the following release prs may have failed the release job failed check the build log the release job failed check the build log the release job failed check the build log the release job failed check the build log
1
46,919
19,553,175,971
IssuesEvent
2022-01-03 03:12:03
PreMiD/Presences
https://api.github.com/repos/PreMiD/Presences
opened
Sushiroll | sushiroll.co.id
💳 Paid Service 🌍 Region Restricted
### Discussed in https://github.com/PreMiD/Presences/discussions/4591 <div type='discussions-op-text'> <sup>Originally posted by **rayyplay** April 12, 2021</sup> **Prerequisites and essential questions** <!--- Required, please answer the following questions as honestly as possible by changing the "[ ]" to "[x]" or by marking it after creating the issue (easier), not marking a question counts as "No". --> - [x] Is it a popular site? - [x] Is the website older than 2 months? <!--- It is necessary for the website to be older than 2 months. --> - [x] Is the site locked to a specific country/region? - [x] Is the site a paid service? (e.g. Netflix, Hulu) - [ ] Does the website feature NSFW content? (e.g. porn, etc...) - [ ] Are you a donator/patron? - [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority? **What's your Discord username?** rayyplay#4999 <!--- Optional, unless you are a donator/patron. Ex. Clyde#0000 --> **What's the name of the service?** sushiroll.co.id | Sushiroll <!--- Required, Ex. www.youtube.com | YouTube --> **What should the Presence display?** Watching Sushiroll, title anime, duration, play/pause/stop <!--- Required, make sure to be as clear as possible on what should be added. --> **If possible, please provide a logo for the service (512x512 minimum)** ![image](https://user-images.githubusercontent.com/42858646/114428192-16aaae80-9be6-11eb-9b10-00713797ded5.png) <!--- Optional, it is recommended to upload the image here instead of using a 3rd-party host. --> </div>
1.0
Sushiroll | sushiroll.co.id - ### Discussed in https://github.com/PreMiD/Presences/discussions/4591 <div type='discussions-op-text'> <sup>Originally posted by **rayyplay** April 12, 2021</sup> **Prerequisites and essential questions** <!--- Required, please answer the following questions as honestly as possible by changing the "[ ]" to "[x]" or by marking it after creating the issue (easier), not marking a question counts as "No". --> - [x] Is it a popular site? - [x] Is the website older than 2 months? <!--- It is necessary for the website to be older than 2 months. --> - [x] Is the site locked to a specific country/region? - [x] Is the site a paid service? (e.g. Netflix, Hulu) - [ ] Does the website feature NSFW content? (e.g. porn, etc...) - [ ] Are you a donator/patron? - [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority? **What's your Discord username?** rayyplay#4999 <!--- Optional, unless you are a donator/patron. Ex. Clyde#0000 --> **What's the name of the service?** sushiroll.co.id | Sushiroll <!--- Required, Ex. www.youtube.com | YouTube --> **What should the Presence display?** Watching Sushiroll, title anime, duration, play/pause/stop <!--- Required, make sure to be as clear as possible on what should be added. --> **If possible, please provide a logo for the service (512x512 minimum)** ![image](https://user-images.githubusercontent.com/42858646/114428192-16aaae80-9be6-11eb-9b10-00713797ded5.png) <!--- Optional, it is recommended to upload the image here instead of using a 3rd-party host. --> </div>
non_process
sushiroll sushiroll co id discussed in originally posted by rayyplay april prerequisites and essential questions is it a popular site is the website older than months is the site locked to a specific country region is the site a paid service e g netflix hulu does the website feature nsfw content e g porn etc are you a donator patron do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority what s your discord username rayyplay what s the name of the service sushiroll co id sushiroll what should the presence display watching sushiroll title anime duration play pause stop if possible please provide a logo for the service minimum
0
3,838
6,802,496,946
IssuesEvent
2017-11-02 20:25:15
WikiWatershed/model-my-watershed
https://api.github.com/repos/WikiWatershed/model-my-watershed
closed
Only Log Geoprocessing API Requests Not From the UI
1 Geoprocessing API
We should only log requests if they're coming from an actual API user, not a user on the app's frontend. Depends on #2270 Depends on #2190
1.0
Only Log Geoprocessing API Requests Not From the UI - We should only log requests if they're coming from an actual API user, not a user on the app's frontend. Depends on #2270 Depends on #2190
process
only log geoprocessing api requests not from the ui we should only log requests if they re coming from an actual api user not a user on the app s frontend depends on depends on
1
7,309
10,449,281,510
IssuesEvent
2019-09-19 08:04:00
stekylsha/CISC210Lab
https://api.github.com/repos/stekylsha/CISC210Lab
opened
Source Configuration Management
Software Process
### Story As a software developer I want to understand SCM, in particular git So that I will be able to control management of software releases, concurrent development, and regression reversion. ### Acceptance Criteria Demonstrate: - Understanding of SCM - Why? How? - Using git, ability to ... - clone git repository - check out specific branch - create new branch - commit changes - understand the different phases of change commit - merge changes from another branch - create pull request - Describe a simple branching strategy for development and release
1.0
Source Configuration Management - ### Story As a software developer I want to understand SCM, in particular git So that I will be able to control management of software releases, concurrent development, and regression reversion. ### Acceptance Criteria Demonstrate: - Understanding of SCM - Why? How? - Using git, ability to ... - clone git repository - check out specific branch - create new branch - commit changes - understand the different phases of change commit - merge changes from another branch - create pull request - Describe a simple branching strategy for development and release
process
source configuration management story as a software developer i want to understand scm in particular git so that i will be able to control management of software releases concurrent development and regression reversion acceptance criteria demonstrate understanding of scm why how using git ability to clone git repository check out specific branch create new branch commit changes understand the different phases of change commit merge changes from another branch create pull request describe a simple branching strategy for development and release
1
196,430
14,859,384,245
IssuesEvent
2021-01-18 18:23:17
kiriDevs/kiriCore
https://api.github.com/repos/kiriDevs/kiriCore
reopened
Write tests for MessageService
tests
Although more complex, as you need a map from at least three prefixes and some preset messages to test it fully, it would make sense to test the MessageService aswell, as it belongs to the API part and is also commonly used by kiriCore itself, making it a crucial part of the Plugin.
1.0
Write tests for MessageService - Although more complex, as you need a map from at least three prefixes and some preset messages to test it fully, it would make sense to test the MessageService aswell, as it belongs to the API part and is also commonly used by kiriCore itself, making it a crucial part of the Plugin.
non_process
write tests for messageservice although more complex as you need a map from at least three prefixes and some preset messages to test it fully it would make sense to test the messageservice aswell as it belongs to the api part and is also commonly used by kiricore itself making it a crucial part of the plugin
0
656,244
21,724,229,865
IssuesEvent
2022-05-11 05:42:45
jordan-sullivan/flashcards-2.5
https://api.github.com/repos/jordan-sullivan/flashcards-2.5
opened
Round class and tests
high priority
Round Your Round class will be the object that takes in responses and records these guesses (as well as if they are correct or incorrect). The currentCard should be the first Card in the Deck (the array of Cards) at the start of the Round Your Round class should meet the following requirements: returnCurrentCard: method that returns the current card being played
1.0
Round class and tests - Round Your Round class will be the object that takes in responses and records these guesses (as well as if they are correct or incorrect). The currentCard should be the first Card in the Deck (the array of Cards) at the start of the Round Your Round class should meet the following requirements: returnCurrentCard: method that returns the current card being played
non_process
round class and tests round your round class will be the object that takes in responses and records these guesses as well as if they are correct or incorrect the currentcard should be the first card in the deck the array of cards at the start of the round your round class should meet the following requirements returncurrentcard method that returns the current card being played
0
144,641
5,543,635,312
IssuesEvent
2017-03-22 17:20:55
jmatheson/OnMe-Backend
https://api.github.com/repos/jmatheson/OnMe-Backend
closed
Join Auth Credentials
high priority
Introduce new rule for `/api/join-` methods. Server should reject join operations if there's an existing user with requested credentials. In case user tries adding credential which belongs to other user, server should response with an error. Example response could look like: { "data" : null, "error" : "Service already belong to another user.", "code" : 1001 } Also, make sure that user is unable to `unjoin` his last authentication server. In case user has only `twitter` linked, he should not be able to `unjoin` it. In case user tries so, server should reject such operation and example response might look like. { "data" : null, "error" : "Illegal operation. Last authentication service can not be unjoined", "code" : 1002 }
1.0
Join Auth Credentials - Introduce new rule for `/api/join-` methods. Server should reject join operations if there's an existing user with requested credentials. In case user tries adding credential which belongs to other user, server should response with an error. Example response could look like: { "data" : null, "error" : "Service already belong to another user.", "code" : 1001 } Also, make sure that user is unable to `unjoin` his last authentication server. In case user has only `twitter` linked, he should not be able to `unjoin` it. In case user tries so, server should reject such operation and example response might look like. { "data" : null, "error" : "Illegal operation. Last authentication service can not be unjoined", "code" : 1002 }
non_process
join auth credentials introduce new rule for api join methods server should reject join operations if there s an existing user with requested credentials in case user tries adding credential which belongs to other user server should response with an error example response could look like data null error service already belong to another user code also make sure that user is unable to unjoin his last authentication server in case user has only twitter linked he should not be able to unjoin it in case user tries so server should reject such operation and example response might look like data null error illegal operation last authentication service can not be unjoined code
0
9,323
12,338,906,602
IssuesEvent
2020-05-14 17:12:35
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
opened
Add chair and vice chair to README
C-process T-task
I volunteered as Chair; @zbraniecki and @nciric volunteered as Vice-Chairs. Document this.
1.0
Add chair and vice chair to README - I volunteered as Chair; @zbraniecki and @nciric volunteered as Vice-Chairs. Document this.
process
add chair and vice chair to readme i volunteered as chair zbraniecki and nciric volunteered as vice chairs document this
1
222,720
17,469,615,443
IssuesEvent
2021-08-06 23:35:57
caforbes/git_fst
https://api.github.com/repos/caforbes/git_fst
opened
big T after long vowel
info needed testing
write tests for big T after long vowels, since those patterns are different than after short V. Requires documentation of various alternations (e.g. amg_oot, amg_oodit)
1.0
big T after long vowel - write tests for big T after long vowels, since those patterns are different than after short V. Requires documentation of various alternations (e.g. amg_oot, amg_oodit)
non_process
big t after long vowel write tests for big t after long vowels since those patterns are different than after short v requires documentation of various alternations e g amg oot amg oodit
0
225,527
17,269,150,363
IssuesEvent
2021-07-22 17:19:40
AyatoKirishima/Portfolio
https://api.github.com/repos/AyatoKirishima/Portfolio
closed
Barre de navigation
documentation enhancement invalid
A faire : - [x] Créer structure site - [x] Editer noms & liens - [x] Editer "logo" ou "accueil" ![image](https://user-images.githubusercontent.com/42692272/125165692-5c9cf180-e198-11eb-9a16-4745cb09f09e.png) Image : Accueil > Photographies > Plus de photographies
1.0
Barre de navigation - A faire : - [x] Créer structure site - [x] Editer noms & liens - [x] Editer "logo" ou "accueil" ![image](https://user-images.githubusercontent.com/42692272/125165692-5c9cf180-e198-11eb-9a16-4745cb09f09e.png) Image : Accueil > Photographies > Plus de photographies
non_process
barre de navigation a faire créer structure site editer noms liens editer logo ou accueil image accueil photographies plus de photographies
0
6,651
9,770,118,046
IssuesEvent
2019-06-06 10:10:06
dzhw/zofar
https://api.github.com/repos/dzhw/zofar
closed
translation process / multilingual surveys
category: service.processes prio: 9999 status: discussion type: backlog.item
optimizing the process to implement the translated survey
1.0
translation process / multilingual surveys - optimizing the process to implement the translated survey
process
translation process multilingual surveys optimizing the process to implement the translated survey
1
21,793
30,299,017,715
IssuesEvent
2023-07-10 03:19:57
winter-telescope/mirar
https://api.github.com/repos/winter-telescope/mirar
opened
[BUG] ZOGY assumes reference images have ZP key
bug processors Critical
**Describe the bug** Similar to #463. The reference image generators do not add the 'ZP' key to the reference images, however downstream it is assumed that the key is in there. You cannot even optionally specify a different key. I think we should consider a combination of: - Allowing the choice of ref ZP key - Picking a reasonable default that matches the photometric calibration processor
1.0
[BUG] ZOGY assumes reference images have ZP key - **Describe the bug** Similar to #463. The reference image generators do not add the 'ZP' key to the reference images, however downstream it is assumed that the key is in there. You cannot even optionally specify a different key. I think we should consider a combination of: - Allowing the choice of ref ZP key - Picking a reasonable default that matches the photometric calibration processor
process
zogy assumes reference images have zp key describe the bug similar to the reference image generators do not add the zp key to the reference images however downstream it is assumed that the key is in there you cannot even optionally specify a different key i think we should consider a combination of allowing the choice of ref zp key picking a reasonable default that matches the photometric calibration processor
1
298,333
25,817,649,044
IssuesEvent
2022-12-12 06:54:17
Azure/azure-sdk-for-net
https://api.github.com/repos/Azure/azure-sdk-for-net
opened
[Storage] [DataMovement] Multiple single transfer tests won't run recorded
Storage Client test-reliability
### Library name Azure.Storage.DataMovement; Azure.Storage.DataMovement.Blobs ### Please describe the feature. Tests for pushing multiple single transfers is not running well under recorded. Making this issue to investigate the reason behind that.
1.0
[Storage] [DataMovement] Multiple single transfer tests won't run recorded - ### Library name Azure.Storage.DataMovement; Azure.Storage.DataMovement.Blobs ### Please describe the feature. Tests for pushing multiple single transfers is not running well under recorded. Making this issue to investigate the reason behind that.
non_process
multiple single transfer tests won t run recorded library name azure storage datamovement azure storage datamovement blobs please describe the feature tests for pushing multiple single transfers is not running well under recorded making this issue to investigate the reason behind that
0
21,291
28,487,810,661
IssuesEvent
2023-04-18 09:04:10
JoTec2002/TINF21C_AAS_Management
https://api.github.com/repos/JoTec2002/TINF21C_AAS_Management
opened
Admin dashboard edit
in Process frontend
Admin dashboard will be able to show the actual accounts and the associated roles.
1.0
Admin dashboard edit - Admin dashboard will be able to show the actual accounts and the associated roles.
process
admin dashboard edit admin dashboard will be able to show the actual accounts and the associated roles
1
765,627
26,854,475,260
IssuesEvent
2023-02-03 13:38:20
KinsonDigital/Velaptor
https://api.github.com/repos/KinsonDigital/Velaptor
closed
🔬Research how to add different type layered rendering
high priority 🔬research
### Complete The Item Below - [X] I have updated the title without removing the 🔬 emoji. ### What To Research Currently, layered rendering has been implemented for all the different types of rendering such as textures, lines, rect, text etc. When rending on different layers for layers in its own rendering category, everything works fine. But, when rendering a rectangle and text, they do not take layering into account between the 2 different types. **Example:** Suppose we had a single rectangle and a single piece of text that was being rendered onto the screen. The rectangle was set to render on layer 10 and the text was set to render on layer 5, the text would still render on top of the rectangle. The reason for this is that the layers are only taken into account for that rendering category. The categories themselves are rendered in a specific order in the `Renderer` class and this is not configurable. Research needs to be done on how to consider layers between categories. This research needs to be done to assess the number of changes that will be required and the risk involved in making this happen. The results from this research will need to be heavily documented and then an issue created to implement this feature. > **Note** Need to look into forcing the text to render after the button face for button controls ### Research Results After doing some prototyping, it was decided that the best course of action would be to do the following: 1. Class name refactoring - Rename the batch manager class to something like `IBatchManagerProxy` - Rename the batching services to something like `IBatchManager`. Basically replace the word _**Service**_ with **Manager** 2. Create 4 new **Carbonate** events. One for each batch type. - These events are going to be pushed out by the `BatchManager` to send data to be rendered between the 4 different renderers. These rendered are coming with issue #451. 3. For performance reasons, the renderers and the `BatchManager` will use pointers to push and receive data to be rendered. 4. An `internal` and `sealed` service will be created to wrap the `Unsafe.AsPointer()` and `Unsafe.AsRef<T>` dotnet types to maintain testability. 5. The `BatchManager` will contain all 4 arrays of data of 1000 items each for each render type. 6. A new `readonly struct` will be created called `RenderItem<T>` which will hold the batch time and metadata. - The metadata will be the layer, date and time of the add, the pointer to the array data item, and the batch type. All of this information is used to properly sort what should be rendered and in what order before telling the renderers what to render. 7. A new `readonly struct` will be created called `ItemToRender` which will hold data to be sent to a renderer. - This `struct` holds data such as the batch type, layer, date/time, and a pointer to the data for the renderer to easily get the data to send to the GPU. ### Acceptance Criteria **This issue is finished when:** - [x] Research complete and issues created _(if needed)_. - [x] If any issues were created, they have been added to the _**Related Work**_ section below. ### ToDo Items - [x] Priority label added to this issue. Refer to the _**Priority Type Labels**_ section below. - [x] Issue linked to the correct project. ### Issue Dependencies _No response_ ### Related Work _No response_ ### Additional Information: **_<details closed><summary>Priority Type Labels</summary>_** | Priority Type | Label | |---------------------|-------------------| | Low Priority | `low priority` | | Medium Priority | `medium priority` | | High Priority | `high priority` | </details> ### Code of Conduct - [X] I agree to follow this project's Code of Conduct.
1.0
🔬Research how to add different type layered rendering - ### Complete The Item Below - [X] I have updated the title without removing the 🔬 emoji. ### What To Research Currently, layered rendering has been implemented for all the different types of rendering such as textures, lines, rect, text etc. When rending on different layers for layers in its own rendering category, everything works fine. But, when rendering a rectangle and text, they do not take layering into account between the 2 different types. **Example:** Suppose we had a single rectangle and a single piece of text that was being rendered onto the screen. The rectangle was set to render on layer 10 and the text was set to render on layer 5, the text would still render on top of the rectangle. The reason for this is that the layers are only taken into account for that rendering category. The categories themselves are rendered in a specific order in the `Renderer` class and this is not configurable. Research needs to be done on how to consider layers between categories. This research needs to be done to assess the number of changes that will be required and the risk involved in making this happen. The results from this research will need to be heavily documented and then an issue created to implement this feature. > **Note** Need to look into forcing the text to render after the button face for button controls ### Research Results After doing some prototyping, it was decided that the best course of action would be to do the following: 1. Class name refactoring - Rename the batch manager class to something like `IBatchManagerProxy` - Rename the batching services to something like `IBatchManager`. Basically replace the word _**Service**_ with **Manager** 2. Create 4 new **Carbonate** events. One for each batch type. - These events are going to be pushed out by the `BatchManager` to send data to be rendered between the 4 different renderers. These rendered are coming with issue #451. 3. For performance reasons, the renderers and the `BatchManager` will use pointers to push and receive data to be rendered. 4. An `internal` and `sealed` service will be created to wrap the `Unsafe.AsPointer()` and `Unsafe.AsRef<T>` dotnet types to maintain testability. 5. The `BatchManager` will contain all 4 arrays of data of 1000 items each for each render type. 6. A new `readonly struct` will be created called `RenderItem<T>` which will hold the batch time and metadata. - The metadata will be the layer, date and time of the add, the pointer to the array data item, and the batch type. All of this information is used to properly sort what should be rendered and in what order before telling the renderers what to render. 7. A new `readonly struct` will be created called `ItemToRender` which will hold data to be sent to a renderer. - This `struct` holds data such as the batch type, layer, date/time, and a pointer to the data for the renderer to easily get the data to send to the GPU. ### Acceptance Criteria **This issue is finished when:** - [x] Research complete and issues created _(if needed)_. - [x] If any issues were created, they have been added to the _**Related Work**_ section below. ### ToDo Items - [x] Priority label added to this issue. Refer to the _**Priority Type Labels**_ section below. - [x] Issue linked to the correct project. ### Issue Dependencies _No response_ ### Related Work _No response_ ### Additional Information: **_<details closed><summary>Priority Type Labels</summary>_** | Priority Type | Label | |---------------------|-------------------| | Low Priority | `low priority` | | Medium Priority | `medium priority` | | High Priority | `high priority` | </details> ### Code of Conduct - [X] I agree to follow this project's Code of Conduct.
non_process
🔬research how to add different type layered rendering complete the item below i have updated the title without removing the 🔬 emoji what to research currently layered rendering has been implemented for all the different types of rendering such as textures lines rect text etc when rending on different layers for layers in its own rendering category everything works fine but when rendering a rectangle and text they do not take layering into account between the different types example suppose we had a single rectangle and a single piece of text that was being rendered onto the screen the rectangle was set to render on layer and the text was set to render on layer the text would still render on top of the rectangle the reason for this is that the layers are only taken into account for that rendering category the categories themselves are rendered in a specific order in the renderer class and this is not configurable research needs to be done on how to consider layers between categories this research needs to be done to assess the number of changes that will be required and the risk involved in making this happen the results from this research will need to be heavily documented and then an issue created to implement this feature note need to look into forcing the text to render after the button face for button controls research results after doing some prototyping it was decided that the best course of action would be to do the following class name refactoring rename the batch manager class to something like ibatchmanagerproxy rename the batching services to something like ibatchmanager basically replace the word service with manager create new carbonate events one for each batch type these events are going to be pushed out by the batchmanager to send data to be rendered between the different renderers these rendered are coming with issue for performance reasons the renderers and the batchmanager will use pointers to push and receive data to be rendered an internal and sealed service will be created to wrap the unsafe aspointer and unsafe asref dotnet types to maintain testability the batchmanager will contain all arrays of data of items each for each render type a new readonly struct will be created called renderitem which will hold the batch time and metadata the metadata will be the layer date and time of the add the pointer to the array data item and the batch type all of this information is used to properly sort what should be rendered and in what order before telling the renderers what to render a new readonly struct will be created called itemtorender which will hold data to be sent to a renderer this struct holds data such as the batch type layer date time and a pointer to the data for the renderer to easily get the data to send to the gpu acceptance criteria this issue is finished when research complete and issues created if needed if any issues were created they have been added to the related work section below todo items priority label added to this issue refer to the priority type labels section below issue linked to the correct project issue dependencies no response related work no response additional information priority type labels priority type label low priority low priority medium priority medium priority high priority high priority code of conduct i agree to follow this project s code of conduct
0
52,473
7,765,327,904
IssuesEvent
2018-06-02 02:00:08
reactjs/react-modal
https://api.github.com/repos/reactjs/react-modal
opened
Sprint for react-modal v4.
dependencies discussion documentation help wanted react testing
It's time for react-modal to get ready for `react 16.3+`, this includes: ## Checklist 1. Deprecate old lifecycle methods and remove polyfills. 2. Decide how long we can maintain react-modal versions below `v4` (backports or another approach). 3. Improve documentation where missing. 4. Add more examples. If you want to give it a try on any item on the list, create a new issue and assign to you. Appreciate ideas regarding this update and help with all the listed items.
1.0
Sprint for react-modal v4. - It's time for react-modal to get ready for `react 16.3+`, this includes: ## Checklist 1. Deprecate old lifecycle methods and remove polyfills. 2. Decide how long we can maintain react-modal versions below `v4` (backports or another approach). 3. Improve documentation where missing. 4. Add more examples. If you want to give it a try on any item on the list, create a new issue and assign to you. Appreciate ideas regarding this update and help with all the listed items.
non_process
sprint for react modal it s time for react modal to get ready for react this includes checklist deprecate old lifecycle methods and remove polyfills decide how long we can maintain react modal versions below backports or another approach improve documentation where missing add more examples if you want to give it a try on any item on the list create a new issue and assign to you appreciate ideas regarding this update and help with all the listed items
0