Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
186,487
21,944,184,418
IssuesEvent
2022-05-23 21:40:15
CMSgov/cms-carts-seds
https://api.github.com/repos/CMSgov/cms-carts-seds
closed
SHF - cms-carts-seds - main - MEDIUM - Instance i-0fc3aad1e5600f745 is vulnerable to CVE-2021-4193
security-hub main
************************************************************** __This issue was generated from Security Hub data and is managed through automation.__ Please do not edit the title or body of this issue, or remove the security-hub tag. All other edits/comments are welcome. Finding Id: inspector/us-east-1/519095364708/81b9be79d70bb03ee622ece378452294cf5c1149 ************************************************************** ## Type of Issue: - [x] Security Hub Finding ## Title: Instance i-0fc3aad1e5600f745 is vulnerable to CVE-2021-4193 ## Id: inspector/us-east-1/519095364708/81b9be79d70bb03ee622ece378452294cf5c1149 (You may use this ID to lookup this finding's details in Security Hub) ## Description vim is vulnerable to Out-of-bounds Read ## Remediation undefined ## AC: - The security hub finding is resolved or suppressed, indicated by a Workflow Status of Resolved or Suppressed.
True
SHF - cms-carts-seds - main - MEDIUM - Instance i-0fc3aad1e5600f745 is vulnerable to CVE-2021-4193 - ************************************************************** __This issue was generated from Security Hub data and is managed through automation.__ Please do not edit the title or body of this issue, or remove the security-hub tag. All other edits/comments are welcome. Finding Id: inspector/us-east-1/519095364708/81b9be79d70bb03ee622ece378452294cf5c1149 ************************************************************** ## Type of Issue: - [x] Security Hub Finding ## Title: Instance i-0fc3aad1e5600f745 is vulnerable to CVE-2021-4193 ## Id: inspector/us-east-1/519095364708/81b9be79d70bb03ee622ece378452294cf5c1149 (You may use this ID to lookup this finding's details in Security Hub) ## Description vim is vulnerable to Out-of-bounds Read ## Remediation undefined ## AC: - The security hub finding is resolved or suppressed, indicated by a Workflow Status of Resolved or Suppressed.
non_process
shf cms carts seds main medium instance i is vulnerable to cve this issue was generated from security hub data and is managed through automation please do not edit the title or body of this issue or remove the security hub tag all other edits comments are welcome finding id inspector us east type of issue security hub finding title instance i is vulnerable to cve id inspector us east you may use this id to lookup this finding s details in security hub description vim is vulnerable to out of bounds read remediation undefined ac the security hub finding is resolved or suppressed indicated by a workflow status of resolved or suppressed
0
19,941
10,563,661,009
IssuesEvent
2019-10-04 21:39:01
SNLComputation/Albany
https://api.github.com/repos/SNLComputation/Albany
closed
'Gather Extruded 2D Nodal Parameter' routine much slower with Epetra than Tpetra
LandIce performance
In doing performance runs, I discovered that 'Gather Extruded 2D Nodal Parameter' is much slower in Epetra than Tpetra. This seems to show up for problems with tetrahedral meshes for ALI much more so than hexahedral meshes. Here are some data: Epetra: ``` Phalanx: Evaluator 5: [<Residual>] Gather Extruded 2D Nodal Parameter: 21.205 - 28.6921% [23025] {min=20.9137, max=40.8765, std dev=1.55442} Phalanx: Evaluator 85: [<Jacobian>] Gather Extruded 2D Nodal Parameter: 12.6985 - 19.9394% [13330] {min=12.5126, max=25.1912, std dev=1.01228} ``` Tpetra: ``` Phalanx: Evaluator 5: [<Residual>] Gather Extruded 2D Nodal Parameter: 0.443377 - 2.6535% [23025] {min=0.427363, max=0.693777, std dev=0.0190795} Phalanx: Evaluator 85: [<Jacobian>] Gather Extruded 2D Nodal Parameter: 0.262684 - 0.8327% [13330] {min=0.252668, max=0.422806, std dev=0.0126421} ``` @bartgol , @mperego : can one of you please have a look at this? I haven't checked if the issue was happening earlier prior to some of the Thyra refactors - I suspect it's likely not necessary to do that, but I can if you think it would be helpful.
True
'Gather Extruded 2D Nodal Parameter' routine much slower with Epetra than Tpetra - In doing performance runs, I discovered that 'Gather Extruded 2D Nodal Parameter' is much slower in Epetra than Tpetra. This seems to show up for problems with tetrahedral meshes for ALI much more so than hexahedral meshes. Here are some data: Epetra: ``` Phalanx: Evaluator 5: [<Residual>] Gather Extruded 2D Nodal Parameter: 21.205 - 28.6921% [23025] {min=20.9137, max=40.8765, std dev=1.55442} Phalanx: Evaluator 85: [<Jacobian>] Gather Extruded 2D Nodal Parameter: 12.6985 - 19.9394% [13330] {min=12.5126, max=25.1912, std dev=1.01228} ``` Tpetra: ``` Phalanx: Evaluator 5: [<Residual>] Gather Extruded 2D Nodal Parameter: 0.443377 - 2.6535% [23025] {min=0.427363, max=0.693777, std dev=0.0190795} Phalanx: Evaluator 85: [<Jacobian>] Gather Extruded 2D Nodal Parameter: 0.262684 - 0.8327% [13330] {min=0.252668, max=0.422806, std dev=0.0126421} ``` @bartgol , @mperego : can one of you please have a look at this? I haven't checked if the issue was happening earlier prior to some of the Thyra refactors - I suspect it's likely not necessary to do that, but I can if you think it would be helpful.
non_process
gather extruded nodal parameter routine much slower with epetra than tpetra in doing performance runs i discovered that gather extruded nodal parameter is much slower in epetra than tpetra this seems to show up for problems with tetrahedral meshes for ali much more so than hexahedral meshes here are some data epetra phalanx evaluator gather extruded nodal parameter min max std dev phalanx evaluator gather extruded nodal parameter min max std dev tpetra phalanx evaluator gather extruded nodal parameter min max std dev phalanx evaluator gather extruded nodal parameter min max std dev bartgol mperego can one of you please have a look at this i haven t checked if the issue was happening earlier prior to some of the thyra refactors i suspect it s likely not necessary to do that but i can if you think it would be helpful
0
49,031
12,268,289,951
IssuesEvent
2020-05-07 12:16:12
lbl-srg/modelica-buildings
https://api.github.com/repos/lbl-srg/modelica-buildings
closed
Add missing comments and info section in the OBC classes
OpenBuildingControl
Running [`modelica-json`](https://github.com/lbl-srg/modelica-json) tool as ``` node app.js -f ../modelica-buildings/Buildings/Controls/OBC -o json -m modelica -d out ``` gives a list of warnings showing that some classes in `Controls.OBC` package have missing comments, empty equation section, or misplaced info section. This is to fix them.
1.0
Add missing comments and info section in the OBC classes - Running [`modelica-json`](https://github.com/lbl-srg/modelica-json) tool as ``` node app.js -f ../modelica-buildings/Buildings/Controls/OBC -o json -m modelica -d out ``` gives a list of warnings showing that some classes in `Controls.OBC` package have missing comments, empty equation section, or misplaced info section. This is to fix them.
non_process
add missing comments and info section in the obc classes running tool as node app js f modelica buildings buildings controls obc o json m modelica d out gives a list of warnings showing that some classes in controls obc package have missing comments empty equation section or misplaced info section this is to fix them
0
109,896
13,865,385,322
IssuesEvent
2020-10-16 04:10:50
greatnewcls/BVUFT6ZTU4ZB376LM2MDD3K6
https://api.github.com/repos/greatnewcls/BVUFT6ZTU4ZB376LM2MDD3K6
reopened
Eihx+KCEYRw++9VG2Ei2YzRVri8EbTD/gW+wXF9ZizflOouP95aFGIFj68WplD3hPtq6otoTaTIl5qqTMFN3vOxMOYioBk7+ybvEQY11Jx6kfgsdAgTHJHPBFELDKujA79iaVMaYTxmNzvyRnQAkfRN+oOdM4WsmLUo19fGlzzU=
design
5SE6faGs48rrgKpYBH4mpbGMnWMFvmGG7rTI8EG5RFjVuNWc+V/eUL1yf6pCW9SgfsAYriJ/mXIYHTxNMQ5VEA1mrR9oJOzoujh97N0UWJ+v/2H6n2Hn8f1M4utdtVU7+2WtZ6pytSlxIdzw9fAWM4w1nt36p2KOTGLvodiJEw5l9Cj1EqSXAFzg+mRHMyWKle97UEN6m0R4qv22iwf1oQnmWbkJqyU/EMFVDjBBUEUg/eXrjxsq37YKwNWxfoI1CRhwgoy+2/ItX2g3AtKifd/yF3seRy3miRerKBtfjDDYXkTahXGq38VMLlSC0c3VikBbGeBHROmVykwthgm4CEB3Tq1kb7aCexmAo6hAbwp+o/e2k+wcpcMwrZv7R4g7EbukIQJQocz3WwIxg57jC+UplUDFa0oKRPbRJxh/Xme1mE6dSIyB2rZt8hpgo8sWEFPfublo88WEt2TTSTS0u9/83Zya0RcYHb///pclmqjf/N2cmtEXGB2///6XJZqo3/zdnJrRFxgdv//+lyWaqN/83Zya0RcYHb///pclmqivaMvyZ0Yag6xgcm50kIx2BukHNH2sQTCIfY8GZaZpc+522h/kmd9TtL7NdPf9+ckchaNePi/MmHUah3eHICp5qOV+Q49GAt+ByXr4YTxvSFkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOQt9gqdGkwvhdjtAhjSO92dckBynlaNT97/zAt6ynJyisfNrsFjmm+ENnoFnPcYWaXjkQ/Qe/+dF5WljY3qmZkGud8TMKrHgNUuYrlD/1I2jLbnfLx8KXCQH9D3nRMXSr7o/rqGLkb/7B/25jsIcrDVkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOfjJY3FdDXWvVj0pDGMZwkFFV7cebZlJi728Bv2JRSSzOR+j+G9vnol1Rq7Z1oGzaG+2iWfCoPSpkQFLEBFpOeEREoEtwqF3JjH9wCgfATUtHX/mORicJPHf2lCbwxceLb+DNQCvHWbrLn+ZWqWTXA1kc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOQt9gqdGkwvhdjtAhjSO92dckBynlaNT97/zAt6ynJyisfNrsFjmm+ENnoFnPcYWadmU2B3ExsziZvHyou2lTkPCl1e9SK83qEApmhqOdjoVI8Vt97+ApXfNUibKZaYJMm+lWGVH++i+62xakuKDqY3dbL/SZutvZub4SpCIjlotZHN1GPBUBdns/ascZrAOOWRzdRjwVAXZ7P2rHGawDjqlS2GAaw4k2V4eR+jIh+VHQP7gFs4GbJ549I1mFr/gxl83P37ahrIj2Q+W882L1BuRQt2UF+59vA04Bo31iLBcF+mY8NTjqWtAjX9IdlG4ElkQSnBzh5YOc1ICnMTaL+Ig76T+gPv3VMcTHZyy6DAZZHN1GPBUBdns/ascZrAOOWRzdRjwVAXZ7P2rHGawDjqlS2GAaw4k2V4eR+jIh+VHQP7gFs4GbJ549I1mFr/gxusy+WQXY3pls6psDwyIfOYzNP5xHTc/WmfCfuy6ZqHof/RgeJgeU8AlZj45kBDpoKue9czOG1WUZ5AoBZjPtvUbHIawFFsM35TLr4UU5BDxZHN1GPBUBdns/ascZrAOOWRzdRjwVAXZ7P2rHGawDjmeVDO/mCxnC9TNt0/gnKxPYHmNtK2Czc+FWloeBCkpUGJQzlOi9uCmbsdtYh73+cMlDXXxTzWl7/cC/HpBnZeFV4zPgma2zApVHgus+/GlPDZjILTCI/iAo+nbqoxwoqetLAYYaiIUpBeh54sSNMO1ZHN1GPBUBdns/ascZrAOOWRzdRjwVAXZ7P2rHGawDjjAazurGnQOsaKwyBCE/eX4eFLs6irgapzJ1H9Jaw++56DXjVFbjCxd5IeA6y5dmoWEiPG29+blCAQC3wjmutl6FU6QDVHVo5nzhhhh4o6oPauwmp75fp25nOirc0hIgSFkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOORMfNuKYswDNNJuRUxbBJNUOVWUMvLGWjX4xz0edk7q3vI+0M/3cId4JO9QLLdY5dugQtR3btSg7UbzV2eRCZlrZDd1VZAuDDUVh5Ije/Qjt/CtwDvPIEvsIYwPlnJ46srFK4VJuwy+PE07fp/Q7SQlkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOgROSVzMrL3c6HXQELwT9sdKLbtvbVF50Wwj/RMRZ5koy4HBTu08szhE3Dh3Q4ZhiGeii3q6hgoAcIVfadwHuRrT3kOTX/AATSRRG5EF9k1PIMv/aaYWPmIrrl6JrQvcwRschrAUWwzflMuvhRTkEPFkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOZ5UM7+YLGcL1M23T+CcrE9geY20rYLNz4VaWh4EKSlR38x3J10RyZOQfzy9mOYm0yUNdfFPNaXv9wL8ekGdl4VXjM+CZrbMClUeC6z78aU+JafDsKfhatcKClILlLGCeWRzdRjwVAXZ7P2rHGawDjlkc3UY8FQF2ez9qxxmsA45Ex824pizAM00m5FTFsEk1Q5VZQy8sZaNfjHPR52Ture8j7Qz/dwh3gk71Ast1jl3gylil+sz5/RokggNpuc4KUsPtBWVwuz/tbs53NNS5sEJY4kvOyrxjS7Q1tzzdcDBvqYFL0DYiZ+d6n1UcVLHJWRzdRjwVAXZ7P2rHGawDjlkc3UY8FQF2ez9qxxmsA46pUthgGsOJNleHkfoyIflR0D+4BbOBmyeePSNZha/4MbrMvlkF2N6ZbOqbA8MiHznWI2z7Sttj2RFuX79hCZJh5SAVL4K2OPLb3o+B7hjRsRR05+WGaSmvuaREA2gAevQDQeuQkOXWg4h6HBRiMT3P1xk7P9UU5sB2lM9XhXfkp9sEjyuxPVYKFNlBWkqaOpcUtMQ2bSBV2BTt9t7wRlDnI7pgHkk+K/7v8atGVklZZ/ApHMGsVYs6jjHQK+SL/hXmjECwBGoUk3ffYF+BzIoQ25w1mzOS36lg2S6TEpT3MduykLwduf04jsHXXJw0t6POZw+lTWf7sueZvN8gZhR1SSA0QxbsSLd2k37dX9qIQr2XClzOj6LolASrcLM4yk9XB9GSfjvdh8TtI5ne0udy
1.0
Eihx+KCEYRw++9VG2Ei2YzRVri8EbTD/gW+wXF9ZizflOouP95aFGIFj68WplD3hPtq6otoTaTIl5qqTMFN3vOxMOYioBk7+ybvEQY11Jx6kfgsdAgTHJHPBFELDKujA79iaVMaYTxmNzvyRnQAkfRN+oOdM4WsmLUo19fGlzzU= - 5SE6faGs48rrgKpYBH4mpbGMnWMFvmGG7rTI8EG5RFjVuNWc+V/eUL1yf6pCW9SgfsAYriJ/mXIYHTxNMQ5VEA1mrR9oJOzoujh97N0UWJ+v/2H6n2Hn8f1M4utdtVU7+2WtZ6pytSlxIdzw9fAWM4w1nt36p2KOTGLvodiJEw5l9Cj1EqSXAFzg+mRHMyWKle97UEN6m0R4qv22iwf1oQnmWbkJqyU/EMFVDjBBUEUg/eXrjxsq37YKwNWxfoI1CRhwgoy+2/ItX2g3AtKifd/yF3seRy3miRerKBtfjDDYXkTahXGq38VMLlSC0c3VikBbGeBHROmVykwthgm4CEB3Tq1kb7aCexmAo6hAbwp+o/e2k+wcpcMwrZv7R4g7EbukIQJQocz3WwIxg57jC+UplUDFa0oKRPbRJxh/Xme1mE6dSIyB2rZt8hpgo8sWEFPfublo88WEt2TTSTS0u9/83Zya0RcYHb///pclmqjf/N2cmtEXGB2///6XJZqo3/zdnJrRFxgdv//+lyWaqN/83Zya0RcYHb///pclmqivaMvyZ0Yag6xgcm50kIx2BukHNH2sQTCIfY8GZaZpc+522h/kmd9TtL7NdPf9+ckchaNePi/MmHUah3eHICp5qOV+Q49GAt+ByXr4YTxvSFkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOQt9gqdGkwvhdjtAhjSO92dckBynlaNT97/zAt6ynJyisfNrsFjmm+ENnoFnPcYWaXjkQ/Qe/+dF5WljY3qmZkGud8TMKrHgNUuYrlD/1I2jLbnfLx8KXCQH9D3nRMXSr7o/rqGLkb/7B/25jsIcrDVkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOfjJY3FdDXWvVj0pDGMZwkFFV7cebZlJi728Bv2JRSSzOR+j+G9vnol1Rq7Z1oGzaG+2iWfCoPSpkQFLEBFpOeEREoEtwqF3JjH9wCgfATUtHX/mORicJPHf2lCbwxceLb+DNQCvHWbrLn+ZWqWTXA1kc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOQt9gqdGkwvhdjtAhjSO92dckBynlaNT97/zAt6ynJyisfNrsFjmm+ENnoFnPcYWadmU2B3ExsziZvHyou2lTkPCl1e9SK83qEApmhqOdjoVI8Vt97+ApXfNUibKZaYJMm+lWGVH++i+62xakuKDqY3dbL/SZutvZub4SpCIjlotZHN1GPBUBdns/ascZrAOOWRzdRjwVAXZ7P2rHGawDjqlS2GAaw4k2V4eR+jIh+VHQP7gFs4GbJ549I1mFr/gxl83P37ahrIj2Q+W882L1BuRQt2UF+59vA04Bo31iLBcF+mY8NTjqWtAjX9IdlG4ElkQSnBzh5YOc1ICnMTaL+Ig76T+gPv3VMcTHZyy6DAZZHN1GPBUBdns/ascZrAOOWRzdRjwVAXZ7P2rHGawDjqlS2GAaw4k2V4eR+jIh+VHQP7gFs4GbJ549I1mFr/gxusy+WQXY3pls6psDwyIfOYzNP5xHTc/WmfCfuy6ZqHof/RgeJgeU8AlZj45kBDpoKue9czOG1WUZ5AoBZjPtvUbHIawFFsM35TLr4UU5BDxZHN1GPBUBdns/ascZrAOOWRzdRjwVAXZ7P2rHGawDjmeVDO/mCxnC9TNt0/gnKxPYHmNtK2Czc+FWloeBCkpUGJQzlOi9uCmbsdtYh73+cMlDXXxTzWl7/cC/HpBnZeFV4zPgma2zApVHgus+/GlPDZjILTCI/iAo+nbqoxwoqetLAYYaiIUpBeh54sSNMO1ZHN1GPBUBdns/ascZrAOOWRzdRjwVAXZ7P2rHGawDjjAazurGnQOsaKwyBCE/eX4eFLs6irgapzJ1H9Jaw++56DXjVFbjCxd5IeA6y5dmoWEiPG29+blCAQC3wjmutl6FU6QDVHVo5nzhhhh4o6oPauwmp75fp25nOirc0hIgSFkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOORMfNuKYswDNNJuRUxbBJNUOVWUMvLGWjX4xz0edk7q3vI+0M/3cId4JO9QLLdY5dugQtR3btSg7UbzV2eRCZlrZDd1VZAuDDUVh5Ije/Qjt/CtwDvPIEvsIYwPlnJ46srFK4VJuwy+PE07fp/Q7SQlkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOgROSVzMrL3c6HXQELwT9sdKLbtvbVF50Wwj/RMRZ5koy4HBTu08szhE3Dh3Q4ZhiGeii3q6hgoAcIVfadwHuRrT3kOTX/AATSRRG5EF9k1PIMv/aaYWPmIrrl6JrQvcwRschrAUWwzflMuvhRTkEPFkc3UY8FQF2ez9qxxmsA45ZHN1GPBUBdns/ascZrAOOZ5UM7+YLGcL1M23T+CcrE9geY20rYLNz4VaWh4EKSlR38x3J10RyZOQfzy9mOYm0yUNdfFPNaXv9wL8ekGdl4VXjM+CZrbMClUeC6z78aU+JafDsKfhatcKClILlLGCeWRzdRjwVAXZ7P2rHGawDjlkc3UY8FQF2ez9qxxmsA45Ex824pizAM00m5FTFsEk1Q5VZQy8sZaNfjHPR52Ture8j7Qz/dwh3gk71Ast1jl3gylil+sz5/RokggNpuc4KUsPtBWVwuz/tbs53NNS5sEJY4kvOyrxjS7Q1tzzdcDBvqYFL0DYiZ+d6n1UcVLHJWRzdRjwVAXZ7P2rHGawDjlkc3UY8FQF2ez9qxxmsA46pUthgGsOJNleHkfoyIflR0D+4BbOBmyeePSNZha/4MbrMvlkF2N6ZbOqbA8MiHznWI2z7Sttj2RFuX79hCZJh5SAVL4K2OPLb3o+B7hjRsRR05+WGaSmvuaREA2gAevQDQeuQkOXWg4h6HBRiMT3P1xk7P9UU5sB2lM9XhXfkp9sEjyuxPVYKFNlBWkqaOpcUtMQ2bSBV2BTt9t7wRlDnI7pgHkk+K/7v8atGVklZZ/ApHMGsVYs6jjHQK+SL/hXmjECwBGoUk3ffYF+BzIoQ25w1mzOS36lg2S6TEpT3MduykLwduf04jsHXXJw0t6POZw+lTWf7sueZvN8gZhR1SSA0QxbsSLd2k37dX9qIQr2XClzOj6LolASrcLM4yk9XB9GSfjvdh8TtI5ne0udy
non_process
eihx kceyrw gw v v emfvdjbbueug o pclmqjf zdnjrrfxgdv lywaqn ckchanepi ennofnpcywaxjkq qe rqglkb j dnqcvhwbrln apxfnuibkzayjmm lwgvh i jih jih gxusy cc glpdzjiltci iao qjt k sl
0
2,252
5,088,652,308
IssuesEvent
2017-01-01 00:04:17
sw4j-org/tool-jpa-processor
https://api.github.com/repos/sw4j-org/tool-jpa-processor
opened
Handle @PrimaryKeyJoinColumns Annotation
annotation processor task
Handle the `@PrimaryKeyJoinColumns` annotation for a property or field. See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf) - 11.1.45 PrimaryKeyJoinColumns Annotation
1.0
Handle @PrimaryKeyJoinColumns Annotation - Handle the `@PrimaryKeyJoinColumns` annotation for a property or field. See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf) - 11.1.45 PrimaryKeyJoinColumns Annotation
process
handle primarykeyjoincolumns annotation handle the primarykeyjoincolumns annotation for a property or field see primarykeyjoincolumns annotation
1
78,477
27,542,345,937
IssuesEvent
2023-03-07 09:25:34
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Long name should be truncated in the pills
T-Defect
### Steps to reproduce ### Steps to reproduce Send a mention for a user with a long name ### Outcome #### What did you expect? a pill with a truncated name #### What happened instead? The pill is cropped <img width="394" alt="image" src="https://user-images.githubusercontent.com/8969772/223379193-07d9b40c-5eb5-494f-aef8-a784c909f874.png"> ### Outcome Here is the suggested design: <img width="1266" alt="image" src="https://user-images.githubusercontent.com/8969772/223365164-af330790-c8c9-41fa-a4ab-a064b54034c6.png"> ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store _No response_ ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? Yes
1.0
Long name should be truncated in the pills - ### Steps to reproduce ### Steps to reproduce Send a mention for a user with a long name ### Outcome #### What did you expect? a pill with a truncated name #### What happened instead? The pill is cropped <img width="394" alt="image" src="https://user-images.githubusercontent.com/8969772/223379193-07d9b40c-5eb5-494f-aef8-a784c909f874.png"> ### Outcome Here is the suggested design: <img width="1266" alt="image" src="https://user-images.githubusercontent.com/8969772/223365164-af330790-c8c9-41fa-a4ab-a064b54034c6.png"> ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store _No response_ ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? Yes
non_process
long name should be truncated in the pills steps to reproduce steps to reproduce send a mention for a user with a long name outcome what did you expect a pill with a truncated name what happened instead the pill is cropped img width alt image src outcome here is the suggested design img width alt image src your phone model no response operating system version no response application version and app store no response homeserver no response will you send logs no are you willing to provide a pr yes
0
64,220
6,896,465,779
IssuesEvent
2017-11-23 17:59:54
joserogerio/promocaldasSite
https://api.github.com/repos/joserogerio/promocaldasSite
closed
Mostrar data de inicio e fim da promoção
melhoramento test
Na listagem de promoções mostrar os dados de inicio e fim da mesma.
1.0
Mostrar data de inicio e fim da promoção - Na listagem de promoções mostrar os dados de inicio e fim da mesma.
non_process
mostrar data de inicio e fim da promoção na listagem de promoções mostrar os dados de inicio e fim da mesma
0
95,683
10,885,503,745
IssuesEvent
2019-11-18 10:31:41
AzureAD/microsoft-authentication-library-for-dotnet
https://api.github.com/repos/AzureAD/microsoft-authentication-library-for-dotnet
closed
Azure AD B2C with SSO implementation in Xamarin Forms
answered documentation question
Is it possible to implement Single Sign On along with Azure AD B2C. If yes, Could you please suggest some sample / Documentation about the same. I am trying to implement the same in my Xamarin Application.
1.0
Azure AD B2C with SSO implementation in Xamarin Forms - Is it possible to implement Single Sign On along with Azure AD B2C. If yes, Could you please suggest some sample / Documentation about the same. I am trying to implement the same in my Xamarin Application.
non_process
azure ad with sso implementation in xamarin forms is it possible to implement single sign on along with azure ad if yes could you please suggest some sample documentation about the same i am trying to implement the same in my xamarin application
0
12,380
14,898,056,730
IssuesEvent
2021-01-21 12:37:54
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Problems with ReviewApp
Pri2 devops-cicd-process/tech devops/prod doc-bug
I'm attempting to configure the reviewApp flow to verify that the dynamic env creation works as expected. This is a fragment of a stage that contains the deployment job ```yaml stages: - stage: deployment_stage_cluster_${{ parameters['Kubernetes.Namespace'] }} displayName: "deploy to ${{ parameters['Kubernetes.Cluster'] }}" dependsOn: [] jobs: - deployment: DeployPullRequest displayName: Deploy Pull request condition: and(succeeded(), not(startsWith(variables['Build.SourceBranch'], 'refs/pull/'))) pool: vmImage: ${{ parameters['Agent.Pool'] }} environment: azureday2020-demo-aks.test strategy: runOnce: deploy: steps: - reviewApp: test - task: KubectlInstaller@0 displayName: "kubectl installer" inputs: kubectlVersion: ${{ parameters['Kubernetes.Kubectl.Version'] }} - task: HelmInstaller@1 displayName: "helm installer" inputs: helmVersionToInstall: ${{ parameters['Kubernetes.Helm.Version'] }} - task: Kubernetes@1 displayName: 'Create a new namespace for the pull request' inputs: command: apply useConfigurationFile: true inline: '{ "kind": "Namespace", "apiVersion": "v1", "metadata": { "name": "test" }}' ``` Problems encountered: - Attempts of running a pipeline containing the above template yield an immediate error: `Job Deployment: Resource test does not exist in environment azureday2020-demo-aks` - This appears to work differntly than advertised? I was under the impression that both the environment and the resource were to be provisioned automatically? - When attempting to use template parameters to define the environment name : `environment: '${{ parameters['ClusterName] }}.${{ parameters['Namespace] }}'` I'm getting consistent failures stating that `environmentName` is invalid/missing --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7730ae4d-4101-9c83-1823-4ff43ff161ce * Version Independent ID: 20a7e263-4819-783e-c984-c4f3b459e22f * Content: [Environment - Kubernetes resource - Azure Pipelines](https://docs.microsoft.com/en-in/azure/devops/pipelines/process/environments-kubernetes?view=azure-devops#setup-review-app) * Content Source: [docs/pipelines/process/environments-kubernetes.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/environments-kubernetes.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Problems with ReviewApp - I'm attempting to configure the reviewApp flow to verify that the dynamic env creation works as expected. This is a fragment of a stage that contains the deployment job ```yaml stages: - stage: deployment_stage_cluster_${{ parameters['Kubernetes.Namespace'] }} displayName: "deploy to ${{ parameters['Kubernetes.Cluster'] }}" dependsOn: [] jobs: - deployment: DeployPullRequest displayName: Deploy Pull request condition: and(succeeded(), not(startsWith(variables['Build.SourceBranch'], 'refs/pull/'))) pool: vmImage: ${{ parameters['Agent.Pool'] }} environment: azureday2020-demo-aks.test strategy: runOnce: deploy: steps: - reviewApp: test - task: KubectlInstaller@0 displayName: "kubectl installer" inputs: kubectlVersion: ${{ parameters['Kubernetes.Kubectl.Version'] }} - task: HelmInstaller@1 displayName: "helm installer" inputs: helmVersionToInstall: ${{ parameters['Kubernetes.Helm.Version'] }} - task: Kubernetes@1 displayName: 'Create a new namespace for the pull request' inputs: command: apply useConfigurationFile: true inline: '{ "kind": "Namespace", "apiVersion": "v1", "metadata": { "name": "test" }}' ``` Problems encountered: - Attempts of running a pipeline containing the above template yield an immediate error: `Job Deployment: Resource test does not exist in environment azureday2020-demo-aks` - This appears to work differntly than advertised? I was under the impression that both the environment and the resource were to be provisioned automatically? - When attempting to use template parameters to define the environment name : `environment: '${{ parameters['ClusterName] }}.${{ parameters['Namespace] }}'` I'm getting consistent failures stating that `environmentName` is invalid/missing --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7730ae4d-4101-9c83-1823-4ff43ff161ce * Version Independent ID: 20a7e263-4819-783e-c984-c4f3b459e22f * Content: [Environment - Kubernetes resource - Azure Pipelines](https://docs.microsoft.com/en-in/azure/devops/pipelines/process/environments-kubernetes?view=azure-devops#setup-review-app) * Content Source: [docs/pipelines/process/environments-kubernetes.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/environments-kubernetes.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
problems with reviewapp i m attempting to configure the reviewapp flow to verify that the dynamic env creation works as expected this is a fragment of a stage that contains the deployment job yaml stages stage deployment stage cluster parameters displayname deploy to parameters dependson jobs deployment deploypullrequest displayname deploy pull request condition and succeeded not startswith variables refs pull pool vmimage parameters environment demo aks test strategy runonce deploy steps reviewapp test task kubectlinstaller displayname kubectl installer inputs kubectlversion parameters task helminstaller displayname helm installer inputs helmversiontoinstall parameters task kubernetes displayname create a new namespace for the pull request inputs command apply useconfigurationfile true inline kind namespace apiversion metadata name test problems encountered attempts of running a pipeline containing the above template yield an immediate error job deployment resource test does not exist in environment demo aks this appears to work differntly than advertised i was under the impression that both the environment and the resource were to be provisioned automatically when attempting to use template parameters to define the environment name environment parameters parameters i m getting consistent failures stating that environmentname is invalid missing document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
14,930
18,359,530,389
IssuesEvent
2021-10-09 01:46:05
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Wrong html processing in the overridden document.write function
TYPE: bug AREA: client health-monitor FREQUENCY: level 1 SYSTEM: client side processing STATE: Stale
Code for reproducing: ```js document.write('<script data-rp-insertionmarker="<body>" '); document.write('id="id">'); document.write('<\/script>'); ``` https://www.jiji.com
1.0
Wrong html processing in the overridden document.write function - Code for reproducing: ```js document.write('<script data-rp-insertionmarker="<body>" '); document.write('id="id">'); document.write('<\/script>'); ``` https://www.jiji.com
process
wrong html processing in the overridden document write function code for reproducing js document write document write id id document write
1
20,226
26,822,263,703
IssuesEvent
2023-02-02 10:18:28
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Unable to open terminal with linux arm and arm64 servers on qemu
bug linux virtual-machine terminal-process
Sanity testing `1.75` on linux servers following the steps at https://github.com/microsoft/vscode-remote-release/wiki/Sanity-Check-VS-Code-Servers#linux-platforms and attempting to open a terminal throws the following error. <img width="571" alt="Screenshot 2023-02-01 at 11 47 03" src="https://user-images.githubusercontent.com/964386/215932968-f0af2873-cd6e-470f-b786-2beb16966c0d.png"> Issue is from the pty module, following snippet can trigger the error on any of the containers ``` > var ptyProcess = pty.spawn('bash', [], {cwd: process.env.HOME}); Unsupported ioctl: cmd=0x5441 Uncaught Error: chdir() failed: Not a directory at new UnixTerminal (/root/.vscode-server/bin/e8bf7514f31cef005c988b206c61948e56aab9cb/node_modules/node-pty/lib/unixTerminal.js:106:24) at Object.spawn (/root/.vscode-server/bin/e8bf7514f31cef005c988b206c61948e56aab9cb/node_modules/node-pty/lib/index.js:29:12) ```
1.0
Unable to open terminal with linux arm and arm64 servers on qemu - Sanity testing `1.75` on linux servers following the steps at https://github.com/microsoft/vscode-remote-release/wiki/Sanity-Check-VS-Code-Servers#linux-platforms and attempting to open a terminal throws the following error. <img width="571" alt="Screenshot 2023-02-01 at 11 47 03" src="https://user-images.githubusercontent.com/964386/215932968-f0af2873-cd6e-470f-b786-2beb16966c0d.png"> Issue is from the pty module, following snippet can trigger the error on any of the containers ``` > var ptyProcess = pty.spawn('bash', [], {cwd: process.env.HOME}); Unsupported ioctl: cmd=0x5441 Uncaught Error: chdir() failed: Not a directory at new UnixTerminal (/root/.vscode-server/bin/e8bf7514f31cef005c988b206c61948e56aab9cb/node_modules/node-pty/lib/unixTerminal.js:106:24) at Object.spawn (/root/.vscode-server/bin/e8bf7514f31cef005c988b206c61948e56aab9cb/node_modules/node-pty/lib/index.js:29:12) ```
process
unable to open terminal with linux arm and servers on qemu sanity testing on linux servers following the steps at and attempting to open a terminal throws the following error img width alt screenshot at src issue is from the pty module following snippet can trigger the error on any of the containers var ptyprocess pty spawn bash cwd process env home unsupported ioctl cmd uncaught error chdir failed not a directory at new unixterminal root vscode server bin node modules node pty lib unixterminal js at object spawn root vscode server bin node modules node pty lib index js
1
677,336
23,158,949,540
IssuesEvent
2022-07-29 15:34:36
ramp4-pcar4/ramp4-pcar4
https://api.github.com/repos/ramp4-pcar4/ramp4-pcar4
closed
Feature Highlighting
effort: far away flavour: feature priority: nice type: adaptive needs: consensus consult: ui/ux
This will likely require some upfront design work. Things to consider - GeoAPI does have a highlight layer, but it is essentially untested - In RAMP2, the "fogging" of the map behind the highlight was done by directly manipulating the SVG that represented the ESRI map. If we do the same, will need to see if things are different in ESRI 4, and be aware that WebGL may be used instead of SVG. - An API function to request a highlight sounds like a fine idea - Functions to get individual geometries from a layer already exist in GeoAPI Related design discussion https://github.com/ramp4-pcar4/ramp4-pcar4/discussions/837
1.0
Feature Highlighting - This will likely require some upfront design work. Things to consider - GeoAPI does have a highlight layer, but it is essentially untested - In RAMP2, the "fogging" of the map behind the highlight was done by directly manipulating the SVG that represented the ESRI map. If we do the same, will need to see if things are different in ESRI 4, and be aware that WebGL may be used instead of SVG. - An API function to request a highlight sounds like a fine idea - Functions to get individual geometries from a layer already exist in GeoAPI Related design discussion https://github.com/ramp4-pcar4/ramp4-pcar4/discussions/837
non_process
feature highlighting this will likely require some upfront design work things to consider geoapi does have a highlight layer but it is essentially untested in the fogging of the map behind the highlight was done by directly manipulating the svg that represented the esri map if we do the same will need to see if things are different in esri and be aware that webgl may be used instead of svg an api function to request a highlight sounds like a fine idea functions to get individual geometries from a layer already exist in geoapi related design discussion
0
6,799
9,937,549,657
IssuesEvent
2019-07-02 22:18:19
googleapis/nodejs-storage
https://api.github.com/repos/googleapis/nodejs-storage
closed
refactor: remove dependency on async module
type: process
we still rely on `async` in a few places, which broke a recent release of the module (as it was a devDependency only).
1.0
refactor: remove dependency on async module - we still rely on `async` in a few places, which broke a recent release of the module (as it was a devDependency only).
process
refactor remove dependency on async module we still rely on async in a few places which broke a recent release of the module as it was a devdependency only
1
19,350
25,481,561,060
IssuesEvent
2022-11-25 22:08:18
kdgregory/log4j-aws-appenders
https://api.github.com/repos/kdgregory/log4j-aws-appenders
reopened
ERROR writer initialization timed out
enhancement in-process
I see this error message when application starts, and it hangs for 60 seconds as timeout code indicates. `2022-10-19 16:04:14,305 main ERROR writer initialization timed out` However, everything else works without a problem... I would like to ask for some hints where it could be hanging? My setup: 1. log4j2.xml configuration ``` xml <?xml version="1.0" encoding="UTF-8"?> <Configuration status="INFO" packages="com.kdgregory.log4j2.aws"> <Appenders> <Console name="ConsoleAppender" target="SYSTEM_OUT"> <PatternLayout pattern="%d{yyy-MM-dd HH:mm:ss.SSS} [%t] %-5level %logger{36} - %X{AWS-XRAY-TRACE-ID} - %msg%n" /> </Console> <CloudWatchAppender name="CLOUDWATCH"> <logGroup>log4j-appender-demo</logGroup> <logStream>demo-{date}-{hostname}-${awslogs:pid}</logStream> <dedicatedWriter>true</dedicatedWriter> <PatternLayout pattern="%d{yyyy-MM-dd HH:mm:ss.SSS} %-5p [%t] %c - %X{AWS-XRAY-TRACE-ID} - %m%n" /> </CloudWatchAppender> </Appenders> <Loggers> <Root level="info"> <AppenderRef ref="ConsoleAppender" /> <AppenderRef ref="CLOUDWATCH" /> </Root> </Loggers> </Configuration> ``` 2. dependencies: ``` xml <dependencies> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-web</artifactId> <exclusions> <exclusion> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-logging</artifactId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.springframework.cloud</groupId> <artifactId>spring-cloud-starter-openfeign</artifactId> </dependency> <dependency> <groupId>io.github.openfeign</groupId> <artifactId>feign-httpclient</artifactId> </dependency> <dependency> <groupId>org.springframework.cloud</groupId> <artifactId>spring-cloud-starter-loadbalancer</artifactId> </dependency> <!-- log related dependencies--> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-log4j2</artifactId> </dependency> <dependency> <groupId>com.kdgregory.logging</groupId> <artifactId>log4j2-aws-appenders</artifactId> <version>3.0.1</version> </dependency> <dependency> <groupId>com.kdgregory.logging</groupId> <artifactId>aws-facade-v2</artifactId> <version>3.0.1</version> <exclusions> <exclusion> <groupId>software.amazon.awssdk</groupId> <artifactId>sts</artifactId> </exclusion> </exclusions> </dependency> <dependency> <groupId>software.amazon.awssdk</groupId> <artifactId>cloudwatchlogs</artifactId> <version>2.17.281</version> </dependency> <dependency> <groupId>software.amazon.awssdk</groupId> <artifactId>sts</artifactId> <version>2.17.281</version> </dependency> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-test</artifactId> <scope>test</scope> </dependency> </dependencies> ``` 3. Scenario: ADOT with Log4j2 - I would like to leverage open-telemetry java agent to inject `%X{AWS-XRAY-TRACE-ID}` into cloud watch logs - credential chain: role assume is from serviceaccount OIDC provider. and this is why I have to upgrade sts SDK version So... With this setup, everything works but hanging for 1 minutes when starts - very annoying... Can anyone bing me a silver-lining? Thanks
1.0
ERROR writer initialization timed out - I see this error message when application starts, and it hangs for 60 seconds as timeout code indicates. `2022-10-19 16:04:14,305 main ERROR writer initialization timed out` However, everything else works without a problem... I would like to ask for some hints where it could be hanging? My setup: 1. log4j2.xml configuration ``` xml <?xml version="1.0" encoding="UTF-8"?> <Configuration status="INFO" packages="com.kdgregory.log4j2.aws"> <Appenders> <Console name="ConsoleAppender" target="SYSTEM_OUT"> <PatternLayout pattern="%d{yyy-MM-dd HH:mm:ss.SSS} [%t] %-5level %logger{36} - %X{AWS-XRAY-TRACE-ID} - %msg%n" /> </Console> <CloudWatchAppender name="CLOUDWATCH"> <logGroup>log4j-appender-demo</logGroup> <logStream>demo-{date}-{hostname}-${awslogs:pid}</logStream> <dedicatedWriter>true</dedicatedWriter> <PatternLayout pattern="%d{yyyy-MM-dd HH:mm:ss.SSS} %-5p [%t] %c - %X{AWS-XRAY-TRACE-ID} - %m%n" /> </CloudWatchAppender> </Appenders> <Loggers> <Root level="info"> <AppenderRef ref="ConsoleAppender" /> <AppenderRef ref="CLOUDWATCH" /> </Root> </Loggers> </Configuration> ``` 2. dependencies: ``` xml <dependencies> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-web</artifactId> <exclusions> <exclusion> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-logging</artifactId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.springframework.cloud</groupId> <artifactId>spring-cloud-starter-openfeign</artifactId> </dependency> <dependency> <groupId>io.github.openfeign</groupId> <artifactId>feign-httpclient</artifactId> </dependency> <dependency> <groupId>org.springframework.cloud</groupId> <artifactId>spring-cloud-starter-loadbalancer</artifactId> </dependency> <!-- log related dependencies--> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-log4j2</artifactId> </dependency> <dependency> <groupId>com.kdgregory.logging</groupId> <artifactId>log4j2-aws-appenders</artifactId> <version>3.0.1</version> </dependency> <dependency> <groupId>com.kdgregory.logging</groupId> <artifactId>aws-facade-v2</artifactId> <version>3.0.1</version> <exclusions> <exclusion> <groupId>software.amazon.awssdk</groupId> <artifactId>sts</artifactId> </exclusion> </exclusions> </dependency> <dependency> <groupId>software.amazon.awssdk</groupId> <artifactId>cloudwatchlogs</artifactId> <version>2.17.281</version> </dependency> <dependency> <groupId>software.amazon.awssdk</groupId> <artifactId>sts</artifactId> <version>2.17.281</version> </dependency> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-test</artifactId> <scope>test</scope> </dependency> </dependencies> ``` 3. Scenario: ADOT with Log4j2 - I would like to leverage open-telemetry java agent to inject `%X{AWS-XRAY-TRACE-ID}` into cloud watch logs - credential chain: role assume is from serviceaccount OIDC provider. and this is why I have to upgrade sts SDK version So... With this setup, everything works but hanging for 1 minutes when starts - very annoying... Can anyone bing me a silver-lining? Thanks
process
error writer initialization timed out i see this error message when application starts and it hangs for seconds as timeout code indicates main error writer initialization timed out however everything else works without a problem i would like to ask for some hints where it could be hanging my setup xml configuration xml appender demo demo date hostname awslogs pid true dependencies xml org springframework boot spring boot starter web org springframework boot spring boot starter logging org springframework cloud spring cloud starter openfeign io github openfeign feign httpclient org springframework cloud spring cloud starter loadbalancer org springframework boot spring boot starter com kdgregory logging aws appenders com kdgregory logging aws facade software amazon awssdk sts software amazon awssdk cloudwatchlogs software amazon awssdk sts org springframework boot spring boot starter test test scenario adot with i would like to leverage open telemetry java agent to inject x aws xray trace id into cloud watch logs credential chain role assume is from serviceaccount oidc provider and this is why i have to upgrade sts sdk version so with this setup everything works but hanging for minutes when starts very annoying can anyone bing me a silver lining thanks
1
17,288
23,096,844,325
IssuesEvent
2022-07-26 20:29:05
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
CASE statements don't evaluate to False if using `, False` and nested CASE statements
Type:Bug Priority:P2 Querying/Processor .Backend Querying/Notebook/Custom Column
**Describe the bug** Creating 2 custom columns that are case statements, then referencing those in a 3rd custom column with a case statement that evaluates to `True` or `False`, results in a bug where the 3rd column doesn't evaluate to `False` - just `NULL`. **Creating custom columns** First <img width="486" alt="image" src="https://user-images.githubusercontent.com/61659989/179753529-81aec074-9e2f-4e33-b8be-2050e8f1da40.png"> Second <img width="486" alt="image" src="https://user-images.githubusercontent.com/61659989/179753789-3aa70339-97ec-452d-8b49-7b165eaf7d22.png"> Third, that references the above 2 - evaluating to True/False <img width="486" alt="image" src="https://user-images.githubusercontent.com/61659989/179753936-3fc69b20-7b7a-4f43-9f46-fdc1ae77350b.png"> **Generated SQL in BigQuery** <img width="904" alt="image" src="https://user-images.githubusercontent.com/61659989/179754328-ed0cfcbd-8bf3-470d-9288-7ce10db28b4a.png"> This should have an `ELSE FALSE` evaluation **Tweaking the third column to return binary output instead** <img width="484" alt="image" src="https://user-images.githubusercontent.com/61659989/179754483-1f5e9fe6-5176-4201-82f0-49d08eaeee70.png"> Generated SQL - now correctly showing the `ELSE` statement <img width="881" alt="image" src="https://user-images.githubusercontent.com/61659989/179754791-3e905f91-5daf-4ee1-a6ec-d426ac6cd715.png">
1.0
CASE statements don't evaluate to False if using `, False` and nested CASE statements - **Describe the bug** Creating 2 custom columns that are case statements, then referencing those in a 3rd custom column with a case statement that evaluates to `True` or `False`, results in a bug where the 3rd column doesn't evaluate to `False` - just `NULL`. **Creating custom columns** First <img width="486" alt="image" src="https://user-images.githubusercontent.com/61659989/179753529-81aec074-9e2f-4e33-b8be-2050e8f1da40.png"> Second <img width="486" alt="image" src="https://user-images.githubusercontent.com/61659989/179753789-3aa70339-97ec-452d-8b49-7b165eaf7d22.png"> Third, that references the above 2 - evaluating to True/False <img width="486" alt="image" src="https://user-images.githubusercontent.com/61659989/179753936-3fc69b20-7b7a-4f43-9f46-fdc1ae77350b.png"> **Generated SQL in BigQuery** <img width="904" alt="image" src="https://user-images.githubusercontent.com/61659989/179754328-ed0cfcbd-8bf3-470d-9288-7ce10db28b4a.png"> This should have an `ELSE FALSE` evaluation **Tweaking the third column to return binary output instead** <img width="484" alt="image" src="https://user-images.githubusercontent.com/61659989/179754483-1f5e9fe6-5176-4201-82f0-49d08eaeee70.png"> Generated SQL - now correctly showing the `ELSE` statement <img width="881" alt="image" src="https://user-images.githubusercontent.com/61659989/179754791-3e905f91-5daf-4ee1-a6ec-d426ac6cd715.png">
process
case statements don t evaluate to false if using false and nested case statements describe the bug creating custom columns that are case statements then referencing those in a custom column with a case statement that evaluates to true or false results in a bug where the column doesn t evaluate to false just null creating custom columns first img width alt image src second img width alt image src third that references the above evaluating to true false img width alt image src generated sql in bigquery img width alt image src this should have an else false evaluation tweaking the third column to return binary output instead img width alt image src generated sql now correctly showing the else statement img width alt image src
1
1,017
3,479,429,277
IssuesEvent
2015-12-28 20:14:13
USC-CSSL/TACIT
https://api.github.com/repos/USC-CSSL/TACIT
opened
preprocessing: correct spelling option
Preprocessing
In preprocessing, can we provide an option to automatically correct misspellings?
1.0
preprocessing: correct spelling option - In preprocessing, can we provide an option to automatically correct misspellings?
process
preprocessing correct spelling option in preprocessing can we provide an option to automatically correct misspellings
1
230,648
25,482,741,430
IssuesEvent
2022-11-26 01:22:46
panasalap/linux-4.1.15
https://api.github.com/repos/panasalap/linux-4.1.15
reopened
CVE-2017-6345 (Medium) detected in linuxlinux-4.1.17, linuxlinux-4.1.17
security vulnerability
## CVE-2017-6345 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.1.17</b>, <b>linuxlinux-4.1.17</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The LLC subsystem in the Linux kernel before 4.9.13 does not ensure that a certain destructor exists in required circumstances, which allows local users to cause a denial of service (BUG_ON) or possibly have unspecified other impact via crafted system calls. <p>Publish Date: 2017-03-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-6345>CVE-2017-6345</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-6345">https://nvd.nist.gov/vuln/detail/CVE-2017-6345</a></p> <p>Release Date: 2017-03-01</p> <p>Fix Resolution: 4.9.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-6345 (Medium) detected in linuxlinux-4.1.17, linuxlinux-4.1.17 - ## CVE-2017-6345 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.1.17</b>, <b>linuxlinux-4.1.17</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The LLC subsystem in the Linux kernel before 4.9.13 does not ensure that a certain destructor exists in required circumstances, which allows local users to cause a denial of service (BUG_ON) or possibly have unspecified other impact via crafted system calls. <p>Publish Date: 2017-03-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-6345>CVE-2017-6345</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-6345">https://nvd.nist.gov/vuln/detail/CVE-2017-6345</a></p> <p>Release Date: 2017-03-01</p> <p>Fix Resolution: 4.9.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linuxlinux linuxlinux cve medium severity vulnerability vulnerable libraries linuxlinux linuxlinux vulnerability details the llc subsystem in the linux kernel before does not ensure that a certain destructor exists in required circumstances which allows local users to cause a denial of service bug on or possibly have unspecified other impact via crafted system calls publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
2,794
5,723,471,012
IssuesEvent
2017-04-20 12:21:04
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
References to topics in ditavalref context lead to broken links in XHTML TOC
bug P2 preprocess/filtering
I'm attaching a sample project: [branchFilteringJobProblem.zip](https://github.com/dita-ot/dita-ot/files/922135/branchFilteringJobProblem.zip) Publish to XHTML and then try to open from the TOC the first link. It points to a non-existing HTML document. It seems that the BranchFilterModule no longer updates the Job object with the new "-1", '-2'... names it gives to topics in ditavalref contexts. Leading to .job.xml not containing <file> mappings for those automatically generated topics. It's a recent change, before this commit @jelovirt made: https://github.com/dita-ot/dita-ot/commit/07331c3193ebd0cf846ff6a35185075599101914 it had a "generateCopies" method which seemed to update the Job object with the new topics....
1.0
References to topics in ditavalref context lead to broken links in XHTML TOC - I'm attaching a sample project: [branchFilteringJobProblem.zip](https://github.com/dita-ot/dita-ot/files/922135/branchFilteringJobProblem.zip) Publish to XHTML and then try to open from the TOC the first link. It points to a non-existing HTML document. It seems that the BranchFilterModule no longer updates the Job object with the new "-1", '-2'... names it gives to topics in ditavalref contexts. Leading to .job.xml not containing <file> mappings for those automatically generated topics. It's a recent change, before this commit @jelovirt made: https://github.com/dita-ot/dita-ot/commit/07331c3193ebd0cf846ff6a35185075599101914 it had a "generateCopies" method which seemed to update the Job object with the new topics....
process
references to topics in ditavalref context lead to broken links in xhtml toc i m attaching a sample project publish to xhtml and then try to open from the toc the first link it points to a non existing html document it seems that the branchfiltermodule no longer updates the job object with the new names it gives to topics in ditavalref contexts leading to job xml not containing mappings for those automatically generated topics it s a recent change before this commit jelovirt made it had a generatecopies method which seemed to update the job object with the new topics
1
7,343
10,479,536,998
IssuesEvent
2019-09-24 04:37:27
OI-wiki/OI-wiki
https://api.github.com/repos/OI-wiki/OI-wiki
closed
LCT 页面代码与描述混乱
需要处理 / Need Processing 需要帮助 / help wanted
首先,十分欢迎你花钱(?)来给 OI WIki 开 issue,在提交之前,请花时间阅读一下这个模板的内容,谢谢合作! - 是出现了什么问题?(最好截图) https://oi-wiki.org/ds/lct/ LCT 页面 文中定义的函数名字与代码中用到的不符(主要是大小写不一样),且建树中的代码风格(数组)与全文的代码风格(指针)不一致。在 建树 中,分裂 和 合并 又被讲了一次(而且写的并不是很详细) - 你是否正在着手修复? 否
1.0
LCT 页面代码与描述混乱 - 首先,十分欢迎你花钱(?)来给 OI WIki 开 issue,在提交之前,请花时间阅读一下这个模板的内容,谢谢合作! - 是出现了什么问题?(最好截图) https://oi-wiki.org/ds/lct/ LCT 页面 文中定义的函数名字与代码中用到的不符(主要是大小写不一样),且建树中的代码风格(数组)与全文的代码风格(指针)不一致。在 建树 中,分裂 和 合并 又被讲了一次(而且写的并不是很详细) - 你是否正在着手修复? 否
process
lct 页面代码与描述混乱 首先,十分欢迎你花钱(?)来给 oi wiki 开 issue,在提交之前,请花时间阅读一下这个模板的内容,谢谢合作! 是出现了什么问题?(最好截图) lct 页面 文中定义的函数名字与代码中用到的不符(主要是大小写不一样),且建树中的代码风格(数组)与全文的代码风格(指针)不一致。在 建树 中,分裂 和 合并 又被讲了一次(而且写的并不是很详细) 你是否正在着手修复? 否
1
11,171
13,957,694,788
IssuesEvent
2020-10-24 08:11:21
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
SE: Harvesting request
Geoportal Harvesting process SE - Sweden
Dear Angelo, Please perform a harvest on the Swedish CSW. We have some updates that needs to be checked. Kind Regards Fredrik Pers&auml;ter
1.0
SE: Harvesting request - Dear Angelo, Please perform a harvest on the Swedish CSW. We have some updates that needs to be checked. Kind Regards Fredrik Pers&auml;ter
process
se harvesting request dear angelo please perform a harvest on the swedish csw we have some updates that needs to be checked kind regards fredrik pers auml ter
1
20,886
27,710,772,156
IssuesEvent
2023-03-14 14:07:43
pcg-platinus/feedback
https://api.github.com/repos/pcg-platinus/feedback
closed
Ich bin Prozessanalyst und möchte den "richtigen" Zeitpunkt in einem Projekt für die Modellierung von Prozessen finden
BPM-BusinessProcessManagement
Bei uns werden in Projekten oft Prozesse in Visio gezeichnet, weil man auf unterster (meist manueller oder sogar applikatorischer) Ebene den GESAMTEN Prozess sehen möchte. In Repositories wie Adonis wären hier für alle Beteiligte entsprechende Read&Explore-Lizenzen notwendig, die man sich (inklusive der Aufwände wie Schulungen) sparen möchte und das Designen einer Zeichnung einfacher ist als einer objektorientierten Datenbank (mit ihren Modellierungs- und Objektablage-Guidelines). Ist es klug, bis zur Abnahme durch den Kunden damit zu warten und dann erst den fertigen Prozess zu übertragen (also am Ende des Prozessschrittes, in dem das To-Be erstellt wird) und ihn in die richtigen Haupt-/Subprozesse zu schneiden? Zu welchem Zeitpunkt sollte man spätestens die Visio-Prozesse ins Adonis übertragen?
1.0
Ich bin Prozessanalyst und möchte den "richtigen" Zeitpunkt in einem Projekt für die Modellierung von Prozessen finden - Bei uns werden in Projekten oft Prozesse in Visio gezeichnet, weil man auf unterster (meist manueller oder sogar applikatorischer) Ebene den GESAMTEN Prozess sehen möchte. In Repositories wie Adonis wären hier für alle Beteiligte entsprechende Read&Explore-Lizenzen notwendig, die man sich (inklusive der Aufwände wie Schulungen) sparen möchte und das Designen einer Zeichnung einfacher ist als einer objektorientierten Datenbank (mit ihren Modellierungs- und Objektablage-Guidelines). Ist es klug, bis zur Abnahme durch den Kunden damit zu warten und dann erst den fertigen Prozess zu übertragen (also am Ende des Prozessschrittes, in dem das To-Be erstellt wird) und ihn in die richtigen Haupt-/Subprozesse zu schneiden? Zu welchem Zeitpunkt sollte man spätestens die Visio-Prozesse ins Adonis übertragen?
process
ich bin prozessanalyst und möchte den richtigen zeitpunkt in einem projekt für die modellierung von prozessen finden bei uns werden in projekten oft prozesse in visio gezeichnet weil man auf unterster meist manueller oder sogar applikatorischer ebene den gesamten prozess sehen möchte in repositories wie adonis wären hier für alle beteiligte entsprechende read explore lizenzen notwendig die man sich inklusive der aufwände wie schulungen sparen möchte und das designen einer zeichnung einfacher ist als einer objektorientierten datenbank mit ihren modellierungs und objektablage guidelines ist es klug bis zur abnahme durch den kunden damit zu warten und dann erst den fertigen prozess zu übertragen also am ende des prozessschrittes in dem das to be erstellt wird und ihn in die richtigen haupt subprozesse zu schneiden zu welchem zeitpunkt sollte man spätestens die visio prozesse ins adonis übertragen
1
14,477
4,937,905,417
IssuesEvent
2016-11-29 09:31:42
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Hidden menu item to connect broken links in search engines to the new link/menu-item
No Code Attached Yet
### Steps to reproduce the issue I changed different menu aliasses to a new nicer alias but the searchengines still have the old one. i want to put in menu-item with the old alias but dont want them to be seen in the menu on the website. so i want functional hidden menu-item that will be link thru the menu-item alias option to the new menu alias. ### Expected result ### Actual result ### System information (as much as possible) ### Additional comments
1.0
Hidden menu item to connect broken links in search engines to the new link/menu-item - ### Steps to reproduce the issue I changed different menu aliasses to a new nicer alias but the searchengines still have the old one. i want to put in menu-item with the old alias but dont want them to be seen in the menu on the website. so i want functional hidden menu-item that will be link thru the menu-item alias option to the new menu alias. ### Expected result ### Actual result ### System information (as much as possible) ### Additional comments
non_process
hidden menu item to connect broken links in search engines to the new link menu item steps to reproduce the issue i changed different menu aliasses to a new nicer alias but the searchengines still have the old one i want to put in menu item with the old alias but dont want them to be seen in the menu on the website so i want functional hidden menu item that will be link thru the menu item alias option to the new menu alias expected result actual result system information as much as possible additional comments
0
14,075
16,945,524,264
IssuesEvent
2021-06-28 06:05:40
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] HTML breakage in review consent screen and consent PDF for custom generated custom document
Bug P1 Process: Fixed Process: Tested QA Process: Tested dev iOS
Steps: 1. Add a custom generated custom document from SB 2. Publish updates 3. Enroll into study from iOS 4. Navigate to Review consent screen 5. Observe Actual: HTML breakage in review consent screen for custom generated custom document Expected: Review consent screen for custom generated custom document should display proper Note: Issue not observed in other screens Issue not observed in Android Study details to verify: 1. Instance: Dev 2. Study: FinalStudy Screenshots: ![unnamed](https://user-images.githubusercontent.com/60386291/123210657-f0579800-d4df-11eb-93b2-89eda671c701.png) ![pdf](https://user-images.githubusercontent.com/60386291/123210714-082f1c00-d4e0-11eb-95bc-0b8656774028.png)
3.0
[iOS] HTML breakage in review consent screen and consent PDF for custom generated custom document - Steps: 1. Add a custom generated custom document from SB 2. Publish updates 3. Enroll into study from iOS 4. Navigate to Review consent screen 5. Observe Actual: HTML breakage in review consent screen for custom generated custom document Expected: Review consent screen for custom generated custom document should display proper Note: Issue not observed in other screens Issue not observed in Android Study details to verify: 1. Instance: Dev 2. Study: FinalStudy Screenshots: ![unnamed](https://user-images.githubusercontent.com/60386291/123210657-f0579800-d4df-11eb-93b2-89eda671c701.png) ![pdf](https://user-images.githubusercontent.com/60386291/123210714-082f1c00-d4e0-11eb-95bc-0b8656774028.png)
process
html breakage in review consent screen and consent pdf for custom generated custom document steps add a custom generated custom document from sb publish updates enroll into study from ios navigate to review consent screen observe actual html breakage in review consent screen for custom generated custom document expected review consent screen for custom generated custom document should display proper note issue not observed in other screens issue not observed in android study details to verify instance dev study finalstudy screenshots
1
65,058
7,854,986,390
IssuesEvent
2018-06-20 22:59:19
Opentrons/opentrons
https://api.github.com/repos/Opentrons/opentrons
closed
Error: Require Ingredient Name and Volume
feature protocol designer small
As a user, I would like to be alerted if I attempt to create an ingredient without a volume ## Acceptance Criteria - If user attempts to click 'save' when volume is blank or 0, show error on field and prevent save - Same if the user attempts to click 'save' when the name is empty -- Volume Field (and/or Name Field) has orange error state -- Volume Field (and/or Name Field) shows error copy ## Design ![image](https://user-images.githubusercontent.com/35570080/41000398-d60e4f70-68db-11e8-8ff9-3bbffe65a19c.png)
1.0
Error: Require Ingredient Name and Volume - As a user, I would like to be alerted if I attempt to create an ingredient without a volume ## Acceptance Criteria - If user attempts to click 'save' when volume is blank or 0, show error on field and prevent save - Same if the user attempts to click 'save' when the name is empty -- Volume Field (and/or Name Field) has orange error state -- Volume Field (and/or Name Field) shows error copy ## Design ![image](https://user-images.githubusercontent.com/35570080/41000398-d60e4f70-68db-11e8-8ff9-3bbffe65a19c.png)
non_process
error require ingredient name and volume as a user i would like to be alerted if i attempt to create an ingredient without a volume acceptance criteria if user attempts to click save when volume is blank or show error on field and prevent save same if the user attempts to click save when the name is empty volume field and or name field has orange error state volume field and or name field shows error copy design
0
52,554
13,224,834,551
IssuesEvent
2020-08-17 19:56:43
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[serialization] std::auto_ptr is depricated (Trac #2417)
Incomplete Migration Migrated from Trac combo core defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2417">https://code.icecube.wisc.edu/projects/icecube/ticket/2417</a>, reported by kjmeagherand owned by olivas</em></summary> <p> ```json { "status": "accepted", "changetime": "2020-06-30T12:16:59", "_ts": "1593519419169166", "description": "I am getting a lot of warnings with gcc 9.2.1 that auto_ptr is depricated. And word on the street is that it will be deprecated at some point in the future. I think that it can just be replaced with unique_ptr.", "reporter": "kjmeagher", "cc": "", "resolution": "", "time": "2020-03-11T00:00:43", "component": "combo core", "summary": "[serialization] std::auto_ptr is depricated", "priority": "normal", "keywords": "", "milestone": "Winter Solstice 2020", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
[serialization] std::auto_ptr is depricated (Trac #2417) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2417">https://code.icecube.wisc.edu/projects/icecube/ticket/2417</a>, reported by kjmeagherand owned by olivas</em></summary> <p> ```json { "status": "accepted", "changetime": "2020-06-30T12:16:59", "_ts": "1593519419169166", "description": "I am getting a lot of warnings with gcc 9.2.1 that auto_ptr is depricated. And word on the street is that it will be deprecated at some point in the future. I think that it can just be replaced with unique_ptr.", "reporter": "kjmeagher", "cc": "", "resolution": "", "time": "2020-03-11T00:00:43", "component": "combo core", "summary": "[serialization] std::auto_ptr is depricated", "priority": "normal", "keywords": "", "milestone": "Winter Solstice 2020", "owner": "olivas", "type": "defect" } ``` </p> </details>
non_process
std auto ptr is depricated trac migrated from json status accepted changetime ts description i am getting a lot of warnings with gcc that auto ptr is depricated and word on the street is that it will be deprecated at some point in the future i think that it can just be replaced with unique ptr reporter kjmeagher cc resolution time component combo core summary std auto ptr is depricated priority normal keywords milestone winter solstice owner olivas type defect
0
8,249
11,421,369,779
IssuesEvent
2020-02-03 12:02:31
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
Logging postcss warnings and messages in Parcel
:bug: Bug CSS Preprocessing Stale
# ❔ Question Hello, I recently installed `'postcss-font-base64` (with `npm install --save-dev`), but it doesn't seem to be working for me (output is not changed in any way). My `postcss.config.js` is shown below, and can verify it is being used - if I remove the `cssnano` entry, the CSS is no longer minimized. How can I debug this? I saw the `postcss-reporter` plugin and tried to install it in a similar fashion, adding it tot he end of my config, but didn't see any messages being reported when running `npm run prod` (which calls `parcel build`). ```javascript module.exports = { plugins: [ require('autoprefixer'), require('postcss-font-base64')({ // no options yet }), require('cssnano')({ preset: ['default', { reduceTransforms: true }] }) ] }; ``` ## 🔦 Context I want to embed fonts (all the kinds supported by `postcss-font-base64`) within css files, to make shipping my CSS easier for users of it. I understand this may be an issue with a postcss plugin and not parcel, but I don't know how to log warnings and such when using postcss with parcel. ## 💻 Code Sample (can try to put something together) ## 🌍 Your Environment | Software | Version(s) | | ---------------- | ---------- | | Parcel | 1.12.3 | Node | v12.2.0 | npm/Yarn | 6.9.0 | Operating System | Linux 035dc2ced347 4.15.0-48-generic #51-Ubuntu SMP Wed Apr 3 08:28:49 UTC 2019 x86_64 GNU/Linux
1.0
Logging postcss warnings and messages in Parcel - # ❔ Question Hello, I recently installed `'postcss-font-base64` (with `npm install --save-dev`), but it doesn't seem to be working for me (output is not changed in any way). My `postcss.config.js` is shown below, and can verify it is being used - if I remove the `cssnano` entry, the CSS is no longer minimized. How can I debug this? I saw the `postcss-reporter` plugin and tried to install it in a similar fashion, adding it tot he end of my config, but didn't see any messages being reported when running `npm run prod` (which calls `parcel build`). ```javascript module.exports = { plugins: [ require('autoprefixer'), require('postcss-font-base64')({ // no options yet }), require('cssnano')({ preset: ['default', { reduceTransforms: true }] }) ] }; ``` ## 🔦 Context I want to embed fonts (all the kinds supported by `postcss-font-base64`) within css files, to make shipping my CSS easier for users of it. I understand this may be an issue with a postcss plugin and not parcel, but I don't know how to log warnings and such when using postcss with parcel. ## 💻 Code Sample (can try to put something together) ## 🌍 Your Environment | Software | Version(s) | | ---------------- | ---------- | | Parcel | 1.12.3 | Node | v12.2.0 | npm/Yarn | 6.9.0 | Operating System | Linux 035dc2ced347 4.15.0-48-generic #51-Ubuntu SMP Wed Apr 3 08:28:49 UTC 2019 x86_64 GNU/Linux
process
logging postcss warnings and messages in parcel ❔ question hello i recently installed postcss font with npm install save dev but it doesn t seem to be working for me output is not changed in any way my postcss config js is shown below and can verify it is being used if i remove the cssnano entry the css is no longer minimized how can i debug this i saw the postcss reporter plugin and tried to install it in a similar fashion adding it tot he end of my config but didn t see any messages being reported when running npm run prod which calls parcel build javascript module exports plugins require autoprefixer require postcss font no options yet require cssnano preset default reducetransforms true 🔦 context i want to embed fonts all the kinds supported by postcss font within css files to make shipping my css easier for users of it i understand this may be an issue with a postcss plugin and not parcel but i don t know how to log warnings and such when using postcss with parcel 💻 code sample can try to put something together 🌍 your environment software version s parcel node npm yarn operating system linux generic ubuntu smp wed apr utc gnu linux
1
15,699
19,848,265,749
IssuesEvent
2022-01-21 09:23:00
ooi-data/CE07SHSM-MFD35-02-PRESFB000-recovered_host-presf_abc_dcl_tide_measurement_recovered
https://api.github.com/repos/ooi-data/CE07SHSM-MFD35-02-PRESFB000-recovered_host-presf_abc_dcl_tide_measurement_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:22:59.931299. ## Details Flow name: `CE07SHSM-MFD35-02-PRESFB000-recovered_host-presf_abc_dcl_tide_measurement_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:22:59.931299. ## Details Flow name: `CE07SHSM-MFD35-02-PRESFB000-recovered_host-presf_abc_dcl_tide_measurement_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host presf abc dcl tide measurement recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
43,103
11,157,882,047
IssuesEvent
2019-12-25 15:47:23
JonasWanke/Unicorn
https://api.github.com/repos/JonasWanke/Unicorn
closed
Update libraries
T: build
<!-- Thanks for taking the time to file an issue! Please select the component label (C: abc) this feature is related to from the right (if applicable). --> **Description** <!-- A clear and concise description of the problem or missing capability and possibly its solution --> <!-- Add any other context about the problem here **Additional context** -->
1.0
Update libraries - <!-- Thanks for taking the time to file an issue! Please select the component label (C: abc) this feature is related to from the right (if applicable). --> **Description** <!-- A clear and concise description of the problem or missing capability and possibly its solution --> <!-- Add any other context about the problem here **Additional context** -->
non_process
update libraries thanks for taking the time to file an issue please select the component label c abc this feature is related to from the right if applicable description add any other context about the problem here additional context
0
44,201
17,909,283,713
IssuesEvent
2021-09-09 01:24:29
hashicorp/terraform-provider-azurerm
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
closed
azurerm_virtual_network_gateway does not allow multiple vpn_client_configuration blocks
enhancement service/virtual-network-gateway
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform (and AzureRM Provider) Version ``` Terraform v0.14.4 + provider registry.terraform.io/hashicorp/azurerm v2.56.0 ``` ### Affected Resource(s) * azurerm_virtual_network_gateway ### Terraform Configuration Files ```hcl locals { location = "westeurope" } resource "azurerm_resource_group" "example" { name = "example-rg" location = local.location } resource "azurerm_virtual_network" "example" { name = "example-vnet" resource_group_name = azurerm_resource_group.example.name location = local.location address_space = ["10.0.0.0/24"] } resource "azurerm_subnet" "example" { name = "example-snet" resource_group_name = azurerm_resource_group.example.name virtual_network_name = azurerm_virtual_network.example.name address_prefixes = ["10.0.0.0/24"] } resource "azurerm_public_ip" "example" { name = "example-pip" resource_group_name = azurerm_resource_group.example.name location = local.location allocation_method = "Dynamic" ip_version = "IPv4" } ``` and ```hcl resource "azurerm_virtual_network_gateway" "example" { name = "example-vgw" resource_group_name = azurerm_resource_group.example.name location = local.location generation = "Generation2" sku = "VpnGw2" type = "Vpn" vpn_type = "RouteBased" ip_configuration { name = "example-vgw-ipconfig" public_ip_address_id = azurerm_public_ip.example.id subnet_id = azurerm_subnet.example.id } vpn_client_configuration { address_space = ["10.0.1.0/24"] vpn_client_protocols = ["OpenVPN"] aad_tenant = "" root_certificate { name = "P2SRootCert" public_cert_data = <certificate> } } } ``` or (less preferable, as it would not reflect the configuration options in the Azure portal) ```hcl resource "azurerm_virtual_network_gateway" "example" { name = "example-vgw" resource_group_name = azurerm_resource_group.example.name location = local.location generation = "Generation2" sku = "VpnGw2" type = "Vpn" vpn_type = "RouteBased" ip_configuration { name = "example-vgw-ipconfig" public_ip_address_id = azurerm_public_ip.example.id subnet_id = azurerm_subnet.example.id } vpn_client_configuration { address_space = ["10.0.1.0/24"] vpn_client_protocols = ["OpenVPN"] root_certificate { name = "P2SRootCert" public_cert_data = "" } } vpn_client_configuration { address_space = ["10.0.1.0/24"] vpn_client_protocols = ["OpenVPN"] aad_tenant = "" } } } ``` ### Expected Behaviour Azure Virtual Network Gateway being created with two authentication types. In the Azure Portal it is possible to select multiple authentication types for a Point-to-site configuration: ![grafik](https://user-images.githubusercontent.com/16082432/115560505-f13c3580-a2b4-11eb-8c50-1d19c4898e49.png) According to the Terraform documentation, there can be only one vpn_client_configuration block and within that block, aad_tenant and root_certificate are incompatible. https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/virtual_network_gateway#vpn_client_configuration ### Actual Behaviour ```hcl Error: "vpn_client_configuration.0.root_certificate": conflicts with vpn_client_configuration.0.aad_tenant ``` or ```hcl Error: vpn_client_configuration: attribute supports 1 item maximum, config has 2 declared ``` ### Steps to Reproduce 1. `terraform plan`
1.0
azurerm_virtual_network_gateway does not allow multiple vpn_client_configuration blocks - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform (and AzureRM Provider) Version ``` Terraform v0.14.4 + provider registry.terraform.io/hashicorp/azurerm v2.56.0 ``` ### Affected Resource(s) * azurerm_virtual_network_gateway ### Terraform Configuration Files ```hcl locals { location = "westeurope" } resource "azurerm_resource_group" "example" { name = "example-rg" location = local.location } resource "azurerm_virtual_network" "example" { name = "example-vnet" resource_group_name = azurerm_resource_group.example.name location = local.location address_space = ["10.0.0.0/24"] } resource "azurerm_subnet" "example" { name = "example-snet" resource_group_name = azurerm_resource_group.example.name virtual_network_name = azurerm_virtual_network.example.name address_prefixes = ["10.0.0.0/24"] } resource "azurerm_public_ip" "example" { name = "example-pip" resource_group_name = azurerm_resource_group.example.name location = local.location allocation_method = "Dynamic" ip_version = "IPv4" } ``` and ```hcl resource "azurerm_virtual_network_gateway" "example" { name = "example-vgw" resource_group_name = azurerm_resource_group.example.name location = local.location generation = "Generation2" sku = "VpnGw2" type = "Vpn" vpn_type = "RouteBased" ip_configuration { name = "example-vgw-ipconfig" public_ip_address_id = azurerm_public_ip.example.id subnet_id = azurerm_subnet.example.id } vpn_client_configuration { address_space = ["10.0.1.0/24"] vpn_client_protocols = ["OpenVPN"] aad_tenant = "" root_certificate { name = "P2SRootCert" public_cert_data = <certificate> } } } ``` or (less preferable, as it would not reflect the configuration options in the Azure portal) ```hcl resource "azurerm_virtual_network_gateway" "example" { name = "example-vgw" resource_group_name = azurerm_resource_group.example.name location = local.location generation = "Generation2" sku = "VpnGw2" type = "Vpn" vpn_type = "RouteBased" ip_configuration { name = "example-vgw-ipconfig" public_ip_address_id = azurerm_public_ip.example.id subnet_id = azurerm_subnet.example.id } vpn_client_configuration { address_space = ["10.0.1.0/24"] vpn_client_protocols = ["OpenVPN"] root_certificate { name = "P2SRootCert" public_cert_data = "" } } vpn_client_configuration { address_space = ["10.0.1.0/24"] vpn_client_protocols = ["OpenVPN"] aad_tenant = "" } } } ``` ### Expected Behaviour Azure Virtual Network Gateway being created with two authentication types. In the Azure Portal it is possible to select multiple authentication types for a Point-to-site configuration: ![grafik](https://user-images.githubusercontent.com/16082432/115560505-f13c3580-a2b4-11eb-8c50-1d19c4898e49.png) According to the Terraform documentation, there can be only one vpn_client_configuration block and within that block, aad_tenant and root_certificate are incompatible. https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/virtual_network_gateway#vpn_client_configuration ### Actual Behaviour ```hcl Error: "vpn_client_configuration.0.root_certificate": conflicts with vpn_client_configuration.0.aad_tenant ``` or ```hcl Error: vpn_client_configuration: attribute supports 1 item maximum, config has 2 declared ``` ### Steps to Reproduce 1. `terraform plan`
non_process
azurerm virtual network gateway does not allow multiple vpn client configuration blocks community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version terraform provider registry terraform io hashicorp azurerm affected resource s azurerm virtual network gateway terraform configuration files hcl locals location westeurope resource azurerm resource group example name example rg location local location resource azurerm virtual network example name example vnet resource group name azurerm resource group example name location local location address space resource azurerm subnet example name example snet resource group name azurerm resource group example name virtual network name azurerm virtual network example name address prefixes resource azurerm public ip example name example pip resource group name azurerm resource group example name location local location allocation method dynamic ip version and hcl resource azurerm virtual network gateway example name example vgw resource group name azurerm resource group example name location local location generation sku type vpn vpn type routebased ip configuration name example vgw ipconfig public ip address id azurerm public ip example id subnet id azurerm subnet example id vpn client configuration address space vpn client protocols aad tenant root certificate name public cert data or less preferable as it would not reflect the configuration options in the azure portal hcl resource azurerm virtual network gateway example name example vgw resource group name azurerm resource group example name location local location generation sku type vpn vpn type routebased ip configuration name example vgw ipconfig public ip address id azurerm public ip example id subnet id azurerm subnet example id vpn client configuration address space vpn client protocols root certificate name public cert data vpn client configuration address space vpn client protocols aad tenant expected behaviour azure virtual network gateway being created with two authentication types in the azure portal it is possible to select multiple authentication types for a point to site configuration according to the terraform documentation there can be only one vpn client configuration block and within that block aad tenant and root certificate are incompatible actual behaviour hcl error vpn client configuration root certificate conflicts with vpn client configuration aad tenant or hcl error vpn client configuration attribute supports item maximum config has declared steps to reproduce terraform plan
0
11,001
13,789,293,449
IssuesEvent
2020-10-09 08:39:25
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Test failure: System.ServiceProcess.Tests.ServiceBaseTests.TestOnExecuteCustomCommand
area-System.ServiceProcess
failed in job:[runtime-libraries-coreclr outerloop 20201008.1 ](https://dev.azure.com/dnceng/public/_build/results?buildId=845001&view=ms.vss-test-web.build-test-results-tab&runId=26906848&paneView=debug&resultId=103443) net6.0-Windows_NT-Release-x64-CoreCLR_release-(Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c Error message ~~~ Assert.Equal() Failure Expected: 129 Actual: 128 Stack trace at System.ServiceProcess.Tests.ServiceBaseTests.TestOnExecuteCustomCommand() in /_/src/libraries/System.ServiceProcess.ServiceController/tests/ServiceBaseTests.cs:line 156 ~~~
1.0
Test failure: System.ServiceProcess.Tests.ServiceBaseTests.TestOnExecuteCustomCommand - failed in job:[runtime-libraries-coreclr outerloop 20201008.1 ](https://dev.azure.com/dnceng/public/_build/results?buildId=845001&view=ms.vss-test-web.build-test-results-tab&runId=26906848&paneView=debug&resultId=103443) net6.0-Windows_NT-Release-x64-CoreCLR_release-(Windows.Server.Core.1909.Amd64.Open)windows.10.amd64.server20h1.open@mcr.microsoft.com/dotnet-buildtools/prereqs:windowsservercore-2004-helix-amd64-20200904200251-272704c Error message ~~~ Assert.Equal() Failure Expected: 129 Actual: 128 Stack trace at System.ServiceProcess.Tests.ServiceBaseTests.TestOnExecuteCustomCommand() in /_/src/libraries/System.ServiceProcess.ServiceController/tests/ServiceBaseTests.cs:line 156 ~~~
process
test failure system serviceprocess tests servicebasetests testonexecutecustomcommand failed in job windows nt release coreclr release windows server core open windows open mcr microsoft com dotnet buildtools prereqs windowsservercore helix error message assert equal failure expected actual stack trace at system serviceprocess tests servicebasetests testonexecutecustomcommand in src libraries system serviceprocess servicecontroller tests servicebasetests cs line
1
6,702
9,814,875,477
IssuesEvent
2019-06-13 11:15:44
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
ogr2ogr to merge several shapefiles in a folder
Feature Request Processing
Author Name: **Filipe Dias** (@fsdias) Original Redmine Issue: [11635](https://issues.qgis.org/issues/11635) Redmine category:processing/gdal Assignee: Giovanni Manghi --- ogr2ogr can be used to merge several shapefiles. It would be useful to have an ogr2ogr based tool that merges several shapefiles stored in a folder. Here's an example code written by Darren Cope: mkdir merged for %f in (*.shp) do ( if not exist merged\\merged.shp ( ogr2ogr -f “esri shapefile” merged\\merged.shp %f) else ( ogr2ogr -f “esri shapefile” -update -append merged\\merged.shp %f -nln Merged ) ) Source: http://darrencope.com/2010/05/07/merge-a-directory-of-shapefiles-using-ogr/
1.0
ogr2ogr to merge several shapefiles in a folder - Author Name: **Filipe Dias** (@fsdias) Original Redmine Issue: [11635](https://issues.qgis.org/issues/11635) Redmine category:processing/gdal Assignee: Giovanni Manghi --- ogr2ogr can be used to merge several shapefiles. It would be useful to have an ogr2ogr based tool that merges several shapefiles stored in a folder. Here's an example code written by Darren Cope: mkdir merged for %f in (*.shp) do ( if not exist merged\\merged.shp ( ogr2ogr -f “esri shapefile” merged\\merged.shp %f) else ( ogr2ogr -f “esri shapefile” -update -append merged\\merged.shp %f -nln Merged ) ) Source: http://darrencope.com/2010/05/07/merge-a-directory-of-shapefiles-using-ogr/
process
to merge several shapefiles in a folder author name filipe dias fsdias original redmine issue redmine category processing gdal assignee giovanni manghi can be used to merge several shapefiles it would be useful to have an based tool that merges several shapefiles stored in a folder here s an example code written by darren cope mkdir merged for f in shp do if not exist merged merged shp f “esri shapefile” merged merged shp f else f “esri shapefile” update append merged merged shp f nln merged source
1
13,195
15,615,223,684
IssuesEvent
2021-03-19 18:51:24
googleapis/nodejs-monitoring-dashboards
https://api.github.com/repos/googleapis/nodejs-monitoring-dashboards
closed
Update protocol definitions to include mosaic layouts
api: monitoring type: process
**Is your feature request related to a problem? Please describe.** The protocol definition implemented in the library does not match the RPC protocol made public and documented by Google. The protocol implemented by the library is missing the MosaicLayout definition. Please update the library to implement the latest version of the protocol definition, as it is [documented](https://cloud.google.com/monitoring/api/ref_v3/rpc/google.monitoring.dashboard.v1#google.monitoring.dashboard.v1.MosaicLayout) and [published](https://github.com/googleapis/googleapis/tree/master/google/monitoring/dashboard/v1) by Google. **Describe alternatives you've considered** 1. Using the REST API (prior to [googleapis/googleapis](https://github.com/googleapis/googleapis) having been [updated](https://github.com/googleapis/googleapis/commit/0dafa3963ef6fcb8a7f5daaa4bec12adb04de518)). 2. Generating an RPC client from the [googleapis](https://github.com/googleapis/googleapis) [monitoring dashboards protocol definitions](https://github.com/googleapis/googleapis/tree/master/google/monitoring/dashboard/v1).
1.0
Update protocol definitions to include mosaic layouts - **Is your feature request related to a problem? Please describe.** The protocol definition implemented in the library does not match the RPC protocol made public and documented by Google. The protocol implemented by the library is missing the MosaicLayout definition. Please update the library to implement the latest version of the protocol definition, as it is [documented](https://cloud.google.com/monitoring/api/ref_v3/rpc/google.monitoring.dashboard.v1#google.monitoring.dashboard.v1.MosaicLayout) and [published](https://github.com/googleapis/googleapis/tree/master/google/monitoring/dashboard/v1) by Google. **Describe alternatives you've considered** 1. Using the REST API (prior to [googleapis/googleapis](https://github.com/googleapis/googleapis) having been [updated](https://github.com/googleapis/googleapis/commit/0dafa3963ef6fcb8a7f5daaa4bec12adb04de518)). 2. Generating an RPC client from the [googleapis](https://github.com/googleapis/googleapis) [monitoring dashboards protocol definitions](https://github.com/googleapis/googleapis/tree/master/google/monitoring/dashboard/v1).
process
update protocol definitions to include mosaic layouts is your feature request related to a problem please describe the protocol definition implemented in the library does not match the rpc protocol made public and documented by google the protocol implemented by the library is missing the mosaiclayout definition please update the library to implement the latest version of the protocol definition as it is and by google describe alternatives you ve considered using the rest api prior to having been generating an rpc client from the
1
127,961
27,161,418,909
IssuesEvent
2023-02-17 12:11:52
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: existing mysql queries fail with MySqlConnectionException
Bug Low MySQL BE Coders Pod Integrations Pod
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description Two users reported this issue today: [here](https://app.intercom.com/a/inbox/y10e7138/inbox/shared/all/conversation/164629100161495) and [here](https://app.intercom.com/a/inbox/y10e7138/inbox/shared/all/conversation/164629100145140) ![image (1)](https://user-images.githubusercontent.com/94514895/186403967-79acc0b7-91da-406d-8733-a2b846a1bdd2.png) [This](https://docs.appsmith.com/help-and-support/troubleshooting-guide/action-errors/mysql-plugin-errors) could be a clue to fix the issue, but the error should ideally not occur for existing running queries. ### Steps To Reproduce not sure, I was not able to reproduce this ### Public Sample App _No response_ ### Version Cloud
1.0
[Bug]: existing mysql queries fail with MySqlConnectionException - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description Two users reported this issue today: [here](https://app.intercom.com/a/inbox/y10e7138/inbox/shared/all/conversation/164629100161495) and [here](https://app.intercom.com/a/inbox/y10e7138/inbox/shared/all/conversation/164629100145140) ![image (1)](https://user-images.githubusercontent.com/94514895/186403967-79acc0b7-91da-406d-8733-a2b846a1bdd2.png) [This](https://docs.appsmith.com/help-and-support/troubleshooting-guide/action-errors/mysql-plugin-errors) could be a clue to fix the issue, but the error should ideally not occur for existing running queries. ### Steps To Reproduce not sure, I was not able to reproduce this ### Public Sample App _No response_ ### Version Cloud
non_process
existing mysql queries fail with mysqlconnectionexception is there an existing issue for this i have searched the existing issues description two users reported this issue today and could be a clue to fix the issue but the error should ideally not occur for existing running queries steps to reproduce not sure i was not able to reproduce this public sample app no response version cloud
0
1,277
3,812,210,168
IssuesEvent
2016-03-27 12:02:20
pwittchen/ReactiveBeacons
https://api.github.com/repos/pwittchen/ReactiveBeacons
closed
Release 0.5.0
release process
**Initial release notes**: - added support for BLE scanning API available on Android API 21 (Lollipop) mentioned in issue #43 - created two scanning strategies for BLE scanning on pre-Lollipop devices and Lollipop devices - updated body of `Observable<Beacon> observe()` method - added `PreLollipopScanStrategy` class, `LollipopScanStrategy` class and `ScanStrategy` interface - added `ScanCallbackAdapter` class used in `LollipopScanStrategy` - added `Observable<Beacon> observe(ScanStrategy scanStrategy)` method to library API, which allows to use one of the existing scan strategies or custom scan strategy - added note about Compatibility with different Android versions in `REAMDE.md` - bumped RxJava version to `1.1.2` - bumped Kotlin version in sample app to `1.0.1-1` - changed `Schedulers.io()` in `subscribeOn(...)` to `Schedulers.computation()` in sample apps **Things to do**: - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update JavaDoc on gh-pages - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
1.0
Release 0.5.0 - **Initial release notes**: - added support for BLE scanning API available on Android API 21 (Lollipop) mentioned in issue #43 - created two scanning strategies for BLE scanning on pre-Lollipop devices and Lollipop devices - updated body of `Observable<Beacon> observe()` method - added `PreLollipopScanStrategy` class, `LollipopScanStrategy` class and `ScanStrategy` interface - added `ScanCallbackAdapter` class used in `LollipopScanStrategy` - added `Observable<Beacon> observe(ScanStrategy scanStrategy)` method to library API, which allows to use one of the existing scan strategies or custom scan strategy - added note about Compatibility with different Android versions in `REAMDE.md` - bumped RxJava version to `1.1.2` - bumped Kotlin version in sample app to `1.0.1-1` - changed `Schedulers.io()` in `subscribeOn(...)` to `Schedulers.computation()` in sample apps **Things to do**: - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update JavaDoc on gh-pages - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
process
release initial release notes added support for ble scanning api available on android api lollipop mentioned in issue created two scanning strategies for ble scanning on pre lollipop devices and lollipop devices updated body of observable observe method added prelollipopscanstrategy class lollipopscanstrategy class and scanstrategy interface added scancallbackadapter class used in lollipopscanstrategy added observable observe scanstrategy scanstrategy method to library api which allows to use one of the existing scan strategies or custom scan strategy added note about compatibility with different android versions in reamde md bumped rxjava version to bumped kotlin version in sample app to changed schedulers io in subscribeon to schedulers computation in sample apps things to do bump library version upload archives to maven central close and release artifact on maven central update javadoc on gh pages update changelog md after maven sync bump library version in readme md create new github release
1
1,150
3,633,798,281
IssuesEvent
2016-02-11 15:50:37
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Lost input from child processes due to initially-flowing net.Socket.
child_process
When opening child processes with child_process.spawn using 'pipe' for its stdout or stderr, the resulting stream is already in the flowing state when it's returned, and some data can be lost when consuming the stream using the 'readable' event + read(). A minimal example: ```js const spawn = require('child_process').spawn; const ls = spawn('echo', ['123']); console.log("flowing:", ls.stdout._readableState.flowing); ls.stdout.on('readable', function() { console.log("read:", ls.stdout.read()); }); ``` On recent node versions, no data is read; the 'readable' event only fires once, for end-of-stream. On older versions, the first chunk of data is read. I'm not sure which exact version introduced the bug, but: ```bash % node --version && node /tmp/test.js v0.10.36 flowing: false read: null read: \<Buffer 31 32 33 0a\> read: null % node --version && node /tmp/test.js v5.1.0 flowing: null read: null % node --version && node /tmp/test.js v5.5.0 flowing: null read: null ``` Potentially related to #445. Seems to be the root cause of a few issues affecting GHCJS: - https://github.com/ghcjs/ghcjs/issues/453 - https://github.com/ghcjs/ghcjs/issues/436 - https://github.com/luigy/try-stack-reflex/issues/5 - https://github.com/commercialhaskell/stack/issues/1496 - https://github.com/commercialhaskell/stack/issues/1437 - https://github.com/commercialhaskell/stack/issues/1629 The interaction between https://github.com/nodejs/node/blob/master/lib/internal/child_process.js#L228 and https://github.com/nodejs/node/blob/master/lib/net.js#L165 looks suspicious to me, but I don't know enough about nodejs' internals to suggest an appropriate fix.
1.0
Lost input from child processes due to initially-flowing net.Socket. - When opening child processes with child_process.spawn using 'pipe' for its stdout or stderr, the resulting stream is already in the flowing state when it's returned, and some data can be lost when consuming the stream using the 'readable' event + read(). A minimal example: ```js const spawn = require('child_process').spawn; const ls = spawn('echo', ['123']); console.log("flowing:", ls.stdout._readableState.flowing); ls.stdout.on('readable', function() { console.log("read:", ls.stdout.read()); }); ``` On recent node versions, no data is read; the 'readable' event only fires once, for end-of-stream. On older versions, the first chunk of data is read. I'm not sure which exact version introduced the bug, but: ```bash % node --version && node /tmp/test.js v0.10.36 flowing: false read: null read: \<Buffer 31 32 33 0a\> read: null % node --version && node /tmp/test.js v5.1.0 flowing: null read: null % node --version && node /tmp/test.js v5.5.0 flowing: null read: null ``` Potentially related to #445. Seems to be the root cause of a few issues affecting GHCJS: - https://github.com/ghcjs/ghcjs/issues/453 - https://github.com/ghcjs/ghcjs/issues/436 - https://github.com/luigy/try-stack-reflex/issues/5 - https://github.com/commercialhaskell/stack/issues/1496 - https://github.com/commercialhaskell/stack/issues/1437 - https://github.com/commercialhaskell/stack/issues/1629 The interaction between https://github.com/nodejs/node/blob/master/lib/internal/child_process.js#L228 and https://github.com/nodejs/node/blob/master/lib/net.js#L165 looks suspicious to me, but I don't know enough about nodejs' internals to suggest an appropriate fix.
process
lost input from child processes due to initially flowing net socket when opening child processes with child process spawn using pipe for its stdout or stderr the resulting stream is already in the flowing state when it s returned and some data can be lost when consuming the stream using the readable event read a minimal example js const spawn require child process spawn const ls spawn echo console log flowing ls stdout readablestate flowing ls stdout on readable function console log read ls stdout read on recent node versions no data is read the readable event only fires once for end of stream on older versions the first chunk of data is read i m not sure which exact version introduced the bug but bash node version node tmp test js flowing false read null read read null node version node tmp test js flowing null read null node version node tmp test js flowing null read null potentially related to seems to be the root cause of a few issues affecting ghcjs the interaction between and looks suspicious to me but i don t know enough about nodejs internals to suggest an appropriate fix
1
102,194
21,930,075,752
IssuesEvent
2022-05-23 08:58:45
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
Submarine equality check failed when host starts a campaign before any client has joined
Bug Code Networking High prio
### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? If a host starts campaign before any client has joined the server, any client who tries to join the already running campaign will get a submarine equality check failed error and are unable to join. Host has to return back to lobby in order to get the clients to join, once there's at least one client playing on the campaign all new clients are able to join without any issues. ### Reproduction steps 1. Host a SteamP2P server 2. Start campaign round 3. Ask someone to join your server 4. They will receive an error once they try to join the already running campaign Not sure if this is possible to do with multiple clients on the same machine ### Bug prevalence Happens every time I play ### Version 0.18.1.0 (Unstable) ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports ```shell There was an error initializing the round. {Submarine equality check failed. The submarine loaded at your end doesn't match the one loaded by the server. There may have been an error in receiving the up-to-date submarine file from the server.} at Barotrauma.Networking.GameClient.ReadStartGameFinalize(IReadMessage inc) in <DEV>\Barotrauma\BarotraumaClient\ClientSource\Networking\GameClient.cs:line 1023 at Barotrauma.Networking.GameClient.StartGame(IReadMessage inc)+MoveNext() in <DEV>\Barotrauma\BarotraumaClient\ClientSource\Networking\GameClient.cs:line 1743 Coroutine "Barotrauma.Networking.GameClient+<StartGame>d__99" has failed ```
1.0
Submarine equality check failed when host starts a campaign before any client has joined - ### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? If a host starts campaign before any client has joined the server, any client who tries to join the already running campaign will get a submarine equality check failed error and are unable to join. Host has to return back to lobby in order to get the clients to join, once there's at least one client playing on the campaign all new clients are able to join without any issues. ### Reproduction steps 1. Host a SteamP2P server 2. Start campaign round 3. Ask someone to join your server 4. They will receive an error once they try to join the already running campaign Not sure if this is possible to do with multiple clients on the same machine ### Bug prevalence Happens every time I play ### Version 0.18.1.0 (Unstable) ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports ```shell There was an error initializing the round. {Submarine equality check failed. The submarine loaded at your end doesn't match the one loaded by the server. There may have been an error in receiving the up-to-date submarine file from the server.} at Barotrauma.Networking.GameClient.ReadStartGameFinalize(IReadMessage inc) in <DEV>\Barotrauma\BarotraumaClient\ClientSource\Networking\GameClient.cs:line 1023 at Barotrauma.Networking.GameClient.StartGame(IReadMessage inc)+MoveNext() in <DEV>\Barotrauma\BarotraumaClient\ClientSource\Networking\GameClient.cs:line 1743 Coroutine "Barotrauma.Networking.GameClient+<StartGame>d__99" has failed ```
non_process
submarine equality check failed when host starts a campaign before any client has joined disclaimers i have searched the issue tracker to check if the issue has already been reported my issue happened while using mods what happened if a host starts campaign before any client has joined the server any client who tries to join the already running campaign will get a submarine equality check failed error and are unable to join host has to return back to lobby in order to get the clients to join once there s at least one client playing on the campaign all new clients are able to join without any issues reproduction steps host a server start campaign round ask someone to join your server they will receive an error once they try to join the already running campaign not sure if this is possible to do with multiple clients on the same machine bug prevalence happens every time i play version unstable no response which operating system did you encounter this bug on windows relevant error messages and crash reports shell there was an error initializing the round submarine equality check failed the submarine loaded at your end doesn t match the one loaded by the server there may have been an error in receiving the up to date submarine file from the server at barotrauma networking gameclient readstartgamefinalize ireadmessage inc in barotrauma barotraumaclient clientsource networking gameclient cs line at barotrauma networking gameclient startgame ireadmessage inc movenext in barotrauma barotraumaclient clientsource networking gameclient cs line coroutine barotrauma networking gameclient d has failed
0
46,329
9,923,895,892
IssuesEvent
2019-07-01 08:25:43
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
Remove "mofo-style" from package.json
code cleanup engineering
Seems like we no longer uses `mofo-style`. Let's remove it from `package.json`.
1.0
Remove "mofo-style" from package.json - Seems like we no longer uses `mofo-style`. Let's remove it from `package.json`.
non_process
remove mofo style from package json seems like we no longer uses mofo style let s remove it from package json
0
241,470
26,256,803,579
IssuesEvent
2023-01-06 01:59:26
Techini/WebGoat
https://api.github.com/repos/Techini/WebGoat
reopened
CVE-2021-32640 (Medium) detected in ws-6.1.2.tgz
security vulnerability
## CVE-2021-32640 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ws-6.1.2.tgz</b></p></summary> <p>Simple to use, blazing fast and thoroughly tested websocket client and server for Node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/ws/-/ws-6.1.2.tgz">https://registry.npmjs.org/ws/-/ws-6.1.2.tgz</a></p> <p>Path to dependency file: /docs/package.json</p> <p>Path to vulnerable library: /docs/node_modules/ws/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - browser-sync-ui-2.26.2.tgz - socket.io-client-2.2.0.tgz - engine.io-client-3.3.1.tgz - :x: **ws-6.1.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Techini/WebGoat/commit/d33cc0e32a0d1b949ff1b85af16890cd452276f8">d33cc0e32a0d1b949ff1b85af16890cd452276f8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ws is an open source WebSocket client and server library for Node.js. A specially crafted value of the `Sec-Websocket-Protocol` header can be used to significantly slow down a ws server. The vulnerability has been fixed in ws@7.4.6 (https://github.com/websockets/ws/commit/00c425ec77993773d823f018f64a5c44e17023ff). In vulnerable versions of ws, the issue can be mitigated by reducing the maximum allowed length of the request headers using the [`--max-http-header-size=size`](https://nodejs.org/api/cli.html#cli_max_http_header_size_size) and/or the [`maxHeaderSize`](https://nodejs.org/api/http.html#http_http_createserver_options_requestlistener) options. <p>Publish Date: 2021-05-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32640>CVE-2021-32640</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/websockets/ws/security/advisories/GHSA-6fc8-4gx4-v693">https://github.com/websockets/ws/security/advisories/GHSA-6fc8-4gx4-v693</a></p> <p>Release Date: 2021-05-25</p> <p>Fix Resolution (ws): 6.2.2</p> <p>Direct dependency fix Resolution (browser-sync): 2.26.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-32640 (Medium) detected in ws-6.1.2.tgz - ## CVE-2021-32640 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ws-6.1.2.tgz</b></p></summary> <p>Simple to use, blazing fast and thoroughly tested websocket client and server for Node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/ws/-/ws-6.1.2.tgz">https://registry.npmjs.org/ws/-/ws-6.1.2.tgz</a></p> <p>Path to dependency file: /docs/package.json</p> <p>Path to vulnerable library: /docs/node_modules/ws/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - browser-sync-ui-2.26.2.tgz - socket.io-client-2.2.0.tgz - engine.io-client-3.3.1.tgz - :x: **ws-6.1.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Techini/WebGoat/commit/d33cc0e32a0d1b949ff1b85af16890cd452276f8">d33cc0e32a0d1b949ff1b85af16890cd452276f8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ws is an open source WebSocket client and server library for Node.js. A specially crafted value of the `Sec-Websocket-Protocol` header can be used to significantly slow down a ws server. The vulnerability has been fixed in ws@7.4.6 (https://github.com/websockets/ws/commit/00c425ec77993773d823f018f64a5c44e17023ff). In vulnerable versions of ws, the issue can be mitigated by reducing the maximum allowed length of the request headers using the [`--max-http-header-size=size`](https://nodejs.org/api/cli.html#cli_max_http_header_size_size) and/or the [`maxHeaderSize`](https://nodejs.org/api/http.html#http_http_createserver_options_requestlistener) options. <p>Publish Date: 2021-05-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32640>CVE-2021-32640</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/websockets/ws/security/advisories/GHSA-6fc8-4gx4-v693">https://github.com/websockets/ws/security/advisories/GHSA-6fc8-4gx4-v693</a></p> <p>Release Date: 2021-05-25</p> <p>Fix Resolution (ws): 6.2.2</p> <p>Direct dependency fix Resolution (browser-sync): 2.26.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in ws tgz cve medium severity vulnerability vulnerable library ws tgz simple to use blazing fast and thoroughly tested websocket client and server for node js library home page a href path to dependency file docs package json path to vulnerable library docs node modules ws package json dependency hierarchy browser sync tgz root library browser sync ui tgz socket io client tgz engine io client tgz x ws tgz vulnerable library found in head commit a href vulnerability details ws is an open source websocket client and server library for node js a specially crafted value of the sec websocket protocol header can be used to significantly slow down a ws server the vulnerability has been fixed in ws in vulnerable versions of ws the issue can be mitigated by reducing the maximum allowed length of the request headers using the and or the options publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ws direct dependency fix resolution browser sync step up your open source security game with whitesource
0
191,661
15,300,489,030
IssuesEvent
2021-02-24 12:23:13
EnterpriseDB/docs
https://api.github.com/repos/EnterpriseDB/docs
closed
EPAS Guide - Duplicate Content in ToC
documentation
As discussed, duplicate content is encountered in Advanced Server guide, refer snippet to remove the content from guide. <img width="285" alt="Screenshot 2021-02-04 at 5 47 02 PM" src="https://user-images.githubusercontent.com/62722976/106912631-18925600-66fb-11eb-983f-985413235931.png"> One more observation: Index advisor content is again duplicated and appearing twice in the toc, remove the config parameters and index advisor content as shown in the attached image. <img width="285" alt="Screenshot 2021-02-04 at 5 47 02 PM" src="https://user-images.githubusercontent.com/62722976/106912734-35c72480-66fb-11eb-9620-d8dcfc4eb13c.png"> -- Thanks to @sheet20
1.0
EPAS Guide - Duplicate Content in ToC - As discussed, duplicate content is encountered in Advanced Server guide, refer snippet to remove the content from guide. <img width="285" alt="Screenshot 2021-02-04 at 5 47 02 PM" src="https://user-images.githubusercontent.com/62722976/106912631-18925600-66fb-11eb-983f-985413235931.png"> One more observation: Index advisor content is again duplicated and appearing twice in the toc, remove the config parameters and index advisor content as shown in the attached image. <img width="285" alt="Screenshot 2021-02-04 at 5 47 02 PM" src="https://user-images.githubusercontent.com/62722976/106912734-35c72480-66fb-11eb-9620-d8dcfc4eb13c.png"> -- Thanks to @sheet20
non_process
epas guide duplicate content in toc as discussed duplicate content is encountered in advanced server guide refer snippet to remove the content from guide img width alt screenshot at pm src one more observation index advisor content is again duplicated and appearing twice in the toc remove the config parameters and index advisor content as shown in the attached image img width alt screenshot at pm src thanks to
0
45,281
5,711,915,408
IssuesEvent
2017-04-19 01:05:00
medic/medic-webapp
https://api.github.com/repos/medic/medic-webapp
closed
The user needs an associated contact to create a contact
4 - Acceptance Testing Bug Regression
A recent refactor caused a regression whereby the logged in user must have an associated contact in order to create a contact. This is how filing reports works, so the resultant report can be assigned to the correct person. However there is no need for contacts to be assigned to a person so this is not required. Furthermore it creates a chicken and egg problem where the first user (root admin) cannot create a person because they don't have a contact person assigned.
1.0
The user needs an associated contact to create a contact - A recent refactor caused a regression whereby the logged in user must have an associated contact in order to create a contact. This is how filing reports works, so the resultant report can be assigned to the correct person. However there is no need for contacts to be assigned to a person so this is not required. Furthermore it creates a chicken and egg problem where the first user (root admin) cannot create a person because they don't have a contact person assigned.
non_process
the user needs an associated contact to create a contact a recent refactor caused a regression whereby the logged in user must have an associated contact in order to create a contact this is how filing reports works so the resultant report can be assigned to the correct person however there is no need for contacts to be assigned to a person so this is not required furthermore it creates a chicken and egg problem where the first user root admin cannot create a person because they don t have a contact person assigned
0
6,453
9,546,534,604
IssuesEvent
2019-05-01 20:14:00
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Department of State - Next Steps Page
Apply Process Approved Requirements Ready State Dept.
Who: Student Applicant What: Next Steps page with a list of things they will need to do to apply Why: As a student applicant I would like to know what to expect before I start my application so I can be prepared. A/C - This will be the first page presented to **eligible applicants** when they click "Apply" - There will be a header "Next Steps" at the top with an explanation - There will be a smaller header "Complete your profile" with an explanation - There will be a numbered list - Select internship opportunities - with explanation - Review your experience and references - with explanation - Review your education and transcript - with an explanation - Review your languages and skills - with explanation - Include a statement of interest - with explanation - There will be a "Continue" button at the bottom of the page that will take the user to the "Select Internship Opportunities" Page InVision Link: https://opm.invisionapp.com/d/main/#/console/15360465/319289299/preview
1.0
Department of State - Next Steps Page - Who: Student Applicant What: Next Steps page with a list of things they will need to do to apply Why: As a student applicant I would like to know what to expect before I start my application so I can be prepared. A/C - This will be the first page presented to **eligible applicants** when they click "Apply" - There will be a header "Next Steps" at the top with an explanation - There will be a smaller header "Complete your profile" with an explanation - There will be a numbered list - Select internship opportunities - with explanation - Review your experience and references - with explanation - Review your education and transcript - with an explanation - Review your languages and skills - with explanation - Include a statement of interest - with explanation - There will be a "Continue" button at the bottom of the page that will take the user to the "Select Internship Opportunities" Page InVision Link: https://opm.invisionapp.com/d/main/#/console/15360465/319289299/preview
process
department of state next steps page who student applicant what next steps page with a list of things they will need to do to apply why as a student applicant i would like to know what to expect before i start my application so i can be prepared a c this will be the first page presented to eligible applicants when they click apply there will be a header next steps at the top with an explanation there will be a smaller header complete your profile with an explanation there will be a numbered list select internship opportunities with explanation review your experience and references with explanation review your education and transcript with an explanation review your languages and skills with explanation include a statement of interest with explanation there will be a continue button at the bottom of the page that will take the user to the select internship opportunities page invision link
1
87,461
8,075,210,133
IssuesEvent
2018-08-07 04:09:53
okTurtles/group-income-simple
https://api.github.com/repos/okTurtles/group-income-simple
closed
Add frontend tests for verifying that GroupSetGroupProfile action works
App:Frontend Kind:Bug Kind:Core Kind:Test Priority:High
### Problem This functionality seem to be tested at all. We only test modifying the global identity profile that's in `contracts/identity.js`. ### Solution Add frontend tests to test that modifying your group profile actually works.
1.0
Add frontend tests for verifying that GroupSetGroupProfile action works - ### Problem This functionality seem to be tested at all. We only test modifying the global identity profile that's in `contracts/identity.js`. ### Solution Add frontend tests to test that modifying your group profile actually works.
non_process
add frontend tests for verifying that groupsetgroupprofile action works problem this functionality seem to be tested at all we only test modifying the global identity profile that s in contracts identity js solution add frontend tests to test that modifying your group profile actually works
0
11,146
13,957,692,871
IssuesEvent
2020-10-24 08:10:49
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
NL: harvesting request
Geoportal Harvesting process NL - The Netherlands
Dear Angelo, I am aware of the still ongoing security problems and know that the harvesting proces is not a validator, but in NL the pressure to harvest is growing day by day. We gave our dataproviders a deadline of march 15 to make their data available through view and download services. Dataproviders want to know if they are now working in the right direction and need to see some prove of their work. Can you give any indication when harvesting will be possible again. Will JRC start the harvesting proces, or will the proposed change to memberstate harvesting controlled become available. Who will be informed about the change? are that the persons wich ahave now access to the INSPIRE registry for discovery services? Best regards Ine
1.0
NL: harvesting request - Dear Angelo, I am aware of the still ongoing security problems and know that the harvesting proces is not a validator, but in NL the pressure to harvest is growing day by day. We gave our dataproviders a deadline of march 15 to make their data available through view and download services. Dataproviders want to know if they are now working in the right direction and need to see some prove of their work. Can you give any indication when harvesting will be possible again. Will JRC start the harvesting proces, or will the proposed change to memberstate harvesting controlled become available. Who will be informed about the change? are that the persons wich ahave now access to the INSPIRE registry for discovery services? Best regards Ine
process
nl harvesting request dear angelo i am aware of the still ongoing security problems and know that the harvesting proces is not a validator but in nl the pressure to harvest is growing day by day we gave our dataproviders a deadline of march to make their data available through view and download services dataproviders want to know if they are now working in the right direction and need to see some prove of their work can you give any indication when harvesting will be possible again will jrc start the harvesting proces or will the proposed change to memberstate harvesting controlled become available who will be informed about the change are that the persons wich ahave now access to the inspire registry for discovery services best regards ine
1
6,024
8,823,797,884
IssuesEvent
2019-01-02 14:56:20
docker/docker.github.io
https://api.github.com/repos/docker/docker.github.io
closed
Compose File docs do not provide Windows examples
area/compose platform/win process/top25
File: [compose/compose-file/index.md](https://docs.docker.com/compose/compose-file/), CC @gbarr01 Many things like volumes have different syntax for volumes. We should provide examples for Windows for compose. /cc @friism @brandonroyal
1.0
Compose File docs do not provide Windows examples - File: [compose/compose-file/index.md](https://docs.docker.com/compose/compose-file/), CC @gbarr01 Many things like volumes have different syntax for volumes. We should provide examples for Windows for compose. /cc @friism @brandonroyal
process
compose file docs do not provide windows examples file cc many things like volumes have different syntax for volumes we should provide examples for windows for compose cc friism brandonroyal
1
24,432
7,497,221,429
IssuesEvent
2018-04-08 17:35:06
scylladb/scylla
https://api.github.com/repos/scylladb/scylla
reopened
Traditional "make install"
Build
Traditionally, installation is done from `make install`, and the packaging scripts pick up the files from the installed location. This allows installs without going through the packaging process. This will help installations from source on distributions that don't use .rpm/.deb, and will eliminate some duplication.
1.0
Traditional "make install" - Traditionally, installation is done from `make install`, and the packaging scripts pick up the files from the installed location. This allows installs without going through the packaging process. This will help installations from source on distributions that don't use .rpm/.deb, and will eliminate some duplication.
non_process
traditional make install traditionally installation is done from make install and the packaging scripts pick up the files from the installed location this allows installs without going through the packaging process this will help installations from source on distributions that don t use rpm deb and will eliminate some duplication
0
63,860
15,726,854,261
IssuesEvent
2021-03-29 11:54:31
rticommunity/rticonnextdds-examples
https://api.github.com/repos/rticommunity/rticonnextdds-examples
closed
Port flat_data_latency C++11 example to CMake
build documentation enhancement good first issue style
The example `/examples/connext_dds/flat_data_latency/c++11` is missing the new `README.md` template and the `CMakeLists.txt`.
1.0
Port flat_data_latency C++11 example to CMake - The example `/examples/connext_dds/flat_data_latency/c++11` is missing the new `README.md` template and the `CMakeLists.txt`.
non_process
port flat data latency c example to cmake the example examples connext dds flat data latency c is missing the new readme md template and the cmakelists txt
0
8,524
5,794,651,643
IssuesEvent
2017-05-02 15:21:24
coala/coala-bears
https://api.github.com/repos/coala/coala-bears
closed
KeywordBear in 0.11 RC doesn't allow unsetting 'keywords' setting anymore
area/usability importance/critical needs backport type/bug
Try this `.coafile`: ``` [keywords] files = .coafile language = coafile bears = KeywordBear regex_keywords = some_(useful|weird)_regex_pattern # next line unsets the default 'FIXME' value keywords = ``` Instead of yielding 0 results as expected, it will yield 907; one for each occurrence of the empty string. Pinging @Mixih since it's a release blocker I assume.
True
KeywordBear in 0.11 RC doesn't allow unsetting 'keywords' setting anymore - Try this `.coafile`: ``` [keywords] files = .coafile language = coafile bears = KeywordBear regex_keywords = some_(useful|weird)_regex_pattern # next line unsets the default 'FIXME' value keywords = ``` Instead of yielding 0 results as expected, it will yield 907; one for each occurrence of the empty string. Pinging @Mixih since it's a release blocker I assume.
non_process
keywordbear in rc doesn t allow unsetting keywords setting anymore try this coafile files coafile language coafile bears keywordbear regex keywords some useful weird regex pattern next line unsets the default fixme value keywords instead of yielding results as expected it will yield one for each occurrence of the empty string pinging mixih since it s a release blocker i assume
0
16,302
20,959,747,320
IssuesEvent
2022-03-27 16:14:14
googleapis/python-dialogflow-cx
https://api.github.com/repos/googleapis/python-dialogflow-cx
closed
Resolve sphinx docs warnings and enable warnings as errors
type: process api: dialogflow
In the python client libraries template for [noxfile.py](https://github.com/googleapis/synthtool/blob/master/synthtool/gcp/templates/python_library/noxfile.py.j2#L231), we enable warnings as errors in the nox session that generates the docs. There is [code in owlbot.py](https://github.com/googleapis/python-dialogflow-cx/blob/master/owlbot.py#L48) that reverts this logic so that warnings are not treated as errors when running `nox -s docs`. This workaround should be removed once the warnings are resolved.
1.0
Resolve sphinx docs warnings and enable warnings as errors - In the python client libraries template for [noxfile.py](https://github.com/googleapis/synthtool/blob/master/synthtool/gcp/templates/python_library/noxfile.py.j2#L231), we enable warnings as errors in the nox session that generates the docs. There is [code in owlbot.py](https://github.com/googleapis/python-dialogflow-cx/blob/master/owlbot.py#L48) that reverts this logic so that warnings are not treated as errors when running `nox -s docs`. This workaround should be removed once the warnings are resolved.
process
resolve sphinx docs warnings and enable warnings as errors in the python client libraries template for we enable warnings as errors in the nox session that generates the docs there is that reverts this logic so that warnings are not treated as errors when running nox s docs this workaround should be removed once the warnings are resolved
1
734,457
25,350,016,255
IssuesEvent
2022-11-19 17:00:24
codymikol/git-down
https://api.github.com/repos/codymikol/git-down
closed
Pass full full FileDeltaNode context to GitDownExtensions
HIGH PRIORITY
It turns out we need some additional context when staging files
1.0
Pass full full FileDeltaNode context to GitDownExtensions - It turns out we need some additional context when staging files
non_process
pass full full filedeltanode context to gitdownextensions it turns out we need some additional context when staging files
0
10,942
13,752,639,517
IssuesEvent
2020-10-06 14:44:48
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Introduce stand-alone integration test package
kind/improvement process/candidate team/typescript
We have a bunch of testing logic we would like to share across packages. Some of the logic right now lives in `@prisma/cli`, some in `@prisma/client`. We should move it out to a separate packages, which lives in our workspace. That package will not be published to NPM, but just include integration tests and test helpers. We would then also move the integration tests, that are in the CLI package there. The ones living in the Client package probably as well.
1.0
Introduce stand-alone integration test package - We have a bunch of testing logic we would like to share across packages. Some of the logic right now lives in `@prisma/cli`, some in `@prisma/client`. We should move it out to a separate packages, which lives in our workspace. That package will not be published to NPM, but just include integration tests and test helpers. We would then also move the integration tests, that are in the CLI package there. The ones living in the Client package probably as well.
process
introduce stand alone integration test package we have a bunch of testing logic we would like to share across packages some of the logic right now lives in prisma cli some in prisma client we should move it out to a separate packages which lives in our workspace that package will not be published to npm but just include integration tests and test helpers we would then also move the integration tests that are in the cli package there the ones living in the client package probably as well
1
132,432
12,507,925,851
IssuesEvent
2020-06-02 14:49:22
LucHermitte/mu-template
https://api.github.com/repos/LucHermitte/mu-template
closed
Missing documentation on typical workflows
Documentation
I found this repository, looking for a way to simplify some boilerplate code generation in my project. Looking at the feature list, I *think* that this can do exactly what I need, but I can't quite figure out how. I've read the README, which mostly specifies what is available, and the documentation under `doc/`, which mostly specifies what moving parts are available, but I'm missing some actual examples of how you would use these templates (in particular snippets). Would you type the name of a snippet and then press a key combo? Or select a bit of text and then add a key combo? Would you enter the name of the snippet in a prompt? Also, there is mention of placeholders, which also sound powerful, but I can't quite figure out what they are intended to do and how. Note that this issue is not so much asking for help (I'll just experiment a bit and can probably figure things out), but a suggestion for adding some additional documentation.
1.0
Missing documentation on typical workflows - I found this repository, looking for a way to simplify some boilerplate code generation in my project. Looking at the feature list, I *think* that this can do exactly what I need, but I can't quite figure out how. I've read the README, which mostly specifies what is available, and the documentation under `doc/`, which mostly specifies what moving parts are available, but I'm missing some actual examples of how you would use these templates (in particular snippets). Would you type the name of a snippet and then press a key combo? Or select a bit of text and then add a key combo? Would you enter the name of the snippet in a prompt? Also, there is mention of placeholders, which also sound powerful, but I can't quite figure out what they are intended to do and how. Note that this issue is not so much asking for help (I'll just experiment a bit and can probably figure things out), but a suggestion for adding some additional documentation.
non_process
missing documentation on typical workflows i found this repository looking for a way to simplify some boilerplate code generation in my project looking at the feature list i think that this can do exactly what i need but i can t quite figure out how i ve read the readme which mostly specifies what is available and the documentation under doc which mostly specifies what moving parts are available but i m missing some actual examples of how you would use these templates in particular snippets would you type the name of a snippet and then press a key combo or select a bit of text and then add a key combo would you enter the name of the snippet in a prompt also there is mention of placeholders which also sound powerful but i can t quite figure out what they are intended to do and how note that this issue is not so much asking for help i ll just experiment a bit and can probably figure things out but a suggestion for adding some additional documentation
0
15,397
19,582,233,822
IssuesEvent
2022-01-04 23:16:50
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Release checklist 0.48
enhancement P1 process
### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] PRs and issues that are merged/closed have milestone populated - [x] PRs and issues targeting release are merged, closed, or re-targeted - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [x] Publish release ## Integration - [ ] Deploy to VM - [ ] Deploy Rosetta API - [ ] Rosetta tests ## Performance - [ ] Deploy to Kubernetes - [ ] Deploy to VM - [ ] gRPC API performance tests - [ ] Importer performance tests - [ ] REST API performance tests - [ ] Migrations tested against mainnet clone ## Previewnet - [ ] Deploy to VM - [ ] Deploy Rosetta API - [ ] Rosetta tests ## Testnet - [ ] Deploy to VM - [ ] Deploy Rosetta API - [ ] Rosetta tests ## Mainnet - [ ] Deploy to Kubernetes EU - [ ] Deploy to Kubernetes NA - [ ] Deploy to VM - [ ] Rosetta tests ### Alternatives _No response_
1.0
Release checklist 0.48 - ### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] PRs and issues that are merged/closed have milestone populated - [x] PRs and issues targeting release are merged, closed, or re-targeted - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [x] Publish release ## Integration - [ ] Deploy to VM - [ ] Deploy Rosetta API - [ ] Rosetta tests ## Performance - [ ] Deploy to Kubernetes - [ ] Deploy to VM - [ ] gRPC API performance tests - [ ] Importer performance tests - [ ] REST API performance tests - [ ] Migrations tested against mainnet clone ## Previewnet - [ ] Deploy to VM - [ ] Deploy Rosetta API - [ ] Rosetta tests ## Testnet - [ ] Deploy to VM - [ ] Deploy Rosetta API - [ ] Rosetta tests ## Mainnet - [ ] Deploy to Kubernetes EU - [ ] Deploy to Kubernetes NA - [ ] Deploy to VM - [ ] Rosetta tests ### Alternatives _No response_
process
release checklist problem we need a checklist to verify the release is rolled out successfully solution preparation prs and issues that are merged closed have milestone populated prs and issues targeting release are merged closed or re targeted github checks for branch are passing automated kubernetes deployment successful tag release upload release artifacts publish release integration deploy to vm deploy rosetta api rosetta tests performance deploy to kubernetes deploy to vm grpc api performance tests importer performance tests rest api performance tests migrations tested against mainnet clone previewnet deploy to vm deploy rosetta api rosetta tests testnet deploy to vm deploy rosetta api rosetta tests mainnet deploy to kubernetes eu deploy to kubernetes na deploy to vm rosetta tests alternatives no response
1
41,905
5,408,719,911
IssuesEvent
2017-03-01 01:04:57
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
[Webhooks: scaleHost] Webhook request to scale up cloning a disconnected host is stuck without any response
area/webhooks kind/bug status/resolved status/to-test
**Rancher Versions:** Server: Master **Docker Version:**1.13.1 **OS and where are the hosts located? (cloud, bare metal, etc):**Ubuntu, Cloud **Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)**Single node **Environment Type: (Cattle/Kubernetes/Swarm/Mesos)**Cattle **Steps to Reproduce:** 1. Create a 4 host setup with few hosts having label "test=value" 2. Create a webhook to scale up hosts by amount=1 3. Power off the oldest host with the label "test=value" and wait till it goes to disconnected state 4. Execute the webhook to scale up while the oldest host is in disconnected state. A new host gets added as per the request 5. Edit the local storage resource limit to a lower value in the host settings on the disconnected host 6. Execute the webhook again, the request is stuck without any response and the cattle logs have timeouts as below: ``` 2017-02-22 18:02:58,034 INFO [:] [] [] [] [ecutorService-6] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3583] on [218] : Timeout 2017-02-22 18:03:28,077 INFO [:] [] [] [] [ecutorService-9] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3553] on [226] : Timeout 2017-02-22 18:03:28,079 INFO [:] [] [] [] [cutorService-12] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3661] on [222] : Timeout 2017-02-22 18:03:28,080 INFO [:] [] [] [] [ecutorService-3] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3559] on [212] : Timeout 2017-02-22 18:03:28,083 INFO [:] [] [] [] [cutorService-10] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3580] on [227] : Timeout 2017-02-22 18:03:28,083 INFO [:] [] [] [] [cutorService-15] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3584] on [216] : Timeout 2017-02-22 18:03:28,086 INFO [:] [] [] [] [ecutorService-5] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3567] on [217] : Timeout 2017-02-22 18:03:28,086 INFO [:] [] [] [] [ecutorService-8] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3498] on [213] : Timeout ``` top output for the rancher server machine: ``` PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 22923 root 20 0 356256 169940 8972 S 98.7 4.5 1390:42 webhook-service 16173 root 20 0 4345488 703796 19156 S 0.7 18.6 30:59.10 java 15866 root 20 0 242176 15752 7276 S 0.3 0.4 2:34.79 docker-containe 22920 root 20 0 268168 16796 7212 S 0.3 0.4 7:15.95 websocket-proxy 1 root 20 0 38120 6224 4036 S 0.0 0.2 0:05.15 systemd 2 root 20 0 0 0 0 S 0.0 0.0 0:00.00 kthreadd 3 root 20 0 0 0 0 S 0.0 0.0 0:04.63 ksoftirqd/0 5 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kworker/0:0H 7 root 20 0 0 0 0 R 0.0 0.0 0:21.09 rcu_sched 8 root 20 0 0 0 0 S 0.0 0.0 0:00.00 rcu_bh 9 root rt 0 0 0 0 S 0.0 0.0 0:00.00 migration/0 10 root rt 0 0 0 0 S 0.0 0.0 0:00.97 watchdog/0 11 root 20 0 0 0 0 S 0.0 0.0 0:00.00 kdevtmpfs 12 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 netns 13 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 perf 14 root 20 0 0 0 0 S 0.0 0.0 0:00.13 khungtaskd 15 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 writeback 16 root 25 5 0 0 0 S 0.0 0.0 0:00.00 ksmd 17 root 39 19 0 0 0 S 0.0 0.0 0:01.75 khugepaged 18 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 crypto 19 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kintegrityd 20 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 bioset 21 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kblockd 22 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 ata_sff 23 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 md 24 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 devfreq_wq 28 root 20 0 0 0 0 S 0.0 0.0 0:00.43 kswapd0 29 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 vmstat ```
1.0
[Webhooks: scaleHost] Webhook request to scale up cloning a disconnected host is stuck without any response - **Rancher Versions:** Server: Master **Docker Version:**1.13.1 **OS and where are the hosts located? (cloud, bare metal, etc):**Ubuntu, Cloud **Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)**Single node **Environment Type: (Cattle/Kubernetes/Swarm/Mesos)**Cattle **Steps to Reproduce:** 1. Create a 4 host setup with few hosts having label "test=value" 2. Create a webhook to scale up hosts by amount=1 3. Power off the oldest host with the label "test=value" and wait till it goes to disconnected state 4. Execute the webhook to scale up while the oldest host is in disconnected state. A new host gets added as per the request 5. Edit the local storage resource limit to a lower value in the host settings on the disconnected host 6. Execute the webhook again, the request is stuck without any response and the cattle logs have timeouts as below: ``` 2017-02-22 18:02:58,034 INFO [:] [] [] [] [ecutorService-6] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3583] on [218] : Timeout 2017-02-22 18:03:28,077 INFO [:] [] [] [] [ecutorService-9] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3553] on [226] : Timeout 2017-02-22 18:03:28,079 INFO [:] [] [] [] [cutorService-12] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3661] on [222] : Timeout 2017-02-22 18:03:28,080 INFO [:] [] [] [] [ecutorService-3] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3559] on [212] : Timeout 2017-02-22 18:03:28,083 INFO [:] [] [] [] [cutorService-10] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3580] on [227] : Timeout 2017-02-22 18:03:28,083 INFO [:] [] [] [] [cutorService-15] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3584] on [216] : Timeout 2017-02-22 18:03:28,086 INFO [:] [] [] [] [ecutorService-5] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3567] on [217] : Timeout 2017-02-22 18:03:28,086 INFO [:] [] [] [] [ecutorService-8] [.e.s.i.ProcessInstanceDispatcherImpl] Timeout on process [volumestoragepoolmap.remove:3498] on [213] : Timeout ``` top output for the rancher server machine: ``` PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 22923 root 20 0 356256 169940 8972 S 98.7 4.5 1390:42 webhook-service 16173 root 20 0 4345488 703796 19156 S 0.7 18.6 30:59.10 java 15866 root 20 0 242176 15752 7276 S 0.3 0.4 2:34.79 docker-containe 22920 root 20 0 268168 16796 7212 S 0.3 0.4 7:15.95 websocket-proxy 1 root 20 0 38120 6224 4036 S 0.0 0.2 0:05.15 systemd 2 root 20 0 0 0 0 S 0.0 0.0 0:00.00 kthreadd 3 root 20 0 0 0 0 S 0.0 0.0 0:04.63 ksoftirqd/0 5 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kworker/0:0H 7 root 20 0 0 0 0 R 0.0 0.0 0:21.09 rcu_sched 8 root 20 0 0 0 0 S 0.0 0.0 0:00.00 rcu_bh 9 root rt 0 0 0 0 S 0.0 0.0 0:00.00 migration/0 10 root rt 0 0 0 0 S 0.0 0.0 0:00.97 watchdog/0 11 root 20 0 0 0 0 S 0.0 0.0 0:00.00 kdevtmpfs 12 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 netns 13 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 perf 14 root 20 0 0 0 0 S 0.0 0.0 0:00.13 khungtaskd 15 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 writeback 16 root 25 5 0 0 0 S 0.0 0.0 0:00.00 ksmd 17 root 39 19 0 0 0 S 0.0 0.0 0:01.75 khugepaged 18 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 crypto 19 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kintegrityd 20 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 bioset 21 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 kblockd 22 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 ata_sff 23 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 md 24 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 devfreq_wq 28 root 20 0 0 0 0 S 0.0 0.0 0:00.43 kswapd0 29 root 0 -20 0 0 0 S 0.0 0.0 0:00.00 vmstat ```
non_process
webhook request to scale up cloning a disconnected host is stuck without any response rancher versions server master docker version os and where are the hosts located cloud bare metal etc ubuntu cloud setup details single node rancher vs ha rancher internal db vs external db single node environment type cattle kubernetes swarm mesos cattle steps to reproduce create a host setup with few hosts having label test value create a webhook to scale up hosts by amount power off the oldest host with the label test value and wait till it goes to disconnected state execute the webhook to scale up while the oldest host is in disconnected state a new host gets added as per the request edit the local storage resource limit to a lower value in the host settings on the disconnected host execute the webhook again the request is stuck without any response and the cattle logs have timeouts as below info timeout on process on timeout info timeout on process on timeout info timeout on process on timeout info timeout on process on timeout info timeout on process on timeout info timeout on process on timeout info timeout on process on timeout info timeout on process on timeout top output for the rancher server machine pid user pr ni virt res shr s cpu mem time command root s webhook service root s java root s docker containe root s websocket proxy root s systemd root s kthreadd root s ksoftirqd root s kworker root r rcu sched root s rcu bh root rt s migration root rt s watchdog root s kdevtmpfs root s netns root s perf root s khungtaskd root s writeback root s ksmd root s khugepaged root s crypto root s kintegrityd root s bioset root s kblockd root s ata sff root s md root s devfreq wq root s root s vmstat
0
19,562
25,885,019,002
IssuesEvent
2022-12-14 13:59:51
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
metricKeyToDimensions cache is printing many stack traces
bug processor/spanmetrics
**Describe the bug** https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/6503 the additions of this LRU cache treats the missing key as an error, and prints a big stacktrace in the logs. **Steps to reproduce** Have either high dimensions or small `dimensions_cache_size` **What did you expect to see?** Either a metric for cache/hit ratio or a much lower log level, such as debug. **What did you see instead?** A lot of stacktraces ``` 2022-04-04T03:10:26.171Z error spanmetricsprocessor@v0.47.0/processor.go:330 value not found in metricKeyToDimensions cache by key "copernic\x00getUserInfo\x00SPAN_KIND_SERVER\x00STATUS_CODE_UNSET" {"kind": "processor", "name": "spanmetrics", "pipeline": "traces"} github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor.(*processorImp).collectCallMetrics github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor@v0.47.0/processor.go:330 github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor.(*processorImp).buildMetrics github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor@v0.47.0/processor.go:258 github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor.(*processorImp).ConsumeTraces github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor@v0.47.0/processor.go:232 go.opentelemetry.io/collector/processor/processorhelper.NewTracesProcessor.func1 go.opentelemetry.io/collector@v0.47.0/processor/processorhelper/traces.go:70 go.opentelemetry.io/collector/consumer.ConsumeTracesFunc.ConsumeTraces go.opentelemetry.io/collector@v0.47.0/consumer/traces.go:36 github.com/open-telemetry/opentelemetry-collector-contrib/receiver/jaegerreceiver.(*jReceiver).PostSpans github.com/open-telemetry/opentelemetry-collector-contrib/receiver/jaegerreceiver@v0.47.0/trace_receiver.go:290 github.com/jaegertracing/jaeger/proto-gen/api_v2._CollectorService_PostSpans_Handler.func1 github.com/jaegertracing/jaeger@v1.32.0/proto-gen/api_v2/collector.pb.go:208 go.opentelemetry.io/collector/config/configgrpc.enhanceWithClientInformation.func1 go.opentelemetry.io/collector@v0.47.0/config/configgrpc/configgrpc.go:382 google.golang.org/grpc.chainUnaryInterceptors.func1.1 google.golang.org/grpc@v1.45.0/server.go:1116 go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc.UnaryServerInterceptor.func1 go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc@v0.29.0/interceptor.go:325 google.golang.org/grpc.chainUnaryInterceptors.func1.1 google.golang.org/grpc@v1.45.0/server.go:1119 google.golang.org/grpc.chainUnaryInterceptors.func1 google.golang.org/grpc@v1.45.0/server.go:1121 github.com/jaegertracing/jaeger/proto-gen/api_v2._CollectorService_PostSpans_Handler github.com/jaegertracing/jaeger@v1.32.0/proto-gen/api_v2/collector.pb.go:210 google.golang.org/grpc.(*Server).processUnaryRPC google.golang.org/grpc@v1.45.0/server.go:1282 google.golang.org/grpc.(*Server).handleStream google.golang.org/grpc@v1.45.0/server.go:1619 google.golang.org/grpc.(*Server).serveStreams.func1.2 google.golang.org/grpc@v1.45.0/server.go:921 ``` **What version did you use?** Version: v0.48.0 **What config did you use?** Config: (e.g. the yaml config file) **Environment** OS: Debian 11 Compiler(if manually compiled): vanilla docker image **Additional context** :blue_heart:
1.0
metricKeyToDimensions cache is printing many stack traces - **Describe the bug** https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/6503 the additions of this LRU cache treats the missing key as an error, and prints a big stacktrace in the logs. **Steps to reproduce** Have either high dimensions or small `dimensions_cache_size` **What did you expect to see?** Either a metric for cache/hit ratio or a much lower log level, such as debug. **What did you see instead?** A lot of stacktraces ``` 2022-04-04T03:10:26.171Z error spanmetricsprocessor@v0.47.0/processor.go:330 value not found in metricKeyToDimensions cache by key "copernic\x00getUserInfo\x00SPAN_KIND_SERVER\x00STATUS_CODE_UNSET" {"kind": "processor", "name": "spanmetrics", "pipeline": "traces"} github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor.(*processorImp).collectCallMetrics github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor@v0.47.0/processor.go:330 github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor.(*processorImp).buildMetrics github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor@v0.47.0/processor.go:258 github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor.(*processorImp).ConsumeTraces github.com/open-telemetry/opentelemetry-collector-contrib/processor/spanmetricsprocessor@v0.47.0/processor.go:232 go.opentelemetry.io/collector/processor/processorhelper.NewTracesProcessor.func1 go.opentelemetry.io/collector@v0.47.0/processor/processorhelper/traces.go:70 go.opentelemetry.io/collector/consumer.ConsumeTracesFunc.ConsumeTraces go.opentelemetry.io/collector@v0.47.0/consumer/traces.go:36 github.com/open-telemetry/opentelemetry-collector-contrib/receiver/jaegerreceiver.(*jReceiver).PostSpans github.com/open-telemetry/opentelemetry-collector-contrib/receiver/jaegerreceiver@v0.47.0/trace_receiver.go:290 github.com/jaegertracing/jaeger/proto-gen/api_v2._CollectorService_PostSpans_Handler.func1 github.com/jaegertracing/jaeger@v1.32.0/proto-gen/api_v2/collector.pb.go:208 go.opentelemetry.io/collector/config/configgrpc.enhanceWithClientInformation.func1 go.opentelemetry.io/collector@v0.47.0/config/configgrpc/configgrpc.go:382 google.golang.org/grpc.chainUnaryInterceptors.func1.1 google.golang.org/grpc@v1.45.0/server.go:1116 go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc.UnaryServerInterceptor.func1 go.opentelemetry.io/contrib/instrumentation/google.golang.org/grpc/otelgrpc@v0.29.0/interceptor.go:325 google.golang.org/grpc.chainUnaryInterceptors.func1.1 google.golang.org/grpc@v1.45.0/server.go:1119 google.golang.org/grpc.chainUnaryInterceptors.func1 google.golang.org/grpc@v1.45.0/server.go:1121 github.com/jaegertracing/jaeger/proto-gen/api_v2._CollectorService_PostSpans_Handler github.com/jaegertracing/jaeger@v1.32.0/proto-gen/api_v2/collector.pb.go:210 google.golang.org/grpc.(*Server).processUnaryRPC google.golang.org/grpc@v1.45.0/server.go:1282 google.golang.org/grpc.(*Server).handleStream google.golang.org/grpc@v1.45.0/server.go:1619 google.golang.org/grpc.(*Server).serveStreams.func1.2 google.golang.org/grpc@v1.45.0/server.go:921 ``` **What version did you use?** Version: v0.48.0 **What config did you use?** Config: (e.g. the yaml config file) **Environment** OS: Debian 11 Compiler(if manually compiled): vanilla docker image **Additional context** :blue_heart:
process
metrickeytodimensions cache is printing many stack traces describe the bug the additions of this lru cache treats the missing key as an error and prints a big stacktrace in the logs steps to reproduce have either high dimensions or small dimensions cache size what did you expect to see either a metric for cache hit ratio or a much lower log level such as debug what did you see instead a lot of stacktraces error spanmetricsprocessor processor go value not found in metrickeytodimensions cache by key copernic kind server code unset kind processor name spanmetrics pipeline traces github com open telemetry opentelemetry collector contrib processor spanmetricsprocessor processorimp collectcallmetrics github com open telemetry opentelemetry collector contrib processor spanmetricsprocessor processor go github com open telemetry opentelemetry collector contrib processor spanmetricsprocessor processorimp buildmetrics github com open telemetry opentelemetry collector contrib processor spanmetricsprocessor processor go github com open telemetry opentelemetry collector contrib processor spanmetricsprocessor processorimp consumetraces github com open telemetry opentelemetry collector contrib processor spanmetricsprocessor processor go go opentelemetry io collector processor processorhelper newtracesprocessor go opentelemetry io collector processor processorhelper traces go go opentelemetry io collector consumer consumetracesfunc consumetraces go opentelemetry io collector consumer traces go github com open telemetry opentelemetry collector contrib receiver jaegerreceiver jreceiver postspans github com open telemetry opentelemetry collector contrib receiver jaegerreceiver trace receiver go github com jaegertracing jaeger proto gen api collectorservice postspans handler github com jaegertracing jaeger proto gen api collector pb go go opentelemetry io collector config configgrpc enhancewithclientinformation go opentelemetry io collector config configgrpc configgrpc go google golang org grpc chainunaryinterceptors google golang org grpc server go go opentelemetry io contrib instrumentation google golang org grpc otelgrpc unaryserverinterceptor go opentelemetry io contrib instrumentation google golang org grpc otelgrpc interceptor go google golang org grpc chainunaryinterceptors google golang org grpc server go google golang org grpc chainunaryinterceptors google golang org grpc server go github com jaegertracing jaeger proto gen api collectorservice postspans handler github com jaegertracing jaeger proto gen api collector pb go google golang org grpc server processunaryrpc google golang org grpc server go google golang org grpc server handlestream google golang org grpc server go google golang org grpc server servestreams google golang org grpc server go what version did you use version what config did you use config e g the yaml config file environment os debian compiler if manually compiled vanilla docker image additional context blue heart
1
793,558
28,001,761,455
IssuesEvent
2023-03-27 12:37:46
DO2-Momo/yarp
https://api.github.com/repos/DO2-Momo/yarp
closed
Desktop Configuration
Priority: MEDIUM
# Desktop Configuration When issue #7 is finished, we can start configuring the desktop environment Feel free to be multiple assignees on this issue, we can test multiple configurations in parrallel
1.0
Desktop Configuration - # Desktop Configuration When issue #7 is finished, we can start configuring the desktop environment Feel free to be multiple assignees on this issue, we can test multiple configurations in parrallel
non_process
desktop configuration desktop configuration when issue is finished we can start configuring the desktop environment feel free to be multiple assignees on this issue we can test multiple configurations in parrallel
0
7,734
10,854,394,491
IssuesEvent
2019-11-13 16:20:57
prisma/prisma-engine
https://api.github.com/repos/prisma/prisma-engine
opened
Extend Destructive Changes: Warn if default is removed from an underlying column
process/candidate
By looking at the Magento database we learned that there are database schemas that use a default value on columns with a foreign key. If a user would start to use lift on such a database the default values would get silently removed. Hence we want to incorporate this into our destructive changes checking: * The `migration-engine` must return a destructive warning if a column default gets removed.
1.0
Extend Destructive Changes: Warn if default is removed from an underlying column - By looking at the Magento database we learned that there are database schemas that use a default value on columns with a foreign key. If a user would start to use lift on such a database the default values would get silently removed. Hence we want to incorporate this into our destructive changes checking: * The `migration-engine` must return a destructive warning if a column default gets removed.
process
extend destructive changes warn if default is removed from an underlying column by looking at the magento database we learned that there are database schemas that use a default value on columns with a foreign key if a user would start to use lift on such a database the default values would get silently removed hence we want to incorporate this into our destructive changes checking the migration engine must return a destructive warning if a column default gets removed
1
18,759
24,661,401,590
IssuesEvent
2022-10-18 07:00:08
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
Binary encoded data in (cc) field fails mail processing
bug verified prioritised by payment mail processing
<!-- Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓 Since november 15th we handle all requests, except real bugs, at our community board. Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21 Please post: - Feature requests - Development questions - Technical questions on the board -> https://community.zammad.org ! If you think you hit a bug, please continue: - Search existing issues and the CHANGELOG.md for your issue - there might be a solution already - Make sure to use the latest version of Zammad if possible - Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it! - Please write the issue in english - Don't remove the template - otherwise we will close the issue without further comments - Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted). * The upper textblock will be removed automatically when you submit your issue * --> ### Infos: * Used Zammad version: 3.1 * Installation method (source, package, ..): any * Operating system: any * Database + version: any * Elasticsearch version: any * Browser + version: any * Ticket-ID: #1053678 ### Expected behavior: Zammad will decode TO, CC and FROM fields correctly and import the mail. ### Actual behavior: In some special cases, processing of mails fail if TO, CC or FROM contains binary data and move the mail to `unprocessible_mails`. ### Log ``` E, [2019-09-18T11:38:01.789617 #1531-70079947099620] ERROR -- : ERROR: Can't process email, you will find it for bug reporting under /home/666/tmp/unprocessable_mail/2019-09-18T09:38:01Z_9c47d46c14b70987029c0f7859243f70.eml, please create an issue at https://github.com/zammad/zammad/issues E, [2019-09-18T11:38:01.789646 #1531-70079947099620] ERROR -- : incompatible character encodings: UTF-8 and ASCII-8BIT (Encoding::CompatibilityError) /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/encodings.rb:137:in `join' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/encodings.rb:137:in `value_decode' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/encodings.rb:109:in `decode_encode' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/fields/unstructured_field.rb:84:in `do_decode' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/fields/unstructured_field.rb:62:in `decoded' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/field.rb:239:in `method_missing' /home/666/app/models/channel/email_parser.rb:514:in `rescue in block in message_header_hash' /home/666/app/models/channel/email_parser.rb:505:in `block in message_header_hash' /home/666/app/models/channel/email_parser.rb:504:in `map' /home/666/app/models/channel/email_parser.rb:504:in `message_header_hash' /home/666/app/models/channel/email_parser.rb:81:in `parse' /home/666/app/models/channel/email_parser.rb:136:in `_process' /home/666/app/models/channel/email_parser.rb:115:in `block in process' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:93:in `block in timeout' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:33:in `block in catch' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:33:in `catch' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:33:in `catch' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:108:in `timeout' /home/666/app/models/channel/email_parser.rb:114:in `process' /home/666/app/models/channel/driver/imap.rb:264:in `block in fetch' /home/666/app/models/channel/driver/imap.rb:212:in `each' /home/666/app/models/channel/driver/imap.rb:212:in `fetch' /home/666/app/models/channel.rb:56:in `fetch' /usr/local/rvm/gems/ruby-2.5.5/gems/activerecord-5.2.3/lib/active_record/relation/delegation.rb:71:in `each' /usr/local/rvm/gems/ruby-2.5.5/gems/activerecord-5.2.3/lib/active_record/relation/delegation.rb:71:in `each' /home/666/app/models/channel.rb:30:in `fetch' (eval):1:in `_start_job' /home/666/app/models/scheduler.rb:286:in `eval' /home/666/app/models/scheduler.rb:286:in `_start_job' /home/666/app/models/scheduler.rb:239:in `block (2 levels) in start_job' /home/666/app/models/scheduler.rb:237:in `loop' /home/666/app/models/scheduler.rb:237:in `block in start_job' /usr/local/rvm/gems/ruby-2.5.5/gems/logging-2.2.2/lib/logging/diagnostic_context.rb:474:in `block in create_with_logging_context' ``` ### Steps to reproduce the behavior: Put the following mail in a file: ``` Mime-Version: 1.0 X-Mailer: GroupWise 2014 R2 Subject: Broken mail Date: Thu, 12 Sep 2019 09:09:48 +0200 Message-ID: <5D79EF3C020000D2000A2654@example.de> From: "Lassen Sie sich =?UTF-8?B?VmVyd8O2aG5lbg==?=" <Verwöhnen@example.de> To: zammad@example.com Content-Type: multipart/alternative; boundary="____TAYQQEUZELYJMCBZPMES____" --____TAYQQEUZELYJMCBZPMES____ ``` Process it via: `cat /path/to/your.eml | time rails r 'Channel::Driver::MailStdin.new'` Yes I'm sure this is a bug and no feature request or a general question.
1.0
Binary encoded data in (cc) field fails mail processing - <!-- Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓 Since november 15th we handle all requests, except real bugs, at our community board. Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21 Please post: - Feature requests - Development questions - Technical questions on the board -> https://community.zammad.org ! If you think you hit a bug, please continue: - Search existing issues and the CHANGELOG.md for your issue - there might be a solution already - Make sure to use the latest version of Zammad if possible - Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it! - Please write the issue in english - Don't remove the template - otherwise we will close the issue without further comments - Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted). * The upper textblock will be removed automatically when you submit your issue * --> ### Infos: * Used Zammad version: 3.1 * Installation method (source, package, ..): any * Operating system: any * Database + version: any * Elasticsearch version: any * Browser + version: any * Ticket-ID: #1053678 ### Expected behavior: Zammad will decode TO, CC and FROM fields correctly and import the mail. ### Actual behavior: In some special cases, processing of mails fail if TO, CC or FROM contains binary data and move the mail to `unprocessible_mails`. ### Log ``` E, [2019-09-18T11:38:01.789617 #1531-70079947099620] ERROR -- : ERROR: Can't process email, you will find it for bug reporting under /home/666/tmp/unprocessable_mail/2019-09-18T09:38:01Z_9c47d46c14b70987029c0f7859243f70.eml, please create an issue at https://github.com/zammad/zammad/issues E, [2019-09-18T11:38:01.789646 #1531-70079947099620] ERROR -- : incompatible character encodings: UTF-8 and ASCII-8BIT (Encoding::CompatibilityError) /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/encodings.rb:137:in `join' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/encodings.rb:137:in `value_decode' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/encodings.rb:109:in `decode_encode' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/fields/unstructured_field.rb:84:in `do_decode' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/fields/unstructured_field.rb:62:in `decoded' /usr/local/rvm/gems/ruby-2.5.5/bundler/gems/mail-9265cf75bbe3/lib/mail/field.rb:239:in `method_missing' /home/666/app/models/channel/email_parser.rb:514:in `rescue in block in message_header_hash' /home/666/app/models/channel/email_parser.rb:505:in `block in message_header_hash' /home/666/app/models/channel/email_parser.rb:504:in `map' /home/666/app/models/channel/email_parser.rb:504:in `message_header_hash' /home/666/app/models/channel/email_parser.rb:81:in `parse' /home/666/app/models/channel/email_parser.rb:136:in `_process' /home/666/app/models/channel/email_parser.rb:115:in `block in process' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:93:in `block in timeout' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:33:in `block in catch' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:33:in `catch' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:33:in `catch' /usr/local/rvm/rubies/ruby-2.5.5/lib/ruby/2.5.0/timeout.rb:108:in `timeout' /home/666/app/models/channel/email_parser.rb:114:in `process' /home/666/app/models/channel/driver/imap.rb:264:in `block in fetch' /home/666/app/models/channel/driver/imap.rb:212:in `each' /home/666/app/models/channel/driver/imap.rb:212:in `fetch' /home/666/app/models/channel.rb:56:in `fetch' /usr/local/rvm/gems/ruby-2.5.5/gems/activerecord-5.2.3/lib/active_record/relation/delegation.rb:71:in `each' /usr/local/rvm/gems/ruby-2.5.5/gems/activerecord-5.2.3/lib/active_record/relation/delegation.rb:71:in `each' /home/666/app/models/channel.rb:30:in `fetch' (eval):1:in `_start_job' /home/666/app/models/scheduler.rb:286:in `eval' /home/666/app/models/scheduler.rb:286:in `_start_job' /home/666/app/models/scheduler.rb:239:in `block (2 levels) in start_job' /home/666/app/models/scheduler.rb:237:in `loop' /home/666/app/models/scheduler.rb:237:in `block in start_job' /usr/local/rvm/gems/ruby-2.5.5/gems/logging-2.2.2/lib/logging/diagnostic_context.rb:474:in `block in create_with_logging_context' ``` ### Steps to reproduce the behavior: Put the following mail in a file: ``` Mime-Version: 1.0 X-Mailer: GroupWise 2014 R2 Subject: Broken mail Date: Thu, 12 Sep 2019 09:09:48 +0200 Message-ID: <5D79EF3C020000D2000A2654@example.de> From: "Lassen Sie sich =?UTF-8?B?VmVyd8O2aG5lbg==?=" <Verwöhnen@example.de> To: zammad@example.com Content-Type: multipart/alternative; boundary="____TAYQQEUZELYJMCBZPMES____" --____TAYQQEUZELYJMCBZPMES____ ``` Process it via: `cat /path/to/your.eml | time rails r 'Channel::Driver::MailStdin.new'` Yes I'm sure this is a bug and no feature request or a general question.
process
binary encoded data in cc field fails mail processing hi there thanks for filing an issue please ensure the following things before creating an issue thank you 🤓 since november we handle all requests except real bugs at our community board full explanation please post feature requests development questions technical questions on the board if you think you hit a bug please continue search existing issues and the changelog md for your issue there might be a solution already make sure to use the latest version of zammad if possible add the log production log file from your system attention make sure no confidential data is in it please write the issue in english don t remove the template otherwise we will close the issue without further comments ask questions about zammad configuration and usage at our mailinglist see note we always do our best unfortunately sometimes there are too many requests and we can t handle everything at once if you want to prioritize escalate your issue you can do so by means of a support contract see the upper textblock will be removed automatically when you submit your issue infos used zammad version installation method source package any operating system any database version any elasticsearch version any browser version any ticket id expected behavior zammad will decode to cc and from fields correctly and import the mail actual behavior in some special cases processing of mails fail if to cc or from contains binary data and move the mail to unprocessible mails log e error error can t process email you will find it for bug reporting under home tmp unprocessable mail eml please create an issue at e error incompatible character encodings utf and ascii encoding compatibilityerror usr local rvm gems ruby bundler gems mail lib mail encodings rb in join usr local rvm gems ruby bundler gems mail lib mail encodings rb in value decode usr local rvm gems ruby bundler gems mail lib mail encodings rb in decode encode usr local rvm gems ruby bundler gems mail lib mail fields unstructured field rb in do decode usr local rvm gems ruby bundler gems mail lib mail fields unstructured field rb in decoded usr local rvm gems ruby bundler gems mail lib mail field rb in method missing home app models channel email parser rb in rescue in block in message header hash home app models channel email parser rb in block in message header hash home app models channel email parser rb in map home app models channel email parser rb in message header hash home app models channel email parser rb in parse home app models channel email parser rb in process home app models channel email parser rb in block in process usr local rvm rubies ruby lib ruby timeout rb in block in timeout usr local rvm rubies ruby lib ruby timeout rb in block in catch usr local rvm rubies ruby lib ruby timeout rb in catch usr local rvm rubies ruby lib ruby timeout rb in catch usr local rvm rubies ruby lib ruby timeout rb in timeout home app models channel email parser rb in process home app models channel driver imap rb in block in fetch home app models channel driver imap rb in each home app models channel driver imap rb in fetch home app models channel rb in fetch usr local rvm gems ruby gems activerecord lib active record relation delegation rb in each usr local rvm gems ruby gems activerecord lib active record relation delegation rb in each home app models channel rb in fetch eval in start job home app models scheduler rb in eval home app models scheduler rb in start job home app models scheduler rb in block levels in start job home app models scheduler rb in loop home app models scheduler rb in block in start job usr local rvm gems ruby gems logging lib logging diagnostic context rb in block in create with logging context steps to reproduce the behavior put the following mail in a file mime version x mailer groupwise subject broken mail date thu sep message id from lassen sie sich utf b to zammad example com content type multipart alternative boundary tayqqeuzelyjmcbzpmes tayqqeuzelyjmcbzpmes process it via cat path to your eml time rails r channel driver mailstdin new yes i m sure this is a bug and no feature request or a general question
1
4,131
7,087,263,576
IssuesEvent
2018-01-11 17:13:10
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
xref to web page kills PDF, hangs HTML5 for a while
P1 bug html5 pdf preprocess preprocess2
## Expected Behavior One of my customers had this link in a topic -- it's missing `@scope="external"` but is otherwise normal and worked in the past: ``` <p>For more information about each edition, see the <xref format="html" href="http://www-01.ibm.com/common/ssi/ShowDoc.wss?docURL=/common/ssi/rep_ca/0/897/ENUS217-240/index.html" >IBM Software Announcement</xref> ``` Expected: the link works. Based on the specification, processing default for links beginning with http, https, ftp, and mailto should be treated as external, so my customers shouldn't be forced to add it (though I tend to encourage adding it), and certainly shouldn't get failing builds because of it: https://docs.oasis-open.org/dita/dita/v1.3/errata01/os/complete/part1-base/langRef/attributes/thescopeattribute.html ## Actual Behavior With a simple one-topic map, and that link in the topic, HTML5 hangs for an extended period at `topicpull`. In 2.5.4, the HTML5 build otherwise works locally on my Windows machine, but when I run it on a Linux machine I get this error from the `topic-fragment` step -- I don't know why only there: `Error: java.lang.IllegalArgumentException: Only file URI scheme supported: http://www-01.ibm.com/common/ssi/ShowDoc.wss?docURL=/common/ssi/rep_ca/0/897/ENUS217-240/index.html` I don't have 3.0 on that machine to test HTML5 there. But locally, with PDF and `preprocess2`, that link now kills the build in the merge step: ``` map2pdf2: [topic-merge] Processing C:\Users\IBM_AD~1\AppData\Local\Temp\temp20171204144909813\e17ddccb7429a132b18bfa84b28d3bed6f844f1a.ditamap [topic-merge] Processing C:\Users\IBM_AD~1\AppData\Local\Temp\temp20171204144909813\b7e631acad95a349a712d265a6b2a2d6792ac95e.dita [topic-merge] null Error: Failed to run pipeline: Failed to process merged topics: Supplied element name is a zero-length string ``` ## Possible Solution Per the specification, treat the 4 known prefixes as "external" when no other scope is specified or inherited (should be relatively straightforward for `xref` as in this case but not sure how this will affect cascading in maps). Apart from the crash, which is obviously a bug itself -- not treating this as external is also a bug because the spec says it should be treated as external; the spec also allows other prefixes to be treated as external, but implementing this as some sort of configuration option seems more like a new feature. ## Steps to Reproduce Build this map to PDF or HTML5 for the different behaviors described: [scopes.zip](https://github.com/dita-ot/dita-ot/files/1528742/scopes.zip)
2.0
xref to web page kills PDF, hangs HTML5 for a while - ## Expected Behavior One of my customers had this link in a topic -- it's missing `@scope="external"` but is otherwise normal and worked in the past: ``` <p>For more information about each edition, see the <xref format="html" href="http://www-01.ibm.com/common/ssi/ShowDoc.wss?docURL=/common/ssi/rep_ca/0/897/ENUS217-240/index.html" >IBM Software Announcement</xref> ``` Expected: the link works. Based on the specification, processing default for links beginning with http, https, ftp, and mailto should be treated as external, so my customers shouldn't be forced to add it (though I tend to encourage adding it), and certainly shouldn't get failing builds because of it: https://docs.oasis-open.org/dita/dita/v1.3/errata01/os/complete/part1-base/langRef/attributes/thescopeattribute.html ## Actual Behavior With a simple one-topic map, and that link in the topic, HTML5 hangs for an extended period at `topicpull`. In 2.5.4, the HTML5 build otherwise works locally on my Windows machine, but when I run it on a Linux machine I get this error from the `topic-fragment` step -- I don't know why only there: `Error: java.lang.IllegalArgumentException: Only file URI scheme supported: http://www-01.ibm.com/common/ssi/ShowDoc.wss?docURL=/common/ssi/rep_ca/0/897/ENUS217-240/index.html` I don't have 3.0 on that machine to test HTML5 there. But locally, with PDF and `preprocess2`, that link now kills the build in the merge step: ``` map2pdf2: [topic-merge] Processing C:\Users\IBM_AD~1\AppData\Local\Temp\temp20171204144909813\e17ddccb7429a132b18bfa84b28d3bed6f844f1a.ditamap [topic-merge] Processing C:\Users\IBM_AD~1\AppData\Local\Temp\temp20171204144909813\b7e631acad95a349a712d265a6b2a2d6792ac95e.dita [topic-merge] null Error: Failed to run pipeline: Failed to process merged topics: Supplied element name is a zero-length string ``` ## Possible Solution Per the specification, treat the 4 known prefixes as "external" when no other scope is specified or inherited (should be relatively straightforward for `xref` as in this case but not sure how this will affect cascading in maps). Apart from the crash, which is obviously a bug itself -- not treating this as external is also a bug because the spec says it should be treated as external; the spec also allows other prefixes to be treated as external, but implementing this as some sort of configuration option seems more like a new feature. ## Steps to Reproduce Build this map to PDF or HTML5 for the different behaviors described: [scopes.zip](https://github.com/dita-ot/dita-ot/files/1528742/scopes.zip)
process
xref to web page kills pdf hangs for a while expected behavior one of my customers had this link in a topic it s missing scope external but is otherwise normal and worked in the past for more information about each edition see the xref format html href ibm software announcement expected the link works based on the specification processing default for links beginning with http https ftp and mailto should be treated as external so my customers shouldn t be forced to add it though i tend to encourage adding it and certainly shouldn t get failing builds because of it actual behavior with a simple one topic map and that link in the topic hangs for an extended period at topicpull in the build otherwise works locally on my windows machine but when i run it on a linux machine i get this error from the topic fragment step i don t know why only there error java lang illegalargumentexception only file uri scheme supported i don t have on that machine to test there but locally with pdf and that link now kills the build in the merge step processing c users ibm ad appdata local temp ditamap processing c users ibm ad appdata local temp dita null error failed to run pipeline failed to process merged topics supplied element name is a zero length string possible solution per the specification treat the known prefixes as external when no other scope is specified or inherited should be relatively straightforward for xref as in this case but not sure how this will affect cascading in maps apart from the crash which is obviously a bug itself not treating this as external is also a bug because the spec says it should be treated as external the spec also allows other prefixes to be treated as external but implementing this as some sort of configuration option seems more like a new feature steps to reproduce build this map to pdf or for the different behaviors described
1
254,225
19,190,047,105
IssuesEvent
2021-12-05 21:03:15
Azure/missionlz
https://api.github.com/repos/Azure/missionlz
closed
Document development using GitHub Code Spaces
documentation
Benefit/Result/Outcome So that users will know that a simple way to develop and deploy MLZ in commercial Azure (and possibly Azure Gov) is to use GitHub Code Spaces. Description GitHub Code Spaces integrates with our development container, making general development easier, and making it easy to spin up separate development environments for things like PR approvals. Acceptance Criteria - Documentation exists on using GitHub Code Spaces with MLZ
1.0
Document development using GitHub Code Spaces - Benefit/Result/Outcome So that users will know that a simple way to develop and deploy MLZ in commercial Azure (and possibly Azure Gov) is to use GitHub Code Spaces. Description GitHub Code Spaces integrates with our development container, making general development easier, and making it easy to spin up separate development environments for things like PR approvals. Acceptance Criteria - Documentation exists on using GitHub Code Spaces with MLZ
non_process
document development using github code spaces benefit result outcome so that users will know that a simple way to develop and deploy mlz in commercial azure and possibly azure gov is to use github code spaces description github code spaces integrates with our development container making general development easier and making it easy to spin up separate development environments for things like pr approvals acceptance criteria documentation exists on using github code spaces with mlz
0
8,887
11,984,619,107
IssuesEvent
2020-04-07 16:08:07
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
opened
Promote lagging betas
type: process
Looking at our README, we should be *thinking* about promoting these packages to GA: - [ ] Google.Cloud.Asset.V1 (needs Google.Cloud.OrgPolicy.V1 and Google.Identity.AccessContextManager.{Type,V1}) - [ ] Google.Cloud.BigQuery.V2 - [ ] Google.Cloud.Bigtable.V2 (and Admin, and Common) - [ ] Google.Cloud.DataCatalog.V1 - [ ] Google.Cloud.Diagnostics.AspNet (in progress) - [ ] Google.Cloud.Dialogflow.V2 (beta only just released; waiting for in-progress changes to land) - [ ] Google.Cloud.Dlp.V3 - [ ] Google.Cloud.Firestore (and related) - [ ] Google.Cloud.Kms.V1 - [ ] Google.PubSub.V2 (waiting for feature implementation) - [ ] Google.Cloud.RecaptchaEnterprise.V1 - [ ] Google.Cloud.Spanner.{...} (do we need to wait for emulator support?) - [ ] Google.Cloud.Storage.V1 (wait for final features) - [ ] Google.Cloud.WebRisk.V1 Will check these off as we complete them.
1.0
Promote lagging betas - Looking at our README, we should be *thinking* about promoting these packages to GA: - [ ] Google.Cloud.Asset.V1 (needs Google.Cloud.OrgPolicy.V1 and Google.Identity.AccessContextManager.{Type,V1}) - [ ] Google.Cloud.BigQuery.V2 - [ ] Google.Cloud.Bigtable.V2 (and Admin, and Common) - [ ] Google.Cloud.DataCatalog.V1 - [ ] Google.Cloud.Diagnostics.AspNet (in progress) - [ ] Google.Cloud.Dialogflow.V2 (beta only just released; waiting for in-progress changes to land) - [ ] Google.Cloud.Dlp.V3 - [ ] Google.Cloud.Firestore (and related) - [ ] Google.Cloud.Kms.V1 - [ ] Google.PubSub.V2 (waiting for feature implementation) - [ ] Google.Cloud.RecaptchaEnterprise.V1 - [ ] Google.Cloud.Spanner.{...} (do we need to wait for emulator support?) - [ ] Google.Cloud.Storage.V1 (wait for final features) - [ ] Google.Cloud.WebRisk.V1 Will check these off as we complete them.
process
promote lagging betas looking at our readme we should be thinking about promoting these packages to ga google cloud asset needs google cloud orgpolicy and google identity accesscontextmanager type google cloud bigquery google cloud bigtable and admin and common google cloud datacatalog google cloud diagnostics aspnet in progress google cloud dialogflow beta only just released waiting for in progress changes to land google cloud dlp google cloud firestore and related google cloud kms google pubsub waiting for feature implementation google cloud recaptchaenterprise google cloud spanner do we need to wait for emulator support google cloud storage wait for final features google cloud webrisk will check these off as we complete them
1
36,937
18,050,464,132
IssuesEvent
2021-09-19 16:56:21
RHEAGROUP/COMET-WebServices-Community-Edition
https://api.github.com/repos/RHEAGROUP/COMET-WebServices-Community-Edition
opened
[Improve] Performance of ModelReferenceDataLibraryDao.GetSiteReferenceDataLibraryDependency function
major performance ?sp
### Prerequisites - [x] I have written a descriptive issue title - [x] I have verified that I am running the latest version of the COMET Web Services - [x] I have searched [open](https://github.com/RHEAGROUP/COMET-WebServices-Community-Edition/issues) and [closed](https://github.com/RHEAGROUP/COMET-WebServices-Community-Edition/issues?q=is%3Aissue+is%3Aclosed) issues to ensure it has not already been reported ### Description The `ModelReferenceDataLibraryDao.GetSiteReferenceDataLibraryDependency` function has a custom query defined to recursively get the rlds to form a chain of RDLs. This query is making use of the `SiteReferenceDataLibrary_View` and should be refactored to make use of the `SiteReferenceDataLibrary_Cache` for improved performance. ### Steps to Reproduce NA ### System Configuration <!-- Tell us about the environment where you are experiencing the bug --> - COMET Web Services version: 6.1.6 <!-- Thanks for reporting the issue to COMET Web Services! -->
True
[Improve] Performance of ModelReferenceDataLibraryDao.GetSiteReferenceDataLibraryDependency function - ### Prerequisites - [x] I have written a descriptive issue title - [x] I have verified that I am running the latest version of the COMET Web Services - [x] I have searched [open](https://github.com/RHEAGROUP/COMET-WebServices-Community-Edition/issues) and [closed](https://github.com/RHEAGROUP/COMET-WebServices-Community-Edition/issues?q=is%3Aissue+is%3Aclosed) issues to ensure it has not already been reported ### Description The `ModelReferenceDataLibraryDao.GetSiteReferenceDataLibraryDependency` function has a custom query defined to recursively get the rlds to form a chain of RDLs. This query is making use of the `SiteReferenceDataLibrary_View` and should be refactored to make use of the `SiteReferenceDataLibrary_Cache` for improved performance. ### Steps to Reproduce NA ### System Configuration <!-- Tell us about the environment where you are experiencing the bug --> - COMET Web Services version: 6.1.6 <!-- Thanks for reporting the issue to COMET Web Services! -->
non_process
performance of modelreferencedatalibrarydao getsitereferencedatalibrarydependency function prerequisites i have written a descriptive issue title i have verified that i am running the latest version of the comet web services i have searched and issues to ensure it has not already been reported description the modelreferencedatalibrarydao getsitereferencedatalibrarydependency function has a custom query defined to recursively get the rlds to form a chain of rdls this query is making use of the sitereferencedatalibrary view and should be refactored to make use of the sitereferencedatalibrary cache for improved performance steps to reproduce na system configuration comet web services version
0
182,439
14,135,102,992
IssuesEvent
2020-11-10 00:52:55
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
opened
tests/provider: MQ test failing in GovCloud
partition/aws-us-gov tests
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description These tests fails on GovCloud: ``` TestAccDataSourceAWSMqBroker_basic ``` With this error: `no such host`. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * aws_mq_broker ### Terraform Configuration Files Example config: ```hcl resource "aws_mq_broker" "acctest" { auto_minor_version_upgrade = true apply_immediately = true broker_name = "yakbroker" configuration { id = aws_mq_configuration.acctest.id revision = aws_mq_configuration.acctest.latest_revision } deployment_mode = "ACTIVE_STANDBY_MULTI_AZ" engine_type = "ActiveMQ" engine_version = "5.15.0" host_instance_type = "mq.t2.micro" maintenance_window_start_time { day_of_week = "TUESDAY" time_of_day = "02:00" time_zone = "CET" } publicly_accessible = true security_groups = aws_security_group.acctest[*].id subnet_ids = aws_subnet.acctest[*].id user { username = "Ender" password = "AndrewWiggin" } user { username = "Petra" password = "PetraArkanian" console_access = true groups = ["dragon", "salamander", "leopard"] } depends_on = [aws_internet_gateway.acctest] } ``` ### Debug Output ``` TestAccDataSourceAWSMqBroker_basic: data_source_aws_mq_broker_test.go:20: Step 1/2 error: Error running apply: 2020/11/08 10:03:29 [DEBUG] Using modified User-Agent: Terraform/0.12.29 HashiCorp-terraform-exec/0.10.0 Error: RequestError: send request failed caused by: Post "https://mq.us-gov-west-1.amazonaws.com/v1/configurations": dial tcp: lookup mq.us-gov-west-1.amazonaws.com on 192.168.22.2:53: no such host --- FAIL: TestAccDataSourceAWSMqBroker_basic (40.97s) ``` ### References * #0000
1.0
tests/provider: MQ test failing in GovCloud - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description These tests fails on GovCloud: ``` TestAccDataSourceAWSMqBroker_basic ``` With this error: `no such host`. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * aws_mq_broker ### Terraform Configuration Files Example config: ```hcl resource "aws_mq_broker" "acctest" { auto_minor_version_upgrade = true apply_immediately = true broker_name = "yakbroker" configuration { id = aws_mq_configuration.acctest.id revision = aws_mq_configuration.acctest.latest_revision } deployment_mode = "ACTIVE_STANDBY_MULTI_AZ" engine_type = "ActiveMQ" engine_version = "5.15.0" host_instance_type = "mq.t2.micro" maintenance_window_start_time { day_of_week = "TUESDAY" time_of_day = "02:00" time_zone = "CET" } publicly_accessible = true security_groups = aws_security_group.acctest[*].id subnet_ids = aws_subnet.acctest[*].id user { username = "Ender" password = "AndrewWiggin" } user { username = "Petra" password = "PetraArkanian" console_access = true groups = ["dragon", "salamander", "leopard"] } depends_on = [aws_internet_gateway.acctest] } ``` ### Debug Output ``` TestAccDataSourceAWSMqBroker_basic: data_source_aws_mq_broker_test.go:20: Step 1/2 error: Error running apply: 2020/11/08 10:03:29 [DEBUG] Using modified User-Agent: Terraform/0.12.29 HashiCorp-terraform-exec/0.10.0 Error: RequestError: send request failed caused by: Post "https://mq.us-gov-west-1.amazonaws.com/v1/configurations": dial tcp: lookup mq.us-gov-west-1.amazonaws.com on 192.168.22.2:53: no such host --- FAIL: TestAccDataSourceAWSMqBroker_basic (40.97s) ``` ### References * #0000
non_process
tests provider mq test failing in govcloud community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description these tests fails on govcloud testaccdatasourceawsmqbroker basic with this error no such host new or affected resource s aws mq broker terraform configuration files example config hcl resource aws mq broker acctest auto minor version upgrade true apply immediately true broker name yakbroker configuration id aws mq configuration acctest id revision aws mq configuration acctest latest revision deployment mode active standby multi az engine type activemq engine version host instance type mq micro maintenance window start time day of week tuesday time of day time zone cet publicly accessible true security groups aws security group acctest id subnet ids aws subnet acctest id user username ender password andrewwiggin user username petra password petraarkanian console access true groups depends on debug output testaccdatasourceawsmqbroker basic data source aws mq broker test go step error error running apply using modified user agent terraform hashicorp terraform exec error requesterror send request failed caused by post dial tcp lookup mq us gov west amazonaws com on no such host fail testaccdatasourceawsmqbroker basic references
0
7,867
2,938,467,145
IssuesEvent
2015-07-01 10:58:37
molgenis/molgenis
https://api.github.com/repos/molgenis/molgenis
opened
Cannot save new TypeTest entity
bug release test v15.07
-Load emx_all_datatypes.xls - Select TypeTest click + button. - Click 'eye icon'. - Fill in all (required fields). - Click 'create'. -> Nothing happens.
1.0
Cannot save new TypeTest entity - -Load emx_all_datatypes.xls - Select TypeTest click + button. - Click 'eye icon'. - Fill in all (required fields). - Click 'create'. -> Nothing happens.
non_process
cannot save new typetest entity load emx all datatypes xls select typetest click button click eye icon fill in all required fields click create nothing happens
0
17,160
22,718,286,679
IssuesEvent
2022-07-06 05:40:44
Carlosmtp/DomuzSGI
https://api.github.com/repos/Carlosmtp/DomuzSGI
closed
Rutas para actualizar procesos
High New Function Process Management
- [x] Crear las rutas dónde se recibirán los datos del frontend para la actualizacion de un procesos. - [x] Crear la función que permita actualizar un proceso en la base de datos.
1.0
Rutas para actualizar procesos - - [x] Crear las rutas dónde se recibirán los datos del frontend para la actualizacion de un procesos. - [x] Crear la función que permita actualizar un proceso en la base de datos.
process
rutas para actualizar procesos crear las rutas dónde se recibirán los datos del frontend para la actualizacion de un procesos crear la función que permita actualizar un proceso en la base de datos
1
6,298
9,305,184,764
IssuesEvent
2019-03-25 05:17:24
brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
https://api.github.com/repos/brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
closed
Написать Техническое задание
Start of the work Work in process
### @goldmen4ik нужно написать часть технического задания. Сделай последние 2 пункта.
1.0
Написать Техническое задание - ### @goldmen4ik нужно написать часть технического задания. Сделай последние 2 пункта.
process
написать техническое задание нужно написать часть технического задания сделай последние пункта
1
165
2,585,740,896
IssuesEvent
2015-02-17 03:25:35
dalehenrich/metacello-work
https://api.github.com/repos/dalehenrich/metacello-work
closed
re-registration with `MetacelloProjectRegistry>>registerProjectRegistration:` can incorrectly update versionString: in versionInfo
in process
In GemStone, on a restore from backup, old registrations (that do not match the current disk) are brought into the image ... certain commands like `lock` will do an `unregister`, `register` and when that happens the cached `versionInfo.versionString` is overwritten with a fresh value based on the contents of disk ... not quite the expected behavior ... something along the lines of `setIfnOtCached:` is needed here
1.0
re-registration with `MetacelloProjectRegistry>>registerProjectRegistration:` can incorrectly update versionString: in versionInfo - In GemStone, on a restore from backup, old registrations (that do not match the current disk) are brought into the image ... certain commands like `lock` will do an `unregister`, `register` and when that happens the cached `versionInfo.versionString` is overwritten with a fresh value based on the contents of disk ... not quite the expected behavior ... something along the lines of `setIfnOtCached:` is needed here
process
re registration with metacelloprojectregistry registerprojectregistration can incorrectly update versionstring in versioninfo in gemstone on a restore from backup old registrations that do not match the current disk are brought into the image certain commands like lock will do an unregister register and when that happens the cached versioninfo versionstring is overwritten with a fresh value based on the contents of disk not quite the expected behavior something along the lines of setifnotcached is needed here
1
14,928
18,359,530,173
IssuesEvent
2021-10-09 01:46:03
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Incorrect behavior: cross-domain iframe content is loaded
TYPE: bug AREA: client FREQUENCY: level 1 SYSTEM: iframe processing STATE: Stale
Browser behavior: Cross-domain content, loaded via `location.replace` or `location.assign` call, don't display. In 'Console` presents the "Refused to display 'https://mail.ru/' in a frame because it set 'X-Frame-Options' to 'sameorigin'. error." In proxy, the cross-domain iframe content is displayed. Sample: index.js ``` var http = require('http'); var fs = require('fs'); http.createServer(function (req, res) { if (req.url === '/') { var content = fs.readFileSync('index.html'); res.end(content); } }).listen(3000); http.createServer(function (req, res) { if (req.url === '/') { var content = fs.readFileSync('iframe.html'); res.end(content); } }).listen(3001); ``` index.html ``` <!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <title>Title</title> </head> <body> <h1>Same domain content</h1> <iframe src="http://localhost:3001"></iframe> <script> window.setTimeout(function () { var iframeWindow = document.getElementsByTagName('iframe')[0].contentWindow; iframeWindow.location.replace('http://mail.ru'); }, 1000); </script> </body> </html> ``` iframe.html ``` <!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <title>Title</title> </head> <body> <h1>Iframe content</h1> </body> </html> ``` Browsers: IE11, Edge(41.16299.15.0), Chrome (63.0.3239.132), Firefox (58.0.1)
1.0
Incorrect behavior: cross-domain iframe content is loaded - Browser behavior: Cross-domain content, loaded via `location.replace` or `location.assign` call, don't display. In 'Console` presents the "Refused to display 'https://mail.ru/' in a frame because it set 'X-Frame-Options' to 'sameorigin'. error." In proxy, the cross-domain iframe content is displayed. Sample: index.js ``` var http = require('http'); var fs = require('fs'); http.createServer(function (req, res) { if (req.url === '/') { var content = fs.readFileSync('index.html'); res.end(content); } }).listen(3000); http.createServer(function (req, res) { if (req.url === '/') { var content = fs.readFileSync('iframe.html'); res.end(content); } }).listen(3001); ``` index.html ``` <!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <title>Title</title> </head> <body> <h1>Same domain content</h1> <iframe src="http://localhost:3001"></iframe> <script> window.setTimeout(function () { var iframeWindow = document.getElementsByTagName('iframe')[0].contentWindow; iframeWindow.location.replace('http://mail.ru'); }, 1000); </script> </body> </html> ``` iframe.html ``` <!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <title>Title</title> </head> <body> <h1>Iframe content</h1> </body> </html> ``` Browsers: IE11, Edge(41.16299.15.0), Chrome (63.0.3239.132), Firefox (58.0.1)
process
incorrect behavior cross domain iframe content is loaded browser behavior cross domain content loaded via location replace or location assign call don t display in console presents the refused to display in a frame because it set x frame options to sameorigin error in proxy the cross domain iframe content is displayed sample index js var http require http var fs require fs http createserver function req res if req url var content fs readfilesync index html res end content listen http createserver function req res if req url var content fs readfilesync iframe html res end content listen index html title same domain content iframe src window settimeout function var iframewindow document getelementsbytagname iframe contentwindow iframewindow location replace iframe html title iframe content browsers edge chrome firefox
1
9,222
12,257,914,548
IssuesEvent
2020-05-06 14:23:00
MHRA/products
https://api.github.com/repos/MHRA/products
closed
PARs - Upload new PAR
EPIC - PARs process
### User want As a Medical Writer in the licensing team I would like to upload a new PAR document to products.mhra.gov.uk ### Technical acceptance criteria After submitting a new version of the document the: - [ ] The PDF is stored in blob storage after upload - [ ] Azure search service is updated if necessary ### Customer acceptance criteria - [ ] Medical writers can enter the relevant metadata about the document (such as the product name, active substances and PL/NR/THR number) - [ ] Medical writers can upload the new document - [ ] After submitting the user is shown a summary of what they've submitted - [ ] The new document is available on the products.mhra.gov.uk site
1.0
PARs - Upload new PAR - ### User want As a Medical Writer in the licensing team I would like to upload a new PAR document to products.mhra.gov.uk ### Technical acceptance criteria After submitting a new version of the document the: - [ ] The PDF is stored in blob storage after upload - [ ] Azure search service is updated if necessary ### Customer acceptance criteria - [ ] Medical writers can enter the relevant metadata about the document (such as the product name, active substances and PL/NR/THR number) - [ ] Medical writers can upload the new document - [ ] After submitting the user is shown a summary of what they've submitted - [ ] The new document is available on the products.mhra.gov.uk site
process
pars upload new par user want as a medical writer in the licensing team i would like to upload a new par document to products mhra gov uk technical acceptance criteria after submitting a new version of the document the the pdf is stored in blob storage after upload azure search service is updated if necessary customer acceptance criteria medical writers can enter the relevant metadata about the document such as the product name active substances and pl nr thr number medical writers can upload the new document after submitting the user is shown a summary of what they ve submitted the new document is available on the products mhra gov uk site
1
59,036
17,015,010,287
IssuesEvent
2021-07-02 10:42:16
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
TypeError: wrong argument type nil (expected String)
Component: merkaartor Priority: major Type: defect
**[Submitted to the original trac issue database at 9.19am, Friday, 21st May 2010]** Sometimes when I upload, the server refuses it with this message: ``` There was an error uploading this request (500) "TypeError: wrong argument type nil (expected String)" Please redownload the problematic feature to handle the conflict. ``` There doesn't seem to be any way to upload after that, until a New File is started and the current edits discarded.
1.0
TypeError: wrong argument type nil (expected String) - **[Submitted to the original trac issue database at 9.19am, Friday, 21st May 2010]** Sometimes when I upload, the server refuses it with this message: ``` There was an error uploading this request (500) "TypeError: wrong argument type nil (expected String)" Please redownload the problematic feature to handle the conflict. ``` There doesn't seem to be any way to upload after that, until a New File is started and the current edits discarded.
non_process
typeerror wrong argument type nil expected string sometimes when i upload the server refuses it with this message there was an error uploading this request typeerror wrong argument type nil expected string please redownload the problematic feature to handle the conflict there doesn t seem to be any way to upload after that until a new file is started and the current edits discarded
0
86,782
3,729,042,376
IssuesEvent
2016-03-07 05:10:54
cs2103jan2016-t11-3j/main
https://api.github.com/repos/cs2103jan2016-t11-3j/main
closed
A user will receive an alert when the task list exceeds a certain number
priority.low type.story
so that he will be alerted of a heavy workload and adjust accordingly
1.0
A user will receive an alert when the task list exceeds a certain number - so that he will be alerted of a heavy workload and adjust accordingly
non_process
a user will receive an alert when the task list exceeds a certain number so that he will be alerted of a heavy workload and adjust accordingly
0
4,206
7,165,099,750
IssuesEvent
2018-01-29 13:28:29
Incubaid/crm
https://api.github.com/repos/Incubaid/crm
closed
Author last does not show up at Ambassador Deal Type
process_wontfix
Can not see Author Last value on Ambassador Deal Type. PLs fix. Thank you
1.0
Author last does not show up at Ambassador Deal Type - Can not see Author Last value on Ambassador Deal Type. PLs fix. Thank you
process
author last does not show up at ambassador deal type can not see author last value on ambassador deal type pls fix thank you
1
20,613
14,075,643,413
IssuesEvent
2020-11-04 09:20:31
hyperledger-labs/business-partner-agent
https://api.github.com/repos/hyperledger-labs/business-partner-agent
closed
Provide helm chart
Infrastructure needs-discussion
As company I want to easily deploy a BPA as helm chart in order to easily launch it in my existing kubernetes cluster . Currently it I have to build the container , create a chart,....
1.0
Provide helm chart - As company I want to easily deploy a BPA as helm chart in order to easily launch it in my existing kubernetes cluster . Currently it I have to build the container , create a chart,....
non_process
provide helm chart as company i want to easily deploy a bpa as helm chart in order to easily launch it in my existing kubernetes cluster currently it i have to build the container create a chart
0
5,377
8,204,069,173
IssuesEvent
2018-09-03 03:57:32
RennurApps/AwareIM-resources
https://api.github.com/repos/RennurApps/AwareIM-resources
closed
REST Services in Aware IM: Pagination, Offset & Limit parameters
CT: Process CT: Query CT: REST Services FT: Feature SC: Wontfix v8.1
Currently there is a limit of how many items can be exposed by the Service at one time. This can range form anywhere between 10, 25 up to 100 usually. Aware IM can expose up to 1000 records at a time. This can be set by the the TAKE BEST parameter. If your database has more than 1000 items, the service will not expose all of the them to the outside. This is where the pagination feature is required. Expose the items in chunks of 100 for example. Examples: [**https://support.smartbear.com/qacomplete/docs/developer/api/rest/api/reference/paging.html**](https://support.smartbear.com/qacomplete/docs/developer/api/rest/api/reference/paging.html) [**RESTful API Design: can your API give developers just the information they need?**](https://apigee.com/about/blog/technology/restful-api-design-can-your-api-give-developers-just-information-they-need) [**REST API Design: Filtering, Sorting, and Pagination**](https://www.moesif.com/blog/technical/api-design/REST-API-Design-Filtering) Parameter | Type | Description -- | -- | -- limit | int | If a limit count is given, no more than that many rows will be returned (but possibly less, if the query itself yields fewer rows). page | int | This parameter says to skip that many rows before beginning to return rows. If both *page *and *limit *appear, then page rows are skipped before starting to count the limit rows that are returned. The default value is 1. orderBy | string | This parameter is used to order your results. If an order is not given, the default order modifiedAt:desc will be applied. offset | int |
1.0
REST Services in Aware IM: Pagination, Offset & Limit parameters - Currently there is a limit of how many items can be exposed by the Service at one time. This can range form anywhere between 10, 25 up to 100 usually. Aware IM can expose up to 1000 records at a time. This can be set by the the TAKE BEST parameter. If your database has more than 1000 items, the service will not expose all of the them to the outside. This is where the pagination feature is required. Expose the items in chunks of 100 for example. Examples: [**https://support.smartbear.com/qacomplete/docs/developer/api/rest/api/reference/paging.html**](https://support.smartbear.com/qacomplete/docs/developer/api/rest/api/reference/paging.html) [**RESTful API Design: can your API give developers just the information they need?**](https://apigee.com/about/blog/technology/restful-api-design-can-your-api-give-developers-just-information-they-need) [**REST API Design: Filtering, Sorting, and Pagination**](https://www.moesif.com/blog/technical/api-design/REST-API-Design-Filtering) Parameter | Type | Description -- | -- | -- limit | int | If a limit count is given, no more than that many rows will be returned (but possibly less, if the query itself yields fewer rows). page | int | This parameter says to skip that many rows before beginning to return rows. If both *page *and *limit *appear, then page rows are skipped before starting to count the limit rows that are returned. The default value is 1. orderBy | string | This parameter is used to order your results. If an order is not given, the default order modifiedAt:desc will be applied. offset | int |
process
rest services in aware im pagination offset limit parameters currently there is a limit of how many items can be exposed by the service at one time this can range form anywhere between up to usually aware im can expose up to records at a time this can be set by the the take best parameter if your database has more than items the service will not expose all of the them to the outside this is where the pagination feature is required expose the items in chunks of for example examples parameter type description limit int if a limit count is given no more than that many rows will be returned but possibly less if the query itself yields fewer rows page int this parameter says to skip that many rows before beginning to return rows if both page and limit appear then page rows are skipped before starting to count the limit rows that are returned the default value is orderby string this parameter is used to order your results if an order is not given the default order modifiedat desc will be applied offset int
1
22,659
31,895,936,943
IssuesEvent
2023-09-18 01:41:45
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - verbatimLabel
Term - change non-normative Task Group - Material Sample Process - complete Class - MaterialEntity
## Term change * Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/) * Efficacy Justification (why is this change necessary?): Verbatim Label is currently organized in the MaterialSample class. Organizing this term with MaterialEntity will also provide for its use with any existing classes of material things within Darwin Core, as it would be understood that MaterialEntity would be an informal superclass to `dwc:MaterialSample`, `dwc:PreservedSpecimen`, `dwc:LivingSpecimen`, `dwc:FossilSpecimen`. * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): Usage as currently occurs in Global Biodiversity Information Facility (GBIF) Darwin Core Archives would not be affected by this change. Darwin Core does not include formal class hierarchies, but if we ignore that formality and imagine what the hierarchy would look like for the classes, we have MaterialEntity as the highest for material things. All of the other material-based classes in Darwin Core (`dwc:MaterialSample`, `dwc:PreservedSpecimen`, `dwc:LivingSpecimen`, `dwc:FossilSpecimen`) might be expected to have verbatim labels. As there are no other classes in between MaterialEntity and those subtypes, verbatimLabel is best organized with MaterialEntity. * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No Current Term definition: https://dwc.tdwg.org/list/#dwc_verbatimLabel (note term is in review - https://github.com/tdwg/dwc/issues/32 Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): verbatimLabel * Term label (English, not normative): Verbatim Label * Organized in Class (e.g., Occurrence, Event, Location, Taxon): **dwc:MaterialEntity** ~~MaterialSample~~ * Definition of the term (normative): A serialized encoding intended to represent the literal, i.e., character by character, textual content of a label affixed on, near, or explicitly associated with a material entity, free from interpretation, translation, or transliteration. * Usage comments (recommendations regarding content, etc., not normative): The content of this term should include no embellishments, prefixes, headers or other additions made to the text. Abbreviations must not be expanded and supposed misspellings must not be corrected. Lines or breakpoints between blocks of text that could be verified by seeing the original labels or images of them may be used. Examples of material entities include preserved specimens, fossil specimens, and material samples. Best practice is to use UTF-8 for all characters. Best practice is to add comment “verbatimLabel derived from human transcription” in occurrenceRemarks. * Examples (not normative): Examples can be found at https://dwc.tdwg.org/examples/verbatimLabel. * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): /Marks/Mark/MarkText
1.0
Change term - verbatimLabel - ## Term change * Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/) * Efficacy Justification (why is this change necessary?): Verbatim Label is currently organized in the MaterialSample class. Organizing this term with MaterialEntity will also provide for its use with any existing classes of material things within Darwin Core, as it would be understood that MaterialEntity would be an informal superclass to `dwc:MaterialSample`, `dwc:PreservedSpecimen`, `dwc:LivingSpecimen`, `dwc:FossilSpecimen`. * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): Usage as currently occurs in Global Biodiversity Information Facility (GBIF) Darwin Core Archives would not be affected by this change. Darwin Core does not include formal class hierarchies, but if we ignore that formality and imagine what the hierarchy would look like for the classes, we have MaterialEntity as the highest for material things. All of the other material-based classes in Darwin Core (`dwc:MaterialSample`, `dwc:PreservedSpecimen`, `dwc:LivingSpecimen`, `dwc:FossilSpecimen`) might be expected to have verbatim labels. As there are no other classes in between MaterialEntity and those subtypes, verbatimLabel is best organized with MaterialEntity. * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No Current Term definition: https://dwc.tdwg.org/list/#dwc_verbatimLabel (note term is in review - https://github.com/tdwg/dwc/issues/32 Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): verbatimLabel * Term label (English, not normative): Verbatim Label * Organized in Class (e.g., Occurrence, Event, Location, Taxon): **dwc:MaterialEntity** ~~MaterialSample~~ * Definition of the term (normative): A serialized encoding intended to represent the literal, i.e., character by character, textual content of a label affixed on, near, or explicitly associated with a material entity, free from interpretation, translation, or transliteration. * Usage comments (recommendations regarding content, etc., not normative): The content of this term should include no embellishments, prefixes, headers or other additions made to the text. Abbreviations must not be expanded and supposed misspellings must not be corrected. Lines or breakpoints between blocks of text that could be verified by seeing the original labels or images of them may be used. Examples of material entities include preserved specimens, fossil specimens, and material samples. Best practice is to use UTF-8 for all characters. Best practice is to add comment “verbatimLabel derived from human transcription” in occurrenceRemarks. * Examples (not normative): Examples can be found at https://dwc.tdwg.org/examples/verbatimLabel. * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): /Marks/Mark/MarkText
process
change term verbatimlabel term change submitter efficacy justification why is this change necessary verbatim label is currently organized in the materialsample class organizing this term with materialentity will also provide for its use with any existing classes of material things within darwin core as it would be understood that materialentity would be an informal superclass to dwc materialsample dwc preservedspecimen dwc livingspecimen dwc fossilspecimen demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations usage as currently occurs in global biodiversity information facility gbif darwin core archives would not be affected by this change darwin core does not include formal class hierarchies but if we ignore that formality and imagine what the hierarchy would look like for the classes we have materialentity as the highest for material things all of the other material based classes in darwin core dwc materialsample dwc preservedspecimen dwc livingspecimen dwc fossilspecimen might be expected to have verbatim labels as there are no other classes in between materialentity and those subtypes verbatimlabel is best organized with materialentity implications for dwciri namespace does this change affect a dwciri term version no current term definition note term is in review proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes verbatimlabel term label english not normative verbatim label organized in class e g occurrence event location taxon dwc materialentity materialsample definition of the term normative a serialized encoding intended to represent the literal i e character by character textual content of a label affixed on near or explicitly associated with a material entity free from interpretation translation or transliteration usage comments recommendations regarding content etc not normative the content of this term should include no embellishments prefixes headers or other additions made to the text abbreviations must not be expanded and supposed misspellings must not be corrected lines or breakpoints between blocks of text that could be verified by seeing the original labels or images of them may be used examples of material entities include preserved specimens fossil specimens and material samples best practice is to use utf for all characters best practice is to add comment “verbatimlabel derived from human transcription” in occurrenceremarks examples not normative examples can be found at refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative marks mark marktext
1
305,383
26,386,429,153
IssuesEvent
2023-01-12 12:39:07
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
com.hazelcast.cluster.TcpIpJoinTest.test_whenIncompatiblePartitionGroups
Team: Core Type: Test-Failure Source: Internal Module: Discovery SPI
_5.2.z_ (commit c1cb576b60c162b5aa273fcdb96b13cea119445a) Failed on openjdk-17: https://jenkins.hazelcast.com/job/Hazelcast-5.maintenance-ZuluJDK17-nightly/46/testReport/junit/com.hazelcast.cluster/TcpIpJoinTest/test_whenIncompatiblePartitionGroups/ <details><summary>Stacktrace:</summary> ``` java.lang.AssertionError at org.junit.Assert.fail(Assert.java:87) at org.junit.Assert.fail(Assert.java:96) at com.hazelcast.cluster.AbstractJoinTest.assertIncompatible(AbstractJoinTest.java:98) at com.hazelcast.cluster.TcpIpJoinTest.test_whenIncompatiblePartitionGroups(TcpIpJoinTest.java:202) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:833) ``` </details> <details><summary>Standard output:</summary> ``` Finished Running Test: test_whenHostUnresolvable in 1.049 seconds. Started Running Test: test_whenIncompatiblePartitionGroups 18:06:27,188 INFO |test_whenIncompatiblePartitionGroups| - [MetricsConfigHelper] test_whenIncompatiblePartitionGroups - [LOCAL] [dev] [5.2.2-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1' 18:06:27,190 INFO |test_whenIncompatiblePartitionGroups| - [logo] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] + + o o o o---o o----o o o---o o o----o o--o--o + + + + | | / \ / | | / / \ | | + + + + + o----o o o o o----o | o o o o----o | + + + + | | / \ / | | \ / \ | | + + o o o o o---o o----o o----o o---o o o o----o o 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Copyright (c) 2008-2022, Hazelcast, Inc. All Rights Reserved. 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Hazelcast Platform 5.2.2-SNAPSHOT (20230111 - c1cb576) starting at [127.0.0.1]:5701 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Cluster name: dev 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker. 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] The Jet engine is disabled. To enable the Jet engine on the members, do one of the following: - Change member config using Java API: config.getJetConfig().setEnabled(true) - Change XML/YAML configuration property: Set hazelcast.jet.enabled to true - Add system property: -Dhz.jet.enabled=true (for Hazelcast embedded, works only when loading config via Config.load) - Add environment variable: HZ_JET_ENABLED=true (recommended when running container image. For Hazelcast embedded, works only when loading config via Config.load) 18:06:27,201 INFO |test_whenIncompatiblePartitionGroups| - [MetricsConfigHelper] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled 18:06:27,220 INFO |test_whenIncompatiblePartitionGroups| - [Node] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Using TCP/IP discovery 18:06:27,220 WARN |test_whenIncompatiblePartitionGroups| - [CPSubsystem] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees. 18:06:27,234 INFO |test_whenIncompatiblePartitionGroups| - [Diagnostics] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments. 18:06:27,234 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is STARTING 18:06:27,242 INFO |test_whenIncompatiblePartitionGroups| - [TcpIpJoiner] hz.quizzical_shockley.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5703 is added to the blacklist. 18:06:27,243 INFO |test_whenIncompatiblePartitionGroups| - [TcpIpJoiner] hz.quizzical_shockley.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is added to the blacklist. 18:06:28,243 INFO |test_whenIncompatiblePartitionGroups| - [ClusterService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Members {size:1, ver:1} [ Member [127.0.0.1]:5701 - cb0989c3-7be0-4840-af83-ee83f5f79dba this ] 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is STARTED 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [MetricsConfigHelper] test_whenIncompatiblePartitionGroups - [LOCAL] [dev] [5.2.2-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1' 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [logo] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] + + o o o o---o o----o o o---o o o----o o--o--o + + + + | | / \ / | | / / \ | | + + + + + o----o o o o o----o | o o o o----o | + + + + | | / \ / | | \ / \ | | + + o o o o o---o o----o o----o o---o o o o----o o 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Copyright (c) 2008-2022, Hazelcast, Inc. All Rights Reserved. 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Hazelcast Platform 5.2.2-SNAPSHOT (20230111 - c1cb576) starting at [127.0.0.1]:5702 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Cluster name: dev 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker. 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] The Jet engine is disabled. To enable the Jet engine on the members, do one of the following: - Change member config using Java API: config.getJetConfig().setEnabled(true) - Change XML/YAML configuration property: Set hazelcast.jet.enabled to true - Add system property: -Dhz.jet.enabled=true (for Hazelcast embedded, works only when loading config via Config.load) - Add environment variable: HZ_JET_ENABLED=true (recommended when running container image. For Hazelcast embedded, works only when loading config via Config.load) 18:06:28,246 INFO |test_whenIncompatiblePartitionGroups| - [MetricsConfigHelper] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled 18:06:28,250 INFO |test_whenIncompatiblePartitionGroups| - [Node] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Using TCP/IP discovery 18:06:28,250 WARN |test_whenIncompatiblePartitionGroups| - [CPSubsystem] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees. 18:06:28,254 INFO |test_whenIncompatiblePartitionGroups| - [Diagnostics] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments. 18:06:28,254 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is STARTING 18:06:28,255 INFO |test_whenIncompatiblePartitionGroups| - [TcpIpJoiner] hz.eloquent_shockley.cached.thread-1 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5703 is added to the blacklist. 18:06:28,255 INFO |test_whenIncompatiblePartitionGroups| - [TcpServerConnection] hz.quizzical_shockley.IO.thread-in-2 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Initialized new cluster connection between /127.0.0.1:5701 and /127.0.0.1:50817 18:06:28,255 INFO |test_whenIncompatiblePartitionGroups| - [TcpServerConnection] hz.eloquent_shockley.IO.thread-in-1 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Initialized new cluster connection between /0.0.0.0:50817 and /127.0.0.1:5701 18:06:28,264 WARN |test_whenIncompatiblePartitionGroups| - [ClusterJoinManager] hz.eloquent_shockley.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Invalid join request from [127.0.0.1]:5701, cause: Incompatible partition groups! expected: HOST_AWARE, found: CUSTOM 18:06:28,264 WARN |test_whenIncompatiblePartitionGroups| - [ClusterJoinManager] hz.eloquent_shockley.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Received an invalid join request from [127.0.0.1]:5701, cause: Incompatible partition groups! expected: HOST_AWARE, found: CUSTOM 18:06:28,264 ERROR |test_whenIncompatiblePartitionGroups| - [cluster] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Node could not join cluster. A Configuration mismatch was detected: Incompatible partition groups! expected: HOST_AWARE, found: CUSTOM Node is going to shutdown now! 18:06:28,264 WARN |test_whenIncompatiblePartitionGroups| - [Node] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Terminating forcefully... 18:06:28,264 INFO |test_whenIncompatiblePartitionGroups| - [Node] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Shutting down connection manager... 18:06:28,264 INFO |test_whenIncompatiblePartitionGroups| - [TcpServerConnection] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Connection[id=1, /127.0.0.1:5701->/127.0.0.1:50817, qualifier=null, endpoint=[127.0.0.1]:5702, remoteUuid=04b55ecd-d879-4415-90f8-cb6124e46ed0, alive=false, connectionType=MEMBER, planeIndex=0] closed. Reason: TcpServer is stopping 18:06:28,264 INFO |test_whenIncompatiblePartitionGroups| - [TcpServerConnection] hz.eloquent_shockley.IO.thread-in-1 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Connection[id=1, /0.0.0.0:50817->/127.0.0.1:5701, qualifier=null, endpoint=[127.0.0.1]:5701, remoteUuid=cb0989c3-7be0-4840-af83-ee83f5f79dba, alive=false, connectionType=MEMBER, planeIndex=0] closed. Reason: Connection closed by the other side 18:06:28,264 INFO |test_whenIncompatiblePartitionGroups| - [Node] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Shutting down node engine... 18:06:28,266 INFO |test_whenIncompatiblePartitionGroups| - [NodeExtension] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Destroying node NodeExtension. 18:06:28,266 INFO |test_whenIncompatiblePartitionGroups| - [Node] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Hazelcast Shutdown is completed in 2 ms. 18:06:28,355 INFO |test_whenIncompatiblePartitionGroups| - [TcpIpJoiner] hz.eloquent_shockley.cached.thread-1 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is added to the blacklist. 18:06:29,255 INFO |test_whenIncompatiblePartitionGroups| - [ClusterService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Members {size:1, ver:1} [ Member [127.0.0.1]:5702 - 04b55ecd-d879-4415-90f8-cb6124e46ed0 this ] 18:06:29,255 WARN |test_whenIncompatiblePartitionGroups| - [Node] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Config seed port is 5701 and cluster size is 1. Some of the ports seem occupied! 18:06:29,256 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is STARTED 18:06:29,860 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is SHUTTING_DOWN 18:06:29,860 WARN |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Terminating forcefully... 18:06:29,860 INFO |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Shutting down connection manager... 18:06:29,861 INFO |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Shutting down node engine... 18:06:29,863 INFO |test_whenIncompatiblePartitionGroups| - [NodeExtension] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Destroying node NodeExtension. 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Hazelcast Shutdown is completed in 6 ms. 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is SHUTDOWN 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] main - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is SHUTTING_DOWN 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Node is already shutting down... Waiting for shutdown process to complete... 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] main - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is SHUTDOWN BuildInfo right after test_whenIncompatiblePartitionGroups(com.hazelcast.cluster.TcpIpJoinTest): BuildInfo{version='5.2.2-SNAPSHOT', build='20230111', buildNumber=20230111, revision=c1cb576, enterprise=false, serializationVersion=1} Hiccups measured while running test 'test_whenIncompatiblePartitionGroups(com.hazelcast.cluster.TcpIpJoinTest):' 18:06:25, accumulated pauses: 710 ms, max pause: 595 ms, pauses over 1000 ms: 0 No metrics recorded during the test ``` </details> Standard output can be found here - https://s3.console.aws.amazon.com/s3/buckets/j-artifacts/Hazelcast-5.maintenance-ZuluJDK17-nightly/46/
1.0
com.hazelcast.cluster.TcpIpJoinTest.test_whenIncompatiblePartitionGroups - _5.2.z_ (commit c1cb576b60c162b5aa273fcdb96b13cea119445a) Failed on openjdk-17: https://jenkins.hazelcast.com/job/Hazelcast-5.maintenance-ZuluJDK17-nightly/46/testReport/junit/com.hazelcast.cluster/TcpIpJoinTest/test_whenIncompatiblePartitionGroups/ <details><summary>Stacktrace:</summary> ``` java.lang.AssertionError at org.junit.Assert.fail(Assert.java:87) at org.junit.Assert.fail(Assert.java:96) at com.hazelcast.cluster.AbstractJoinTest.assertIncompatible(AbstractJoinTest.java:98) at com.hazelcast.cluster.TcpIpJoinTest.test_whenIncompatiblePartitionGroups(TcpIpJoinTest.java:202) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.lang.Thread.run(Thread.java:833) ``` </details> <details><summary>Standard output:</summary> ``` Finished Running Test: test_whenHostUnresolvable in 1.049 seconds. Started Running Test: test_whenIncompatiblePartitionGroups 18:06:27,188 INFO |test_whenIncompatiblePartitionGroups| - [MetricsConfigHelper] test_whenIncompatiblePartitionGroups - [LOCAL] [dev] [5.2.2-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1' 18:06:27,190 INFO |test_whenIncompatiblePartitionGroups| - [logo] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] + + o o o o---o o----o o o---o o o----o o--o--o + + + + | | / \ / | | / / \ | | + + + + + o----o o o o o----o | o o o o----o | + + + + | | / \ / | | \ / \ | | + + o o o o o---o o----o o----o o---o o o o----o o 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Copyright (c) 2008-2022, Hazelcast, Inc. All Rights Reserved. 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Hazelcast Platform 5.2.2-SNAPSHOT (20230111 - c1cb576) starting at [127.0.0.1]:5701 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Cluster name: dev 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker. 18:06:27,191 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] The Jet engine is disabled. To enable the Jet engine on the members, do one of the following: - Change member config using Java API: config.getJetConfig().setEnabled(true) - Change XML/YAML configuration property: Set hazelcast.jet.enabled to true - Add system property: -Dhz.jet.enabled=true (for Hazelcast embedded, works only when loading config via Config.load) - Add environment variable: HZ_JET_ENABLED=true (recommended when running container image. For Hazelcast embedded, works only when loading config via Config.load) 18:06:27,201 INFO |test_whenIncompatiblePartitionGroups| - [MetricsConfigHelper] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled 18:06:27,220 INFO |test_whenIncompatiblePartitionGroups| - [Node] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Using TCP/IP discovery 18:06:27,220 WARN |test_whenIncompatiblePartitionGroups| - [CPSubsystem] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees. 18:06:27,234 INFO |test_whenIncompatiblePartitionGroups| - [Diagnostics] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments. 18:06:27,234 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is STARTING 18:06:27,242 INFO |test_whenIncompatiblePartitionGroups| - [TcpIpJoiner] hz.quizzical_shockley.cached.thread-3 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5703 is added to the blacklist. 18:06:27,243 INFO |test_whenIncompatiblePartitionGroups| - [TcpIpJoiner] hz.quizzical_shockley.cached.thread-2 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is added to the blacklist. 18:06:28,243 INFO |test_whenIncompatiblePartitionGroups| - [ClusterService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Members {size:1, ver:1} [ Member [127.0.0.1]:5701 - cb0989c3-7be0-4840-af83-ee83f5f79dba this ] 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is STARTED 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [MetricsConfigHelper] test_whenIncompatiblePartitionGroups - [LOCAL] [dev] [5.2.2-SNAPSHOT] Overridden metrics configuration with system property 'hazelcast.metrics.collection.frequency'='1' -> 'MetricsConfig.collectionFrequencySeconds'='1' 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [logo] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] + + o o o o---o o----o o o---o o o----o o--o--o + + + + | | / \ / | | / / \ | | + + + + + o----o o o o o----o | o o o o----o | + + + + | | / \ / | | \ / \ | | + + o o o o o---o o----o o----o o---o o o o----o o 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Copyright (c) 2008-2022, Hazelcast, Inc. All Rights Reserved. 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Hazelcast Platform 5.2.2-SNAPSHOT (20230111 - c1cb576) starting at [127.0.0.1]:5702 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Cluster name: dev 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Integrity Checker is disabled. Fail-fast on corrupted executables will not be performed. For more information, see the documentation for Integrity Checker. 18:06:28,244 INFO |test_whenIncompatiblePartitionGroups| - [system] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] The Jet engine is disabled. To enable the Jet engine on the members, do one of the following: - Change member config using Java API: config.getJetConfig().setEnabled(true) - Change XML/YAML configuration property: Set hazelcast.jet.enabled to true - Add system property: -Dhz.jet.enabled=true (for Hazelcast embedded, works only when loading config via Config.load) - Add environment variable: HZ_JET_ENABLED=true (recommended when running container image. For Hazelcast embedded, works only when loading config via Config.load) 18:06:28,246 INFO |test_whenIncompatiblePartitionGroups| - [MetricsConfigHelper] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Collecting debug metrics and sending to diagnostics is enabled 18:06:28,250 INFO |test_whenIncompatiblePartitionGroups| - [Node] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Using TCP/IP discovery 18:06:28,250 WARN |test_whenIncompatiblePartitionGroups| - [CPSubsystem] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] CP Subsystem is not enabled. CP data structures will operate in UNSAFE mode! Please note that UNSAFE mode will not provide strong consistency guarantees. 18:06:28,254 INFO |test_whenIncompatiblePartitionGroups| - [Diagnostics] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments. 18:06:28,254 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is STARTING 18:06:28,255 INFO |test_whenIncompatiblePartitionGroups| - [TcpIpJoiner] hz.eloquent_shockley.cached.thread-1 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5703 is added to the blacklist. 18:06:28,255 INFO |test_whenIncompatiblePartitionGroups| - [TcpServerConnection] hz.quizzical_shockley.IO.thread-in-2 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Initialized new cluster connection between /127.0.0.1:5701 and /127.0.0.1:50817 18:06:28,255 INFO |test_whenIncompatiblePartitionGroups| - [TcpServerConnection] hz.eloquent_shockley.IO.thread-in-1 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Initialized new cluster connection between /0.0.0.0:50817 and /127.0.0.1:5701 18:06:28,264 WARN |test_whenIncompatiblePartitionGroups| - [ClusterJoinManager] hz.eloquent_shockley.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Invalid join request from [127.0.0.1]:5701, cause: Incompatible partition groups! expected: HOST_AWARE, found: CUSTOM 18:06:28,264 WARN |test_whenIncompatiblePartitionGroups| - [ClusterJoinManager] hz.eloquent_shockley.priority-generic-operation.thread-0 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Received an invalid join request from [127.0.0.1]:5701, cause: Incompatible partition groups! expected: HOST_AWARE, found: CUSTOM 18:06:28,264 ERROR |test_whenIncompatiblePartitionGroups| - [cluster] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Node could not join cluster. A Configuration mismatch was detected: Incompatible partition groups! expected: HOST_AWARE, found: CUSTOM Node is going to shutdown now! 18:06:28,264 WARN |test_whenIncompatiblePartitionGroups| - [Node] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Terminating forcefully... 18:06:28,264 INFO |test_whenIncompatiblePartitionGroups| - [Node] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Shutting down connection manager... 18:06:28,264 INFO |test_whenIncompatiblePartitionGroups| - [TcpServerConnection] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Connection[id=1, /127.0.0.1:5701->/127.0.0.1:50817, qualifier=null, endpoint=[127.0.0.1]:5702, remoteUuid=04b55ecd-d879-4415-90f8-cb6124e46ed0, alive=false, connectionType=MEMBER, planeIndex=0] closed. Reason: TcpServer is stopping 18:06:28,264 INFO |test_whenIncompatiblePartitionGroups| - [TcpServerConnection] hz.eloquent_shockley.IO.thread-in-1 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Connection[id=1, /0.0.0.0:50817->/127.0.0.1:5701, qualifier=null, endpoint=[127.0.0.1]:5701, remoteUuid=cb0989c3-7be0-4840-af83-ee83f5f79dba, alive=false, connectionType=MEMBER, planeIndex=0] closed. Reason: Connection closed by the other side 18:06:28,264 INFO |test_whenIncompatiblePartitionGroups| - [Node] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Shutting down node engine... 18:06:28,266 INFO |test_whenIncompatiblePartitionGroups| - [NodeExtension] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Destroying node NodeExtension. 18:06:28,266 INFO |test_whenIncompatiblePartitionGroups| - [Node] hz.quizzical_shockley.generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Hazelcast Shutdown is completed in 2 ms. 18:06:28,355 INFO |test_whenIncompatiblePartitionGroups| - [TcpIpJoiner] hz.eloquent_shockley.cached.thread-1 - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is added to the blacklist. 18:06:29,255 INFO |test_whenIncompatiblePartitionGroups| - [ClusterService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Members {size:1, ver:1} [ Member [127.0.0.1]:5702 - 04b55ecd-d879-4415-90f8-cb6124e46ed0 this ] 18:06:29,255 WARN |test_whenIncompatiblePartitionGroups| - [Node] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Config seed port is 5701 and cluster size is 1. Some of the ports seem occupied! 18:06:29,256 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] test_whenIncompatiblePartitionGroups - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is STARTED 18:06:29,860 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is SHUTTING_DOWN 18:06:29,860 WARN |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Terminating forcefully... 18:06:29,860 INFO |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Shutting down connection manager... 18:06:29,861 INFO |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Shutting down node engine... 18:06:29,863 INFO |test_whenIncompatiblePartitionGroups| - [NodeExtension] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Destroying node NodeExtension. 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] Hazelcast Shutdown is completed in 6 ms. 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] main - [127.0.0.1]:5702 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5702 is SHUTDOWN 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] main - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is SHUTTING_DOWN 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [Node] main - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] Node is already shutting down... Waiting for shutdown process to complete... 18:06:29,866 INFO |test_whenIncompatiblePartitionGroups| - [LifecycleService] main - [127.0.0.1]:5701 [dev] [5.2.2-SNAPSHOT] [127.0.0.1]:5701 is SHUTDOWN BuildInfo right after test_whenIncompatiblePartitionGroups(com.hazelcast.cluster.TcpIpJoinTest): BuildInfo{version='5.2.2-SNAPSHOT', build='20230111', buildNumber=20230111, revision=c1cb576, enterprise=false, serializationVersion=1} Hiccups measured while running test 'test_whenIncompatiblePartitionGroups(com.hazelcast.cluster.TcpIpJoinTest):' 18:06:25, accumulated pauses: 710 ms, max pause: 595 ms, pauses over 1000 ms: 0 No metrics recorded during the test ``` </details> Standard output can be found here - https://s3.console.aws.amazon.com/s3/buckets/j-artifacts/Hazelcast-5.maintenance-ZuluJDK17-nightly/46/
non_process
com hazelcast cluster tcpipjointest test whenincompatiblepartitiongroups z commit failed on openjdk stacktrace java lang assertionerror at org junit assert fail assert java at org junit assert fail assert java at com hazelcast cluster abstractjointest assertincompatible abstractjointest java at com hazelcast cluster tcpipjointest test whenincompatiblepartitiongroups tcpipjointest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java base java util concurrent futuretask run futuretask java at java base java lang thread run thread java standard output finished running test test whenhostunresolvable in seconds started running test test whenincompatiblepartitiongroups info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups overridden metrics configuration with system property hazelcast metrics collection frequency metricsconfig collectionfrequencyseconds info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups copyright c hazelcast inc all rights reserved info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups hazelcast platform snapshot starting at info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups cluster name dev info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups integrity checker is disabled fail fast on corrupted executables will not be performed for more information see the documentation for integrity checker info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups the jet engine is disabled to enable the jet engine on the members do one of the following change member config using java api config getjetconfig setenabled true change xml yaml configuration property set hazelcast jet enabled to true add system property dhz jet enabled true for hazelcast embedded works only when loading config via config load add environment variable hz jet enabled true recommended when running container image for hazelcast embedded works only when loading config via config load info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups collecting debug metrics and sending to diagnostics is enabled info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups using tcp ip discovery warn test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups cp subsystem is not enabled cp data structures will operate in unsafe mode please note that unsafe mode will not provide strong consistency guarantees info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups diagnostics disabled to enable add dhazelcast diagnostics enabled true to the jvm arguments info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups is starting info test whenincompatiblepartitiongroups hz quizzical shockley cached thread is added to the blacklist info test whenincompatiblepartitiongroups hz quizzical shockley cached thread is added to the blacklist info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups members size ver member this info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups is started info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups overridden metrics configuration with system property hazelcast metrics collection frequency metricsconfig collectionfrequencyseconds info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups copyright c hazelcast inc all rights reserved info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups hazelcast platform snapshot starting at info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups cluster name dev info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups integrity checker is disabled fail fast on corrupted executables will not be performed for more information see the documentation for integrity checker info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups the jet engine is disabled to enable the jet engine on the members do one of the following change member config using java api config getjetconfig setenabled true change xml yaml configuration property set hazelcast jet enabled to true add system property dhz jet enabled true for hazelcast embedded works only when loading config via config load add environment variable hz jet enabled true recommended when running container image for hazelcast embedded works only when loading config via config load info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups collecting debug metrics and sending to diagnostics is enabled info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups using tcp ip discovery warn test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups cp subsystem is not enabled cp data structures will operate in unsafe mode please note that unsafe mode will not provide strong consistency guarantees info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups diagnostics disabled to enable add dhazelcast diagnostics enabled true to the jvm arguments info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups is starting info test whenincompatiblepartitiongroups hz eloquent shockley cached thread is added to the blacklist info test whenincompatiblepartitiongroups hz quizzical shockley io thread in initialized new cluster connection between and info test whenincompatiblepartitiongroups hz eloquent shockley io thread in initialized new cluster connection between and warn test whenincompatiblepartitiongroups hz eloquent shockley priority generic operation thread invalid join request from cause incompatible partition groups expected host aware found custom warn test whenincompatiblepartitiongroups hz eloquent shockley priority generic operation thread received an invalid join request from cause incompatible partition groups expected host aware found custom error test whenincompatiblepartitiongroups hz quizzical shockley generic operation thread node could not join cluster a configuration mismatch was detected incompatible partition groups expected host aware found custom node is going to shutdown now warn test whenincompatiblepartitiongroups hz quizzical shockley generic operation thread terminating forcefully info test whenincompatiblepartitiongroups hz quizzical shockley generic operation thread shutting down connection manager info test whenincompatiblepartitiongroups hz quizzical shockley generic operation thread connection remoteuuid alive false connectiontype member planeindex closed reason tcpserver is stopping info test whenincompatiblepartitiongroups hz eloquent shockley io thread in connection remoteuuid alive false connectiontype member planeindex closed reason connection closed by the other side info test whenincompatiblepartitiongroups hz quizzical shockley generic operation thread shutting down node engine info test whenincompatiblepartitiongroups hz quizzical shockley generic operation thread destroying node nodeextension info test whenincompatiblepartitiongroups hz quizzical shockley generic operation thread hazelcast shutdown is completed in ms info test whenincompatiblepartitiongroups hz eloquent shockley cached thread is added to the blacklist info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups members size ver member this warn test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups config seed port is and cluster size is some of the ports seem occupied info test whenincompatiblepartitiongroups test whenincompatiblepartitiongroups is started info test whenincompatiblepartitiongroups main is shutting down warn test whenincompatiblepartitiongroups main terminating forcefully info test whenincompatiblepartitiongroups main shutting down connection manager info test whenincompatiblepartitiongroups main shutting down node engine info test whenincompatiblepartitiongroups main destroying node nodeextension info test whenincompatiblepartitiongroups main hazelcast shutdown is completed in ms info test whenincompatiblepartitiongroups main is shutdown info test whenincompatiblepartitiongroups main is shutting down info test whenincompatiblepartitiongroups main node is already shutting down waiting for shutdown process to complete info test whenincompatiblepartitiongroups main is shutdown buildinfo right after test whenincompatiblepartitiongroups com hazelcast cluster tcpipjointest buildinfo version snapshot build buildnumber revision enterprise false serializationversion hiccups measured while running test test whenincompatiblepartitiongroups com hazelcast cluster tcpipjointest accumulated pauses ms max pause ms pauses over ms no metrics recorded during the test standard output can be found here
0
241,632
26,256,867,552
IssuesEvent
2023-01-06 02:04:17
vlaship/build-docker-image
https://api.github.com/repos/vlaship/build-docker-image
opened
CVE-2022-22950 (Medium) detected in spring-expression-5.1.10.RELEASE.jar
security vulnerability
## CVE-2022-22950 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-expression-5.1.10.RELEASE.jar</b></p></summary> <p>Spring Expression Language (SpEL)</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.1.10.RELEASE/7dc657305af97015f96135bfbb5036caf8829070/spring-expression-5.1.10.RELEASE.jar,/root/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.1.10.RELEASE/7dc657305af97015f96135bfbb5036caf8829070/spring-expression-5.1.10.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.9.RELEASE.jar (Root Library) - spring-boot-starter-2.1.9.RELEASE.jar - spring-boot-2.1.9.RELEASE.jar - spring-context-5.1.10.RELEASE.jar - :x: **spring-expression-5.1.10.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vlaship/build-docker-image/commit/b43ef51625fbadf06596247b05de8136106b7833">b43ef51625fbadf06596247b05de8136106b7833</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> n Spring Framework versions 5.3.0 - 5.3.16 and older unsupported versions, it is possible for a user to provide a specially crafted SpEL expression that may cause a denial of service condition. <p>Publish Date: 2022-04-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22950>CVE-2022-22950</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22950">https://tanzu.vmware.com/security/cve-2022-22950</a></p> <p>Release Date: 2022-04-01</p> <p>Fix Resolution (org.springframework:spring-expression): 5.2.20.RELEASE</p> <p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-web): 2.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-22950 (Medium) detected in spring-expression-5.1.10.RELEASE.jar - ## CVE-2022-22950 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-expression-5.1.10.RELEASE.jar</b></p></summary> <p>Spring Expression Language (SpEL)</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.1.10.RELEASE/7dc657305af97015f96135bfbb5036caf8829070/spring-expression-5.1.10.RELEASE.jar,/root/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.1.10.RELEASE/7dc657305af97015f96135bfbb5036caf8829070/spring-expression-5.1.10.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.9.RELEASE.jar (Root Library) - spring-boot-starter-2.1.9.RELEASE.jar - spring-boot-2.1.9.RELEASE.jar - spring-context-5.1.10.RELEASE.jar - :x: **spring-expression-5.1.10.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vlaship/build-docker-image/commit/b43ef51625fbadf06596247b05de8136106b7833">b43ef51625fbadf06596247b05de8136106b7833</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> n Spring Framework versions 5.3.0 - 5.3.16 and older unsupported versions, it is possible for a user to provide a specially crafted SpEL expression that may cause a denial of service condition. <p>Publish Date: 2022-04-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22950>CVE-2022-22950</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22950">https://tanzu.vmware.com/security/cve-2022-22950</a></p> <p>Release Date: 2022-04-01</p> <p>Fix Resolution (org.springframework:spring-expression): 5.2.20.RELEASE</p> <p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-web): 2.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in spring expression release jar cve medium severity vulnerability vulnerable library spring expression release jar spring expression language spel library home page a href path to dependency file build gradle path to vulnerable library root gradle caches modules files org springframework spring expression release spring expression release jar root gradle caches modules files org springframework spring expression release spring expression release jar dependency hierarchy spring boot starter web release jar root library spring boot starter release jar spring boot release jar spring context release jar x spring expression release jar vulnerable library found in head commit a href vulnerability details n spring framework versions and older unsupported versions it is possible for a user to provide a specially crafted spel expression that may cause a denial of service condition publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring expression release direct dependency fix resolution org springframework boot spring boot starter web step up your open source security game with mend
0
10,531
13,310,552,358
IssuesEvent
2020-08-26 06:46:19
e4exp/paper_manager_abstract
https://api.github.com/repos/e4exp/paper_manager_abstract
opened
Top-down Tree Long Short-Term Memory Networks
2016 Natural Language Processing Recurrent Neural Network Tree Structure
* https://arxiv.org/abs/1511.00060 * NAACL 2016 より複雑な計算単位を持つリカレントニューラルネットワークの一種であるLong Short-Term Memory (LSTM)ネットワークは、様々なシーケンスモデリングタスクへの応用に成功している。 本論文では、LSTMに基づくニューラルネットワークモデルであるTree Long Short-Short-Term Memory (TreeLSTM)を開発し、線形シーケンスではなくツリーを予測するように設計した。 TreeLSTMは、文の依存関係木の生成確率を推定することで、文の確率を定義する。 各時間ステップでは、生成された部分木の表現に基づいてノードが生成される。 さらに、左右の依存関係の相関を明示的に表現することで、TreeLSTM のモデリング能力を向上させている。 このモデルをMSR文完成課題に適用することで、現在の技術水準を超える結果を得ることができた。 また、依存性解析のリランキングについても、競争力のある結果が得られたことを報告する。
1.0
Top-down Tree Long Short-Term Memory Networks - * https://arxiv.org/abs/1511.00060 * NAACL 2016 より複雑な計算単位を持つリカレントニューラルネットワークの一種であるLong Short-Term Memory (LSTM)ネットワークは、様々なシーケンスモデリングタスクへの応用に成功している。 本論文では、LSTMに基づくニューラルネットワークモデルであるTree Long Short-Short-Term Memory (TreeLSTM)を開発し、線形シーケンスではなくツリーを予測するように設計した。 TreeLSTMは、文の依存関係木の生成確率を推定することで、文の確率を定義する。 各時間ステップでは、生成された部分木の表現に基づいてノードが生成される。 さらに、左右の依存関係の相関を明示的に表現することで、TreeLSTM のモデリング能力を向上させている。 このモデルをMSR文完成課題に適用することで、現在の技術水準を超える結果を得ることができた。 また、依存性解析のリランキングについても、競争力のある結果が得られたことを報告する。
process
top down tree long short term memory networks naacl より複雑な計算単位を持つリカレントニューラルネットワークの一種であるlong short term memory lstm ネットワークは、様々なシーケンスモデリングタスクへの応用に成功している。 本論文では、lstmに基づくニューラルネットワークモデルであるtree long short short term memory treelstm を開発し、線形シーケンスではなくツリーを予測するように設計した。 treelstmは、文の依存関係木の生成確率を推定することで、文の確率を定義する。 各時間ステップでは、生成された部分木の表現に基づいてノードが生成される。 さらに、左右の依存関係の相関を明示的に表現することで、treelstm のモデリング能力を向上させている。 このモデルをmsr文完成課題に適用することで、現在の技術水準を超える結果を得ることができた。 また、依存性解析のリランキングについても、競争力のある結果が得られたことを報告する。
1
82,505
23,802,783,252
IssuesEvent
2022-09-03 15:07:34
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: The Tablet(large) and the Tablet icon looks similar in the canvas size change section
Bug Good First Issue Design System Pod Property Pane QA Low Release UI Builders Pod Responsive Viewport
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior It is observed that the same icon is displayed for Tablet(Large) and Tablet option in the canvas size change section <img width="1680" alt="Screenshot 2021-11-24 at 12 34 10 PM" src="https://user-images.githubusercontent.com/35128442/143191136-187da401-37cc-49c4-ad77-659eb5a977c4.png"> ### Steps To Reproduce 1.Launch Appmsith app in release environment 2.Drag and drop any widget 3.Click any where on the canvas outside of any widget 4.On the property doc pane - Canvas size section will be displayed 5.Check the icon of Tablet(Large) and Tablet ### Environment Release ### Version Cloud
1.0
[Bug]: The Tablet(large) and the Tablet icon looks similar in the canvas size change section - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior It is observed that the same icon is displayed for Tablet(Large) and Tablet option in the canvas size change section <img width="1680" alt="Screenshot 2021-11-24 at 12 34 10 PM" src="https://user-images.githubusercontent.com/35128442/143191136-187da401-37cc-49c4-ad77-659eb5a977c4.png"> ### Steps To Reproduce 1.Launch Appmsith app in release environment 2.Drag and drop any widget 3.Click any where on the canvas outside of any widget 4.On the property doc pane - Canvas size section will be displayed 5.Check the icon of Tablet(Large) and Tablet ### Environment Release ### Version Cloud
non_process
the tablet large and the tablet icon looks similar in the canvas size change section is there an existing issue for this i have searched the existing issues current behavior it is observed that the same icon is displayed for tablet large and tablet option in the canvas size change section img width alt screenshot at pm src steps to reproduce launch appmsith app in release environment drag and drop any widget click any where on the canvas outside of any widget on the property doc pane canvas size section will be displayed check the icon of tablet large and tablet environment release version cloud
0
7,799
10,958,671,680
IssuesEvent
2019-11-27 09:52:08
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
Update standard process model for MVP to contain only one data task
area/process kind/user-story solution/app-frontend status/ready-for-specification team/nusse
**Functional architect/designer:** @-mention **Technical architect:** @-mention **Description** As an application developer I want to create an app where the end user can fill in a form and complete the process without having to start a separate submit. Hence, provide a simple BPMN file with only one task. **Sketch (if relevant)** (Screenshot and link to Figma, make sure your sketch is public!) **Navigation from/to (if relevant)** Altinn studio **Technical considerations** Input (beyond tasks) on how the user story should be solved can be put here. **Acceptance criterea** - What is allowed/not allowed - Validations - Error messages and warnings - ... **Tasks** - [x] Create a new bpmn file with one data task - [x] Document the solution - [x] Create development tasks and add "ready for development" label - [x] Remove "ready for specification" label once - [x] Test Design - [x] Test Execution
1.0
Update standard process model for MVP to contain only one data task - **Functional architect/designer:** @-mention **Technical architect:** @-mention **Description** As an application developer I want to create an app where the end user can fill in a form and complete the process without having to start a separate submit. Hence, provide a simple BPMN file with only one task. **Sketch (if relevant)** (Screenshot and link to Figma, make sure your sketch is public!) **Navigation from/to (if relevant)** Altinn studio **Technical considerations** Input (beyond tasks) on how the user story should be solved can be put here. **Acceptance criterea** - What is allowed/not allowed - Validations - Error messages and warnings - ... **Tasks** - [x] Create a new bpmn file with one data task - [x] Document the solution - [x] Create development tasks and add "ready for development" label - [x] Remove "ready for specification" label once - [x] Test Design - [x] Test Execution
process
update standard process model for mvp to contain only one data task functional architect designer mention technical architect mention description as an application developer i want to create an app where the end user can fill in a form and complete the process without having to start a separate submit hence provide a simple bpmn file with only one task sketch if relevant screenshot and link to figma make sure your sketch is public navigation from to if relevant altinn studio technical considerations input beyond tasks on how the user story should be solved can be put here acceptance criterea what is allowed not allowed validations error messages and warnings tasks create a new bpmn file with one data task document the solution create development tasks and add ready for development label remove ready for specification label once test design test execution
1
20,429
27,091,982,374
IssuesEvent
2023-02-14 21:58:24
dotnet/fabricbot-config
https://api.github.com/repos/dotnet/fabricbot-config
closed
Bot is not removing items from columns when we transfer them to another repo
enhancement process
There was an issue in the runtime repo that we transferred to another repo, and the issue was still showing up in the untriaged column. Is this something that can be implemented? Is the bot able to see a notification when an issue/PR gets transferred to another repo? Example: Msbuild issue 8319, which we transferred from runtime.
1.0
Bot is not removing items from columns when we transfer them to another repo - There was an issue in the runtime repo that we transferred to another repo, and the issue was still showing up in the untriaged column. Is this something that can be implemented? Is the bot able to see a notification when an issue/PR gets transferred to another repo? Example: Msbuild issue 8319, which we transferred from runtime.
process
bot is not removing items from columns when we transfer them to another repo there was an issue in the runtime repo that we transferred to another repo and the issue was still showing up in the untriaged column is this something that can be implemented is the bot able to see a notification when an issue pr gets transferred to another repo example msbuild issue which we transferred from runtime
1
23,569
4,028,297,504
IssuesEvent
2016-05-18 05:20:19
Mr-Kumar-Abhishek/zuzeelik
https://api.github.com/repos/Mr-Kumar-Abhishek/zuzeelik
opened
Test builds for different versions MAC OS X
testing
Tests build for different versions of MAC OS X. Either use a Continuous Integration / Continuous Deployment system or test on MAC OS X machines. (Refer issue #2)
1.0
Test builds for different versions MAC OS X - Tests build for different versions of MAC OS X. Either use a Continuous Integration / Continuous Deployment system or test on MAC OS X machines. (Refer issue #2)
non_process
test builds for different versions mac os x tests build for different versions of mac os x either use a continuous integration continuous deployment system or test on mac os x machines refer issue
0
814,062
30,484,805,680
IssuesEvent
2023-07-18 00:34:33
AdaptiveHaptics/AdapticsDesigner
https://api.github.com/repos/AdaptiveHaptics/AdapticsDesigner
closed
send next eval params along with update_playstart
bug high-priority
causes jump(desync) on playstart when connecting to server mid playback after a cjump has been taken. Also causes playstart in the middle of pattern, after a cjump, to not work as expected.
1.0
send next eval params along with update_playstart - causes jump(desync) on playstart when connecting to server mid playback after a cjump has been taken. Also causes playstart in the middle of pattern, after a cjump, to not work as expected.
non_process
send next eval params along with update playstart causes jump desync on playstart when connecting to server mid playback after a cjump has been taken also causes playstart in the middle of pattern after a cjump to not work as expected
0
8,971
12,086,691,132
IssuesEvent
2020-04-18 11:14:49
CGAL/cgal
https://api.github.com/repos/CGAL/cgal
closed
About normal computation?
Pkg::Point_set_processing_3 question
I would like to know about normal computation. For my data, I used pca_estimate_normal, but it takes so long time, around one night. When I used another library, where they use KDTreeSearch, it was faster. What is the difference between pca_estimate_normals and KDTreeSearch based algorithms? As for example KDTreeSearch (open3d): > pcd.estimate_normals( o3d.geometry.KDTreeSearchParamHybrid(radius=0.1, max_nn=30)) Is there any way to calculate normals using KDTreeSearch in CGAL? My enivornment: Windows 10 Visual studio 2017 CGAL 5.0.2
1.0
About normal computation? - I would like to know about normal computation. For my data, I used pca_estimate_normal, but it takes so long time, around one night. When I used another library, where they use KDTreeSearch, it was faster. What is the difference between pca_estimate_normals and KDTreeSearch based algorithms? As for example KDTreeSearch (open3d): > pcd.estimate_normals( o3d.geometry.KDTreeSearchParamHybrid(radius=0.1, max_nn=30)) Is there any way to calculate normals using KDTreeSearch in CGAL? My enivornment: Windows 10 Visual studio 2017 CGAL 5.0.2
process
about normal computation i would like to know about normal computation for my data i used pca estimate normal but it takes so long time around one night when i used another library where they use kdtreesearch it was faster what is the difference between pca estimate normals and kdtreesearch based algorithms as for example kdtreesearch pcd estimate normals geometry kdtreesearchparamhybrid radius max nn is there any way to calculate normals using kdtreesearch in cgal my enivornment windows visual studio cgal
1
19,074
25,103,846,970
IssuesEvent
2022-11-08 15:20:08
USAID-SA-SI/COVID19
https://api.github.com/repos/USAID-SA-SI/COVID19
opened
Public dataset | mon_yr format not aligned to USAID format
bug Public Dataset Processing
they set full dataset to first day of each month
1.0
Public dataset | mon_yr format not aligned to USAID format - they set full dataset to first day of each month
process
public dataset mon yr format not aligned to usaid format they set full dataset to first day of each month
1
67,120
8,074,841,804
IssuesEvent
2018-08-07 01:49:29
vavr-io/vavr
https://api.github.com/repos/vavr-io/vavr
closed
Remove Either projections in favor of swap
!BACK-COMPAT design/refactoring/improvement help wanted «vavr-controlx»
Currently we support both: projections and the right-biased approach. I think we can abandon the projections for the 1.0.0 release. Details see [here](http://robsscala.blogspot.de/2012/06/fixing-scalaeither-unbiased-vs-biased.html)
1.0
Remove Either projections in favor of swap - Currently we support both: projections and the right-biased approach. I think we can abandon the projections for the 1.0.0 release. Details see [here](http://robsscala.blogspot.de/2012/06/fixing-scalaeither-unbiased-vs-biased.html)
non_process
remove either projections in favor of swap currently we support both projections and the right biased approach i think we can abandon the projections for the release details see
0
5,134
7,920,042,020
IssuesEvent
2018-07-04 20:43:31
UnbFeelings/unb-feelings-docs
https://api.github.com/repos/UnbFeelings/unb-feelings-docs
closed
Backlog de Backlogs
Desenvolvimento Processo wiki
### Épico 01- Apoio à um perfil ..... #### Feature 01 (E01FE01) - Publicar Apoio ..... #### Feature 02 (E01FE02) - Gerenciar Apoio ...... ### Épico 02 - Bloqueio ...... #### Feature 01 (E02FE01) - Bloquear perfil ...... #### Feature 02 (E02FE02) - Gerenciar perfis bloqueados .....
1.0
Backlog de Backlogs - ### Épico 01- Apoio à um perfil ..... #### Feature 01 (E01FE01) - Publicar Apoio ..... #### Feature 02 (E01FE02) - Gerenciar Apoio ...... ### Épico 02 - Bloqueio ...... #### Feature 01 (E02FE01) - Bloquear perfil ...... #### Feature 02 (E02FE02) - Gerenciar perfis bloqueados .....
process
backlog de backlogs épico apoio à um perfil feature publicar apoio feature gerenciar apoio épico bloqueio feature bloquear perfil feature gerenciar perfis bloqueados
1
499,896
14,481,606,899
IssuesEvent
2020-12-10 12:52:52
weaveworks/eksctl
https://api.github.com/repos/weaveworks/eksctl
closed
Arm64 build support
kind/feature priority/backlog
**Why do you want this feature?** With the GA of EKS support for Graviton2 last month, it would be nice to have the `eksctl` binary and docker image built for Arm64. As `eksctl` is 97%+ written in Go, this should be mostly a build/script/packaging change rather than substantive code changes. **What feature/behavior/change do you want?** - [ ] The ability to compile eksctl from source on Arm64 - [ ] Arm64 binaries generated automatically by the existing build processes - [ ] Multi-arch images on the `weaveworks/eksctl` dockerhub
1.0
Arm64 build support - **Why do you want this feature?** With the GA of EKS support for Graviton2 last month, it would be nice to have the `eksctl` binary and docker image built for Arm64. As `eksctl` is 97%+ written in Go, this should be mostly a build/script/packaging change rather than substantive code changes. **What feature/behavior/change do you want?** - [ ] The ability to compile eksctl from source on Arm64 - [ ] Arm64 binaries generated automatically by the existing build processes - [ ] Multi-arch images on the `weaveworks/eksctl` dockerhub
non_process
build support why do you want this feature with the ga of eks support for last month it would be nice to have the eksctl binary and docker image built for as eksctl is written in go this should be mostly a build script packaging change rather than substantive code changes what feature behavior change do you want the ability to compile eksctl from source on binaries generated automatically by the existing build processes multi arch images on the weaveworks eksctl dockerhub
0
16,668
21,771,954,986
IssuesEvent
2022-05-13 09:56:43
OpenDataScotland/the_od_bods
https://api.github.com/repos/OpenDataScotland/the_od_bods
opened
Populate WhatDoTheyKnow IDs
good first issue data processing
![image](https://user-images.githubusercontent.com/46202639/168259557-0737f984-0c8e-4e30-a878-278a470f7432.png) We added a new field to orgs for WhatDoTheyKnow IDs so we can link to to their list of FOI requests published on https://www.whatdotheyknow.com/. However, not all orgs have this value populated. This could be a good exercise for someone to go through all the orgs in opendata.scot and populate them accordingly. To populate, the user needs to find the relevant organisation on WDTK (e.g. https://www.whatdotheyknow.com/body/aberdeen_city_council for Aberdeen City Council), take the ID which is the text after `https://www.whatdotheyknow.com/body/` (i.e. aberdeen_city_council) and copy that into the WhatDoTheyKnow organization ID field for the org.
1.0
Populate WhatDoTheyKnow IDs - ![image](https://user-images.githubusercontent.com/46202639/168259557-0737f984-0c8e-4e30-a878-278a470f7432.png) We added a new field to orgs for WhatDoTheyKnow IDs so we can link to to their list of FOI requests published on https://www.whatdotheyknow.com/. However, not all orgs have this value populated. This could be a good exercise for someone to go through all the orgs in opendata.scot and populate them accordingly. To populate, the user needs to find the relevant organisation on WDTK (e.g. https://www.whatdotheyknow.com/body/aberdeen_city_council for Aberdeen City Council), take the ID which is the text after `https://www.whatdotheyknow.com/body/` (i.e. aberdeen_city_council) and copy that into the WhatDoTheyKnow organization ID field for the org.
process
populate whatdotheyknow ids we added a new field to orgs for whatdotheyknow ids so we can link to to their list of foi requests published on however not all orgs have this value populated this could be a good exercise for someone to go through all the orgs in opendata scot and populate them accordingly to populate the user needs to find the relevant organisation on wdtk e g for aberdeen city council take the id which is the text after i e aberdeen city council and copy that into the whatdotheyknow organization id field for the org
1
16,223
20,752,896,564
IssuesEvent
2022-03-15 09:27:07
hoprnet/hoprnet
https://api.github.com/repos/hoprnet/hoprnet
closed
Update processes in relation to documentation
new issue processes
<!--- Please DO NOT remove the automatically added 'new issue' label --> <!--- Provide a general summary of the issue in the Title above --> - Add docs check step as part of the release process - Move back to rotating release masters
1.0
Update processes in relation to documentation - <!--- Please DO NOT remove the automatically added 'new issue' label --> <!--- Provide a general summary of the issue in the Title above --> - Add docs check step as part of the release process - Move back to rotating release masters
process
update processes in relation to documentation add docs check step as part of the release process move back to rotating release masters
1
77,999
27,271,605,898
IssuesEvent
2023-02-22 22:58:13
idaholab/HERON
https://api.github.com/repos/idaholab/HERON
closed
[DEFECT] Taxes & Inflation flags on recurring cashflows don't do anything
defect priority-critical
-------- Defect Description -------- **Describe the defect** Currently, a user cannot prevent taxes from being applied to a recurring cash flow. According to [`TEAL/src/Cashflows.py#L1127`](https://github.com/idaholab/TEAL/blob/devel/src/CashFlows.py#L1127), any recurring cash flow is automatically taxed by default. If a user specifies `taxable=False` on a `<Cashflow>` that is recurring, it will apply the tax rate anyway. In HERON, users have the ability to specify if a cash flow is taxed or not, but in the code, we do not actually pass this information on to TEAL. See [`HERON/src/DispatchManager.py#L335-L340`](https://github.com/idaholab/HERON/blob/devel/src/DispatchManager.py#L335) This becomes an issue, for example, if a user wants to implement a Production Tax Credit (PTC) for a specific resource on a specific component. This is a recurring cash flow that should not be taxed. ##### What did you expect to see happen? ##### What did you see instead? ##### Do you have a suggested fix for the development team? **Describe how to Reproduce** Steps to reproduce the behavior: 1. 2. 3. 4. **Screenshots and Input Files** Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. **Platform (please complete the following information):** - OS: [e.g. iOS] - Version: [e.g. 22] - Dependencies Installation: [CONDA or PIP] ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or task? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
1.0
[DEFECT] Taxes & Inflation flags on recurring cashflows don't do anything - -------- Defect Description -------- **Describe the defect** Currently, a user cannot prevent taxes from being applied to a recurring cash flow. According to [`TEAL/src/Cashflows.py#L1127`](https://github.com/idaholab/TEAL/blob/devel/src/CashFlows.py#L1127), any recurring cash flow is automatically taxed by default. If a user specifies `taxable=False` on a `<Cashflow>` that is recurring, it will apply the tax rate anyway. In HERON, users have the ability to specify if a cash flow is taxed or not, but in the code, we do not actually pass this information on to TEAL. See [`HERON/src/DispatchManager.py#L335-L340`](https://github.com/idaholab/HERON/blob/devel/src/DispatchManager.py#L335) This becomes an issue, for example, if a user wants to implement a Production Tax Credit (PTC) for a specific resource on a specific component. This is a recurring cash flow that should not be taxed. ##### What did you expect to see happen? ##### What did you see instead? ##### Do you have a suggested fix for the development team? **Describe how to Reproduce** Steps to reproduce the behavior: 1. 2. 3. 4. **Screenshots and Input Files** Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. **Platform (please complete the following information):** - OS: [e.g. iOS] - Version: [e.g. 22] - Dependencies Installation: [CONDA or PIP] ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or task? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [x] 1. If the issue is a defect, is the defect fixed? - [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
non_process
taxes inflation flags on recurring cashflows don t do anything defect description describe the defect currently a user cannot prevent taxes from being applied to a recurring cash flow according to any recurring cash flow is automatically taxed by default if a user specifies taxable false on a that is recurring it will apply the tax rate anyway in heron users have the ability to specify if a cash flow is taxed or not but in the code we do not actually pass this information on to teal see this becomes an issue for example if a user wants to implement a production tax credit ptc for a specific resource on a specific component this is a recurring cash flow that should not be taxed what did you expect to see happen what did you see instead do you have a suggested fix for the development team describe how to reproduce steps to reproduce the behavior screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
0
12,052
14,739,161,440
IssuesEvent
2021-01-07 06:37:24
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Phoenix Synergy - Error Page
anc-process anp-1 ant-bug ant-parent/primary has attachment
In GitLab by @kdjstudios on Aug 31, 2018, 08:02 **Submitted by:** "Jesus Corchado" <jesus.corchado@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-08-31-35869/conversation **Server:** Internal **Client/Site:** Phoenix Synergy **Account:** NA **Issue:** I just clicked finalize to the billing for Phoenix Synergy’s weekly cycle and got the error page. When I try to go back, I continue to get this page and unable to access it. ![image](/uploads/ddef2080e409edb2cc710e4e12796980/image.png)
1.0
Phoenix Synergy - Error Page - In GitLab by @kdjstudios on Aug 31, 2018, 08:02 **Submitted by:** "Jesus Corchado" <jesus.corchado@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-08-31-35869/conversation **Server:** Internal **Client/Site:** Phoenix Synergy **Account:** NA **Issue:** I just clicked finalize to the billing for Phoenix Synergy’s weekly cycle and got the error page. When I try to go back, I continue to get this page and unable to access it. ![image](/uploads/ddef2080e409edb2cc710e4e12796980/image.png)
process
phoenix synergy error page in gitlab by kdjstudios on aug submitted by jesus corchado helpdesk server internal client site phoenix synergy account na issue i just clicked finalize to the billing for phoenix synergy’s weekly cycle and got the error page when i try to go back i continue to get this page and unable to access it uploads image png
1
2,372
2,920,685,720
IssuesEvent
2015-06-24 20:14:34
Homebrew/linuxbrew
https://api.github.com/repos/Homebrew/linuxbrew
closed
php55: /usr/sbin/apxs: No such file or directory
build-error help wanted
Hi, I try to install php55 with homebrew but I have the following problem: 3. Apache was not built using --enable-so (the apxs usage page is displayed) The output of /usr/sbin/apxs follows: ``` ./configure: line 8426: /usr/sbin/apxs: No such file or directory configure: error: Aborting ``` Thank you for your help.
1.0
php55: /usr/sbin/apxs: No such file or directory - Hi, I try to install php55 with homebrew but I have the following problem: 3. Apache was not built using --enable-so (the apxs usage page is displayed) The output of /usr/sbin/apxs follows: ``` ./configure: line 8426: /usr/sbin/apxs: No such file or directory configure: error: Aborting ``` Thank you for your help.
non_process
usr sbin apxs no such file or directory hi i try to install with homebrew but i have the following problem apache was not built using enable so the apxs usage page is displayed the output of usr sbin apxs follows configure line usr sbin apxs no such file or directory configure error aborting thank you for your help
0
7,292
10,439,833,355
IssuesEvent
2019-09-18 07:21:43
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
closed
Release 0.13.0
RxJava1.x release process
**Release notes** - replacing default protocol HTTP with HTTPS in WalledGardenInternetObservingStrategy - PR #376, issue #323 **Note** `RxJava1.x` branch is deprecated and no longer maintained. This release is done due to community request. **To do list** - [x] update docs - [x] update javadocs - [x] bump library version - [x] upload artifact - [x] close and release artifact - [x] update changelog (after maven sync) - [x] create new github release
1.0
Release 0.13.0 - **Release notes** - replacing default protocol HTTP with HTTPS in WalledGardenInternetObservingStrategy - PR #376, issue #323 **Note** `RxJava1.x` branch is deprecated and no longer maintained. This release is done due to community request. **To do list** - [x] update docs - [x] update javadocs - [x] bump library version - [x] upload artifact - [x] close and release artifact - [x] update changelog (after maven sync) - [x] create new github release
process
release release notes replacing default protocol http with https in walledgardeninternetobservingstrategy pr issue note x branch is deprecated and no longer maintained this release is done due to community request to do list update docs update javadocs bump library version upload artifact close and release artifact update changelog after maven sync create new github release
1
7,988
11,183,359,521
IssuesEvent
2019-12-31 12:57:20
carrascomj/drastic
https://api.github.com/repos/carrascomj/drastic
closed
PRE-PROCESSING: whole pipeline
enhancement pre-processing
Having defined all of the pre-processing functions (embeddings not included), we need to define some API for the user to transform the raw data into a suitable format: * [x] Functions in python that joins all of the pre-processing functions into a pipeline. * [x] Filter big genes too in the pipeline. * [x] Command Line Interface in `/bin` to go from paths to files (FASTA and feature table) to the pre-processed dataframe (maybe a CSV).
1.0
PRE-PROCESSING: whole pipeline - Having defined all of the pre-processing functions (embeddings not included), we need to define some API for the user to transform the raw data into a suitable format: * [x] Functions in python that joins all of the pre-processing functions into a pipeline. * [x] Filter big genes too in the pipeline. * [x] Command Line Interface in `/bin` to go from paths to files (FASTA and feature table) to the pre-processed dataframe (maybe a CSV).
process
pre processing whole pipeline having defined all of the pre processing functions embeddings not included we need to define some api for the user to transform the raw data into a suitable format functions in python that joins all of the pre processing functions into a pipeline filter big genes too in the pipeline command line interface in bin to go from paths to files fasta and feature table to the pre processed dataframe maybe a csv
1
42,602
9,256,821,065
IssuesEvent
2019-03-16 22:39:53
fwcd/KotlinLanguageServer
https://api.github.com/repos/fwcd/KotlinLanguageServer
closed
No suggestion options for enum types if companion object exists on enum class
bug code completion
```kotlin public enum class TokenType(val type: String) { ILLEGAL("ILLEGAL"), EOF("EOF"), /* ... */ RBRACE("{"); companion object { private val tokens: HashMap<String, TokenType> = HashMap() init { for(token in TokenType.values()) { tokens.put(token.type, token) } } public fun get(type: String): TokenType = this.tokens.getOrDefault(type, TokenType.ILLEGAL) } } ``` With this code, if i type `TokenType.`, i get the following options: ![image](https://user-images.githubusercontent.com/18282288/47261019-24bc6000-d4be-11e8-92d9-193d303ffabc.png) If i comment out the companion object, i get the full list of enum types: ![image](https://user-images.githubusercontent.com/18282288/47261023-34d43f80-d4be-11e8-9eb9-ef8cf0bda416.png)
1.0
No suggestion options for enum types if companion object exists on enum class - ```kotlin public enum class TokenType(val type: String) { ILLEGAL("ILLEGAL"), EOF("EOF"), /* ... */ RBRACE("{"); companion object { private val tokens: HashMap<String, TokenType> = HashMap() init { for(token in TokenType.values()) { tokens.put(token.type, token) } } public fun get(type: String): TokenType = this.tokens.getOrDefault(type, TokenType.ILLEGAL) } } ``` With this code, if i type `TokenType.`, i get the following options: ![image](https://user-images.githubusercontent.com/18282288/47261019-24bc6000-d4be-11e8-92d9-193d303ffabc.png) If i comment out the companion object, i get the full list of enum types: ![image](https://user-images.githubusercontent.com/18282288/47261023-34d43f80-d4be-11e8-9eb9-ef8cf0bda416.png)
non_process
no suggestion options for enum types if companion object exists on enum class kotlin public enum class tokentype val type string illegal illegal eof eof rbrace companion object private val tokens hashmap hashmap init for token in tokentype values tokens put token type token public fun get type string tokentype this tokens getordefault type tokentype illegal with this code if i type tokentype i get the following options if i comment out the companion object i get the full list of enum types
0
816,624
30,604,820,274
IssuesEvent
2023-07-22 21:45:18
Bentheburrito/cai
https://api.github.com/repos/Bentheburrito/cai
opened
Metrics: persist duplicate events
low priority data insights
It would be interesting to save and graph duplicate events alongside total events. Grouping by fields like world_id would be interesting as well
1.0
Metrics: persist duplicate events - It would be interesting to save and graph duplicate events alongside total events. Grouping by fields like world_id would be interesting as well
non_process
metrics persist duplicate events it would be interesting to save and graph duplicate events alongside total events grouping by fields like world id would be interesting as well
0
240,394
20,026,517,694
IssuesEvent
2022-02-01 22:00:48
carbon-design-system/carbon-for-ibm-dotcom
https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom
opened
[Test Scenario]: Cloud Masthead with Authentication
dev test: e2e
### Component and test scenario title Masthead - Cloud ### Test scenario steps Masthead - Cloud - This component is marked behind a feature flag and it's available to preview via the experimental deploy preview. - Enable this component locally - Edit your `.env` file and add this line `DDS_CLOUD_MASTHEAD=true`. -------------- This is a follow-up issue to https://github.com/carbon-design-system/carbon-for-ibm-dotcom/issues/8133 once, the work outlined in this feature request has been addressed/merged https://github.com/carbon-design-system/carbon-for-ibm-dotcom/issues/8033. **Authentication** 1. Emulate the return of `_dl.ddo.user.segment.isCloudLoggedOn` in the window object - see (https://github.com/carbon-design-system/carbon-for-ibm-dotcom/pull/8149/files#diff-4d335d7e7137c49adbed103940aad93a66985e57608311a93400fe972323c5b9R12) 2. Once authenticated, ensure that the "Console" CTA button is visible and links out to `https://cloud.ibm.com`. 3. Ensure that the "Profile" button returns menu items. The amount doesn't matter as this content might be adjusted over time - https://github.ibm.com/webstandards/ibm-dotcom-translations/blob/master/translations/cloud-masthead/usen.json#L1527
1.0
[Test Scenario]: Cloud Masthead with Authentication - ### Component and test scenario title Masthead - Cloud ### Test scenario steps Masthead - Cloud - This component is marked behind a feature flag and it's available to preview via the experimental deploy preview. - Enable this component locally - Edit your `.env` file and add this line `DDS_CLOUD_MASTHEAD=true`. -------------- This is a follow-up issue to https://github.com/carbon-design-system/carbon-for-ibm-dotcom/issues/8133 once, the work outlined in this feature request has been addressed/merged https://github.com/carbon-design-system/carbon-for-ibm-dotcom/issues/8033. **Authentication** 1. Emulate the return of `_dl.ddo.user.segment.isCloudLoggedOn` in the window object - see (https://github.com/carbon-design-system/carbon-for-ibm-dotcom/pull/8149/files#diff-4d335d7e7137c49adbed103940aad93a66985e57608311a93400fe972323c5b9R12) 2. Once authenticated, ensure that the "Console" CTA button is visible and links out to `https://cloud.ibm.com`. 3. Ensure that the "Profile" button returns menu items. The amount doesn't matter as this content might be adjusted over time - https://github.ibm.com/webstandards/ibm-dotcom-translations/blob/master/translations/cloud-masthead/usen.json#L1527
non_process
cloud masthead with authentication component and test scenario title masthead cloud test scenario steps masthead cloud this component is marked behind a feature flag and it s available to preview via the experimental deploy preview enable this component locally edit your env file and add this line dds cloud masthead true this is a follow up issue to once the work outlined in this feature request has been addressed merged authentication emulate the return of dl ddo user segment iscloudloggedon in the window object see once authenticated ensure that the console cta button is visible and links out to ensure that the profile button returns menu items the amount doesn t matter as this content might be adjusted over time
0
9,515
12,498,194,284
IssuesEvent
2020-06-01 17:51:03
GoogleCloudPlatform/golang-samples
https://api.github.com/repos/GoogleCloudPlatform/golang-samples
closed
testing: add firestore testing to nightly E2E
Fixit api: firestore priority: p3 type: process
Firestore can't run in our regular test projects (incompatible with Datastore). Currently we don't run our Firestore tests because `GOLANG_SAMPLES_FIRESTORE_PROJECT` is unset. We should set this in our nightly E2E tests.
1.0
testing: add firestore testing to nightly E2E - Firestore can't run in our regular test projects (incompatible with Datastore). Currently we don't run our Firestore tests because `GOLANG_SAMPLES_FIRESTORE_PROJECT` is unset. We should set this in our nightly E2E tests.
process
testing add firestore testing to nightly firestore can t run in our regular test projects incompatible with datastore currently we don t run our firestore tests because golang samples firestore project is unset we should set this in our nightly tests
1
308,608
26,617,387,664
IssuesEvent
2023-01-24 08:32:08
Mbed-TLS/mbedtls
https://api.github.com/repos/Mbed-TLS/mbedtls
closed
Depends.py enhancement: key exchange tests
enhancement component-tls size-m component-test
After merging https://github.com/Mbed-TLS/mbedtls/pull/6385, there are configurations using various key exchange types, but there are no performed tests. This task is to change the current solution: ``` 'kex': ExclusiveDomain(key_exchange_symbols, [build_command + ['lib'], build_command + ['-C', 'programs']]), ``` to ``` 'kex': ExclusiveDomain(key_exchange_symbols, build_and_test), ``` And fix all issues that arise.
1.0
Depends.py enhancement: key exchange tests - After merging https://github.com/Mbed-TLS/mbedtls/pull/6385, there are configurations using various key exchange types, but there are no performed tests. This task is to change the current solution: ``` 'kex': ExclusiveDomain(key_exchange_symbols, [build_command + ['lib'], build_command + ['-C', 'programs']]), ``` to ``` 'kex': ExclusiveDomain(key_exchange_symbols, build_and_test), ``` And fix all issues that arise.
non_process
depends py enhancement key exchange tests after merging there are configurations using various key exchange types but there are no performed tests this task is to change the current solution kex exclusivedomain key exchange symbols build command to kex exclusivedomain key exchange symbols build and test and fix all issues that arise
0
11,250
14,016,847,973
IssuesEvent
2020-10-29 14:59:08
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
Import archive mailbox
enhancement mail processing verified
### Infos: * Used Zammad version: 3.4 * Installation method (source, package, ..): any * Operating system: any * Database + version: any * Elasticsearch version: any * Browser + version: any ### Expected behavior: * When configuring a mailbox it should be possible to decide wether this is a fresh inbox or an archive. * When the archive option is not set the import should behave as it does now. * When the archive option is set it should keep the timestamps of the imported mails and don't process mails (that are older than X) as if they were new (e.g. sending auto replies) ### Actual behavior: * When configuring a mailbox every email is treated as new which leads to sending auto replies to e.g. year/moth old tickets. ### Steps to reproduce the behavior: * Configure a mailbox with quite old emails in it. ### Additional information: There is currently a PoC/beta branch available that enables the import of an archive mailbox. However, it keeps treating the mailbox as an archive after everything got imported. However, the UI does not distinguish between a regular channel and the archive one which leads to confusion. This should be resolved. There are (at least) two approaches: 1st) Make the UI clear that it is an archive Channel 2nd) Switch to regular mode once initial sync is completed Yes I'm sure this is a bug and no feature request or a general question.
1.0
Import archive mailbox - ### Infos: * Used Zammad version: 3.4 * Installation method (source, package, ..): any * Operating system: any * Database + version: any * Elasticsearch version: any * Browser + version: any ### Expected behavior: * When configuring a mailbox it should be possible to decide wether this is a fresh inbox or an archive. * When the archive option is not set the import should behave as it does now. * When the archive option is set it should keep the timestamps of the imported mails and don't process mails (that are older than X) as if they were new (e.g. sending auto replies) ### Actual behavior: * When configuring a mailbox every email is treated as new which leads to sending auto replies to e.g. year/moth old tickets. ### Steps to reproduce the behavior: * Configure a mailbox with quite old emails in it. ### Additional information: There is currently a PoC/beta branch available that enables the import of an archive mailbox. However, it keeps treating the mailbox as an archive after everything got imported. However, the UI does not distinguish between a regular channel and the archive one which leads to confusion. This should be resolved. There are (at least) two approaches: 1st) Make the UI clear that it is an archive Channel 2nd) Switch to regular mode once initial sync is completed Yes I'm sure this is a bug and no feature request or a general question.
process
import archive mailbox infos used zammad version installation method source package any operating system any database version any elasticsearch version any browser version any expected behavior when configuring a mailbox it should be possible to decide wether this is a fresh inbox or an archive when the archive option is not set the import should behave as it does now when the archive option is set it should keep the timestamps of the imported mails and don t process mails that are older than x as if they were new e g sending auto replies actual behavior when configuring a mailbox every email is treated as new which leads to sending auto replies to e g year moth old tickets steps to reproduce the behavior configure a mailbox with quite old emails in it additional information there is currently a poc beta branch available that enables the import of an archive mailbox however it keeps treating the mailbox as an archive after everything got imported however the ui does not distinguish between a regular channel and the archive one which leads to confusion this should be resolved there are at least two approaches make the ui clear that it is an archive channel switch to regular mode once initial sync is completed yes i m sure this is a bug and no feature request or a general question
1