Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
14,914
18,344,260,769
IssuesEvent
2021-10-08 02:39:40
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Pre / Post Execution Script
Feedback stale Processing Bug
### What is the bug or the crash? Hello, I use QGIS 3.16.3. I would like to run a script after running a processing model. In the processing tool options I can specify a path for a script. However, this script does not work! In QGIS 2.18 it works. Furthermore, I would like to have this script or this path only in this project, how can this be solved? Thanks in advance ChrisGISv1 ### Steps to reproduce the issue ![issue](https://user-images.githubusercontent.com/88772838/129010409-4b7ea09c-de8b-4764-9160-4b0bba34591d.PNG) ### Versions <!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd"> <html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css"> p, li { white-space: pre-wrap; } </style></head><body> QGIS-Version | 3.16.3-Hannover | QGIS-Codeversion | 94ac9f21b8 -- | -- | -- | -- Kompiliert gegen Qt | 5.11.2 | Laufendes Qt | 5.11.2 Kompiliert mit GDAL/OGR | 3.1.4 | Läuft mit GDAL/OGR | 3.1.4 Kompiliert mit GEOS | 3.8.1-CAPI-1.13.3 | Läuft mit GEOS | 3.8.1-CAPI-1.13.3 Kompiliert mit SQLite | 3.29.0 | Läuft mit SQLite | 3.29.0 PostgreSQL-Client-Version | 11.5 | SpatiaLite-Version | 4.3.0 QWT-Version | 6.1.3 | QScintilla2-Version | 2.10.8 Kompiliert mit PROJ | 6.3.2 | Läuft mit PROJ | Rel. 6.3.2, May 1st, 2020 BS-Version | Windows 10 (10.0) Aktive Python-Erweiterungen | pluginbuilder3; plugin_reloader; db_manager; MetaSearch; processing </body></html><!--EndFragment--> ### Additional context _No response_
1.0
Pre / Post Execution Script - ### What is the bug or the crash? Hello, I use QGIS 3.16.3. I would like to run a script after running a processing model. In the processing tool options I can specify a path for a script. However, this script does not work! In QGIS 2.18 it works. Furthermore, I would like to have this script or this path only in this project, how can this be solved? Thanks in advance ChrisGISv1 ### Steps to reproduce the issue ![issue](https://user-images.githubusercontent.com/88772838/129010409-4b7ea09c-de8b-4764-9160-4b0bba34591d.PNG) ### Versions <!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd"> <html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css"> p, li { white-space: pre-wrap; } </style></head><body> QGIS-Version | 3.16.3-Hannover | QGIS-Codeversion | 94ac9f21b8 -- | -- | -- | -- Kompiliert gegen Qt | 5.11.2 | Laufendes Qt | 5.11.2 Kompiliert mit GDAL/OGR | 3.1.4 | Läuft mit GDAL/OGR | 3.1.4 Kompiliert mit GEOS | 3.8.1-CAPI-1.13.3 | Läuft mit GEOS | 3.8.1-CAPI-1.13.3 Kompiliert mit SQLite | 3.29.0 | Läuft mit SQLite | 3.29.0 PostgreSQL-Client-Version | 11.5 | SpatiaLite-Version | 4.3.0 QWT-Version | 6.1.3 | QScintilla2-Version | 2.10.8 Kompiliert mit PROJ | 6.3.2 | Läuft mit PROJ | Rel. 6.3.2, May 1st, 2020 BS-Version | Windows 10 (10.0) Aktive Python-Erweiterungen | pluginbuilder3; plugin_reloader; db_manager; MetaSearch; processing </body></html><!--EndFragment--> ### Additional context _No response_
process
pre post execution script what is the bug or the crash hello i use qgis i would like to run a script after running a processing model in the processing tool options i can specify a path for a script however this script does not work in qgis it works furthermore i would like to have this script or this path only in this project how can this be solved thanks in advance steps to reproduce the issue versions doctype html public dtd html en p li white space pre wrap qgis version hannover qgis codeversion kompiliert gegen qt laufendes qt kompiliert mit gdal ogr läuft mit gdal ogr kompiliert mit geos capi läuft mit geos capi kompiliert mit sqlite läuft mit sqlite postgresql client version spatialite version qwt version version kompiliert mit proj läuft mit proj rel may bs version windows aktive python erweiterungen plugin reloader db manager metasearch processing additional context no response
1
351,543
10,520,381,937
IssuesEvent
2019-09-30 00:52:26
ropenscilabs/drake-manual
https://api.github.com/repos/ropenscilabs/drake-manual
closed
Tell a more interesting story with the deep learning workflow
difficulty: advanced :three: high priority
The model tuning in the [current deep learning chapter](https://ropenscilabs.github.io/drake-manual/churn.html) does not noticeably affect model performance. If we found more impactful things to tune, we could tell more of a story. Related: https://github.com/leonjessen/RPharma2019/blob/master/Exercises/04_diamonds_regression.md
1.0
Tell a more interesting story with the deep learning workflow - The model tuning in the [current deep learning chapter](https://ropenscilabs.github.io/drake-manual/churn.html) does not noticeably affect model performance. If we found more impactful things to tune, we could tell more of a story. Related: https://github.com/leonjessen/RPharma2019/blob/master/Exercises/04_diamonds_regression.md
non_process
tell a more interesting story with the deep learning workflow the model tuning in the does not noticeably affect model performance if we found more impactful things to tune we could tell more of a story related
0
14,871
18,280,901,229
IssuesEvent
2021-10-05 03:06:57
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Python 3 support
P3 type: process team-Rules-Python
I would like for this issue to collect all issues concerning Python 3. As far as I know there are multiple problems: 1. The supported values for `srcs_version` are `PY2, PY3, PY2AND3, PY2ONLY, PY3ONLY`. - `PY3` is documented wrong. It actually handles similar to `PY2`, which is fine. - `PY3ONLY` is not even documented, although it mostly is what a lot of Python 3 users want. 2. By default Bazel always executes Python 2. Which might be an ok choice for now for `PY2`, `PY3`, `PY2AND3` and `PY2ONLY`. It is however totally wrong for `PY3ONLY`. (#1406) It seems like `checkSourceIsCompatible` is used instead of a hypothetical `selectPythonVersionForSource`. - Question there would be why it is not using Python 3. - How big is Googles interest in Python 3 support on a scale of 0 (want it now) to 5 (who cares)? - Would anyone inside of Google be available for mentorship for extending Python 3 support? 3. Is there a way for Bazel to use a Python version embedded into a repository? 4. `--force_python` and `--host_force_python` are given to users, yet they are not documented. 5. `py_binary` wrapper uses Python 2 although it may not even be present (#544) 6. `py_binary` wrapper spawns a new process for executing python scripts. I do know that we are not talking about great performance with Python but that seems a bit excessive. 7. `pkg_tar` seems to internally use Python2 only. [Does not work with 3](https://github.com/bazelbuild/bazel/issues/2352). 8. Python cannot be used as a label (#2244) 9. pyd cannot be created straight-forward (#2497) 10. Support embeddable Python for Windows (#2509)
1.0
Python 3 support - I would like for this issue to collect all issues concerning Python 3. As far as I know there are multiple problems: 1. The supported values for `srcs_version` are `PY2, PY3, PY2AND3, PY2ONLY, PY3ONLY`. - `PY3` is documented wrong. It actually handles similar to `PY2`, which is fine. - `PY3ONLY` is not even documented, although it mostly is what a lot of Python 3 users want. 2. By default Bazel always executes Python 2. Which might be an ok choice for now for `PY2`, `PY3`, `PY2AND3` and `PY2ONLY`. It is however totally wrong for `PY3ONLY`. (#1406) It seems like `checkSourceIsCompatible` is used instead of a hypothetical `selectPythonVersionForSource`. - Question there would be why it is not using Python 3. - How big is Googles interest in Python 3 support on a scale of 0 (want it now) to 5 (who cares)? - Would anyone inside of Google be available for mentorship for extending Python 3 support? 3. Is there a way for Bazel to use a Python version embedded into a repository? 4. `--force_python` and `--host_force_python` are given to users, yet they are not documented. 5. `py_binary` wrapper uses Python 2 although it may not even be present (#544) 6. `py_binary` wrapper spawns a new process for executing python scripts. I do know that we are not talking about great performance with Python but that seems a bit excessive. 7. `pkg_tar` seems to internally use Python2 only. [Does not work with 3](https://github.com/bazelbuild/bazel/issues/2352). 8. Python cannot be used as a label (#2244) 9. pyd cannot be created straight-forward (#2497) 10. Support embeddable Python for Windows (#2509)
process
python support i would like for this issue to collect all issues concerning python as far as i know there are multiple problems the supported values for srcs version are is documented wrong it actually handles similar to which is fine is not even documented although it mostly is what a lot of python users want by default bazel always executes python which might be an ok choice for now for and it is however totally wrong for it seems like checksourceiscompatible is used instead of a hypothetical selectpythonversionforsource question there would be why it is not using python how big is googles interest in python support on a scale of want it now to who cares would anyone inside of google be available for mentorship for extending python support is there a way for bazel to use a python version embedded into a repository force python and host force python are given to users yet they are not documented py binary wrapper uses python although it may not even be present py binary wrapper spawns a new process for executing python scripts i do know that we are not talking about great performance with python but that seems a bit excessive pkg tar seems to internally use only python cannot be used as a label pyd cannot be created straight forward support embeddable python for windows
1
794,062
28,021,108,114
IssuesEvent
2023-03-28 05:30:11
AY2223S2-CS2103T-T09-2/tp
https://api.github.com/repos/AY2223S2-CS2103T-T09-2/tp
closed
Edit command variant - Append
priority.High v1.3
Allows user to add on groups/tags instead of overwriting the existing groups/tags
1.0
Edit command variant - Append - Allows user to add on groups/tags instead of overwriting the existing groups/tags
non_process
edit command variant append allows user to add on groups tags instead of overwriting the existing groups tags
0
232,004
17,767,999,988
IssuesEvent
2021-08-30 10:01:49
amzn/selling-partner-api-docs
https://api.github.com/repos/amzn/selling-partner-api-docs
opened
Getting Error in CreateInboundShipmentPlan Api
bug documentation
I have used body part for creating shipment plan. $body = array( 'ShipFromAddress' => array( 'Name' => 'Gorilla Store', 'AddressLine1' => '9637 Pierrront ST', 'AddressLine2' => 'BURKE', 'DistrictOrCounty'=> 'US', 'City' => 'Richmond', 'StateOrProvinceCode' => 'VA', 'CountryCode' => 'US', 'PostalCode' => '20015' ), 'LabelPrepPreference' => 'SELLER_LABEL', 'ShipToCountryCode' => 'US', 'ShipToCountrySubdivisionCode' => '', 'InboundShipmentPlanRequestItems' => array( array( 'SellerSKU' => '', 'ASIN' => '', 'Condition' => 'NewItem', 'Quantity' => '5', 'QuantityInCase' => '', 'PrepDetailsList' => array( 'PrepInstruction' => 'BubbleWrapping ', 'PrepOwner' => 'SELLER' ) ) ) ); I am getting following array even I have tried to change country and address. [400] Client error: `POST https://sellingpartnerapi-na.amazon.com/fba/inbound/v0/plans` resulted in a `400 Bad Request` response: { "errors": [ { "code": "InvalidInput", "message": " Reason: Country field is missing, Name field is m (truncated...)
1.0
Getting Error in CreateInboundShipmentPlan Api - I have used body part for creating shipment plan. $body = array( 'ShipFromAddress' => array( 'Name' => 'Gorilla Store', 'AddressLine1' => '9637 Pierrront ST', 'AddressLine2' => 'BURKE', 'DistrictOrCounty'=> 'US', 'City' => 'Richmond', 'StateOrProvinceCode' => 'VA', 'CountryCode' => 'US', 'PostalCode' => '20015' ), 'LabelPrepPreference' => 'SELLER_LABEL', 'ShipToCountryCode' => 'US', 'ShipToCountrySubdivisionCode' => '', 'InboundShipmentPlanRequestItems' => array( array( 'SellerSKU' => '', 'ASIN' => '', 'Condition' => 'NewItem', 'Quantity' => '5', 'QuantityInCase' => '', 'PrepDetailsList' => array( 'PrepInstruction' => 'BubbleWrapping ', 'PrepOwner' => 'SELLER' ) ) ) ); I am getting following array even I have tried to change country and address. [400] Client error: `POST https://sellingpartnerapi-na.amazon.com/fba/inbound/v0/plans` resulted in a `400 Bad Request` response: { "errors": [ { "code": "InvalidInput", "message": " Reason: Country field is missing, Name field is m (truncated...)
non_process
getting error in createinboundshipmentplan api i have used body part for creating shipment plan body array shipfromaddress array name gorilla store pierrront st burke districtorcounty us city richmond stateorprovincecode va countrycode us postalcode labelpreppreference seller label shiptocountrycode us shiptocountrysubdivisioncode inboundshipmentplanrequestitems array array sellersku asin condition newitem quantity quantityincase prepdetailslist array prepinstruction bubblewrapping prepowner seller i am getting following array even i have tried to change country and address client error post resulted in a bad request response errors code invalidinput message reason country field is missing name field is m truncated
0
19,061
25,078,595,652
IssuesEvent
2022-11-07 17:18:10
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
New reset tab: issue when compress history
bug: wip reproduce: confirmed scope: UI scope: image processing
**To Reproduce** 1. Have a tab to reset (when I found the issue, I was on filmic with both look and options to reset) 2. Double-click on tab: tab is reset 3. Compress history (with c shortcut or button) 4. See that reset tab go back on previous state before reset **Expected behavior** Compress history should consider tab reset. @dterrahe: this one for you
1.0
New reset tab: issue when compress history - **To Reproduce** 1. Have a tab to reset (when I found the issue, I was on filmic with both look and options to reset) 2. Double-click on tab: tab is reset 3. Compress history (with c shortcut or button) 4. See that reset tab go back on previous state before reset **Expected behavior** Compress history should consider tab reset. @dterrahe: this one for you
process
new reset tab issue when compress history to reproduce have a tab to reset when i found the issue i was on filmic with both look and options to reset double click on tab tab is reset compress history with c shortcut or button see that reset tab go back on previous state before reset expected behavior compress history should consider tab reset dterrahe this one for you
1
2,569
5,325,644,003
IssuesEvent
2017-02-15 00:28:17
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
opened
[subtitles] [German] L'Union européenne est entrée dans un processus de dislocation
Process: [0] Awaiting subtitles
# Video title L'Union européenne est entrée dans un processus de dislocation # URL https://www.youtube.com/watch?v=QzgNqXhn_Ao&t=68s # Youtube subtitles language German # Duration 3:05 # Subtitles URL https://www.youtube.com/timedtext_editor?bl=vmp&ref=player&tab=captions&v=QzgNqXhn_Ao&lang=de&ui=hd&action_mde_edit_form=1
1.0
[subtitles] [German] L'Union européenne est entrée dans un processus de dislocation - # Video title L'Union européenne est entrée dans un processus de dislocation # URL https://www.youtube.com/watch?v=QzgNqXhn_Ao&t=68s # Youtube subtitles language German # Duration 3:05 # Subtitles URL https://www.youtube.com/timedtext_editor?bl=vmp&ref=player&tab=captions&v=QzgNqXhn_Ao&lang=de&ui=hd&action_mde_edit_form=1
process
l union européenne est entrée dans un processus de dislocation video title l union européenne est entrée dans un processus de dislocation url youtube subtitles language german duration subtitles url
1
11,939
14,707,504,018
IssuesEvent
2021-01-04 21:44:10
googleapis/python-speech
https://api.github.com/repos/googleapis/python-speech
closed
speech/cloud-client test failures
api: speech priority: p2 type: process
Builds https://source.cloud.google.com/results/invocations/ee35c95b-7ef9-4c90-b5ca-b30534a5c83b/targets https://source.cloud.google.com/results/invocations/eace6b03-019a-4f2a-9802-4e07135a922d/targets We don't have xunit xml log file, so build cop bot didn't catch it, but it seems like Python died. Log: ```============================= test session starts ============================== platform linux -- Python 3.7.7, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 -- /workspace/speech/cloud-client/.nox/py-3-7/bin/python cachedir: .pytest_cache rootdir: /workspace, configfile: pytest.ini collecting ... collected 22 items beta_snippets_test.py::test_transcribe_file_with_enhanced_model PASSED [ 4%] beta_snippets_test.py::test_transcribe_file_with_metadata PASSED [ 9%] beta_snippets_test.py::test_transcribe_file_with_auto_punctuation PASSED [ 13%] beta_snippets_test.py::test_transcribe_diarization PASSED [ 18%] beta_snippets_test.py::test_transcribe_multichannel_file E0828 10:27:18.708634490 13111 sync_posix.cc:67] assertion failed: pthread_mutex_lock(mu) == 0 Fatal Python error: Aborted Thread 0x00007f2c36e0e600 (most recent call first): File "/usr/local/lib/python3.7/codecs.py", line 322 in decode File "/workspace/speech/cloud-client/.nox/py-3-7/lib/python3.7/site-packages/_pytest/capture.py", line 484 in snap File "/workspace/speech/cloud-client/.nox/py-3-7/lib/python3.7/site-packages/_pytest/capture.py", line 570 in readouterr File "/workspace/speech/cloud-client/.nox/py-3-7/lib/python3.7/site-packages/_pytest/capture.py", line 657 in read_global_capture File "/workspace/speech/cloud-client/.nox/py-3-7/lib/python3.7/site-packages/_pytest/capture.py", line 718 in item_capture nox > Command pytest --junitxm```
1.0
speech/cloud-client test failures - Builds https://source.cloud.google.com/results/invocations/ee35c95b-7ef9-4c90-b5ca-b30534a5c83b/targets https://source.cloud.google.com/results/invocations/eace6b03-019a-4f2a-9802-4e07135a922d/targets We don't have xunit xml log file, so build cop bot didn't catch it, but it seems like Python died. Log: ```============================= test session starts ============================== platform linux -- Python 3.7.7, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 -- /workspace/speech/cloud-client/.nox/py-3-7/bin/python cachedir: .pytest_cache rootdir: /workspace, configfile: pytest.ini collecting ... collected 22 items beta_snippets_test.py::test_transcribe_file_with_enhanced_model PASSED [ 4%] beta_snippets_test.py::test_transcribe_file_with_metadata PASSED [ 9%] beta_snippets_test.py::test_transcribe_file_with_auto_punctuation PASSED [ 13%] beta_snippets_test.py::test_transcribe_diarization PASSED [ 18%] beta_snippets_test.py::test_transcribe_multichannel_file E0828 10:27:18.708634490 13111 sync_posix.cc:67] assertion failed: pthread_mutex_lock(mu) == 0 Fatal Python error: Aborted Thread 0x00007f2c36e0e600 (most recent call first): File "/usr/local/lib/python3.7/codecs.py", line 322 in decode File "/workspace/speech/cloud-client/.nox/py-3-7/lib/python3.7/site-packages/_pytest/capture.py", line 484 in snap File "/workspace/speech/cloud-client/.nox/py-3-7/lib/python3.7/site-packages/_pytest/capture.py", line 570 in readouterr File "/workspace/speech/cloud-client/.nox/py-3-7/lib/python3.7/site-packages/_pytest/capture.py", line 657 in read_global_capture File "/workspace/speech/cloud-client/.nox/py-3-7/lib/python3.7/site-packages/_pytest/capture.py", line 718 in item_capture nox > Command pytest --junitxm```
process
speech cloud client test failures builds we don t have xunit xml log file so build cop bot didn t catch it but it seems like python died log test session starts platform linux python pytest py pluggy workspace speech cloud client nox py bin python cachedir pytest cache rootdir workspace configfile pytest ini collecting collected items beta snippets test py test transcribe file with enhanced model passed beta snippets test py test transcribe file with metadata passed beta snippets test py test transcribe file with auto punctuation passed beta snippets test py test transcribe diarization passed beta snippets test py test transcribe multichannel file sync posix cc assertion failed pthread mutex lock mu fatal python error aborted thread most recent call first file usr local lib codecs py line in decode file workspace speech cloud client nox py lib site packages pytest capture py line in snap file workspace speech cloud client nox py lib site packages pytest capture py line in readouterr file workspace speech cloud client nox py lib site packages pytest capture py line in read global capture file workspace speech cloud client nox py lib site packages pytest capture py line in item capture nox command pytest junitxm
1
41,359
12,832,000,518
IssuesEvent
2020-07-07 06:48:11
rvvergara/next-js-basic
https://api.github.com/repos/rvvergara/next-js-basic
closed
CVE-2020-7598 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz
security vulnerability
## CVE-2020-7598 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary> <p> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/next-js-basic/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/next-js-basic/node_modules/mkdirp/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - eslint-5.16.0.tgz (Root Library) - mkdirp-0.5.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/next-js-basic/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/next-js-basic/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - next-8.1.0.tgz (Root Library) - webpack-4.29.0.tgz - watchpack-1.6.0.tgz - chokidar-2.1.6.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - rc-1.2.8.tgz - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/rvvergara/next-js-basic/commit/84695a914e1eb9c6b29e2f0eb39bfe4960ad47c8">84695a914e1eb9c6b29e2f0eb39bfe4960ad47c8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p> <p>Release Date: 2020-03-11</p> <p>Fix Resolution: minimist - 0.2.1,1.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7598 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - ## CVE-2020-7598 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary> <p> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/next-js-basic/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/next-js-basic/node_modules/mkdirp/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - eslint-5.16.0.tgz (Root Library) - mkdirp-0.5.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/next-js-basic/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/next-js-basic/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - next-8.1.0.tgz (Root Library) - webpack-4.29.0.tgz - watchpack-1.6.0.tgz - chokidar-2.1.6.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - rc-1.2.8.tgz - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/rvvergara/next-js-basic/commit/84695a914e1eb9c6b29e2f0eb39bfe4960ad47c8">84695a914e1eb9c6b29e2f0eb39bfe4960ad47c8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p> <p>Release Date: 2020-03-11</p> <p>Fix Resolution: minimist - 0.2.1,1.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in minimist tgz minimist tgz cve medium severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz parse argument options library home page a href path to dependency file tmp ws scm next js basic package json path to vulnerable library tmp ws scm next js basic node modules mkdirp node modules minimist package json dependency hierarchy eslint tgz root library mkdirp tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file tmp ws scm next js basic package json path to vulnerable library tmp ws scm next js basic node modules minimist package json dependency hierarchy next tgz root library webpack tgz watchpack tgz chokidar tgz fsevents tgz node pre gyp tgz rc tgz x minimist tgz vulnerable library found in head commit a href vulnerability details minimist before could be tricked into adding or modifying properties of object prototype using a constructor or proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution minimist step up your open source security game with whitesource
0
4,652
7,495,777,937
IssuesEvent
2018-04-08 01:13:35
gkiar/reading
https://api.github.com/repos/gkiar/reading
closed
Paper: testname
processing
URL: [http://testurl.io](http://testurl.io) ## This paper does... ## This paper does not... ## Other comments?
1.0
Paper: testname - URL: [http://testurl.io](http://testurl.io) ## This paper does... ## This paper does not... ## Other comments?
process
paper testname url this paper does this paper does not other comments
1
2,307
5,120,135,203
IssuesEvent
2017-01-09 00:41:00
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
opened
[Subtitles] [FR] MÉLENCHON - ESPRIT DE CAMPAGNE #0 - #EDC0
Language: French Process: Someone is working on this issue Process: [1] Writing in progress
# Video title MÉLENCHON - ESPRIT DE CAMPAGNE #0 - #EDC0 # URL https://www.youtube.com/watch?v=3iOUqUp1r3w&list=PLnAm9o_Xn_3CFeesH6PnFXAi9CzDkayoj # Youtube subtitles language Français # Duration 1:12:48 # Subtitles URL https://www.youtube.com/timedtext_editor?ref=player&action_mde_edit_form=1&v=3iOUqUp1r3w&tab=captions&lang=fr&bl=vmp&ui=hd
2.0
[Subtitles] [FR] MÉLENCHON - ESPRIT DE CAMPAGNE #0 - #EDC0 - # Video title MÉLENCHON - ESPRIT DE CAMPAGNE #0 - #EDC0 # URL https://www.youtube.com/watch?v=3iOUqUp1r3w&list=PLnAm9o_Xn_3CFeesH6PnFXAi9CzDkayoj # Youtube subtitles language Français # Duration 1:12:48 # Subtitles URL https://www.youtube.com/timedtext_editor?ref=player&action_mde_edit_form=1&v=3iOUqUp1r3w&tab=captions&lang=fr&bl=vmp&ui=hd
process
mélenchon esprit de campagne video title mélenchon esprit de campagne url youtube subtitles language français duration subtitles url
1
9,086
12,154,473,164
IssuesEvent
2020-04-25 08:29:40
utopia-rise/godot-kotlin
https://api.github.com/repos/utopia-rise/godot-kotlin
closed
Implement property hint string annotations
feature tools:annotation-processor tools:annotations tools:entry-generator
**Describe the problem or limitation you are having in your project:** With this a user won't have to write error prone property hint strings like this anymore: `0,100000,1000,or_greater` **Describe how this feature / enhancement will help you overcome this problem or limitation:** The usability of the registration system would be much improved when using property hint strings **Show a mock up screenshots/video or a flow diagram explaining how your proposal will work:** We could define annotations like `@HintRange(start = 0, end = 1000, step = 2)` which would provide typesafety **Describe implementation detail for your proposal (in code), if possible:** None to provide atm. **If this enhancement will not be used often, can it be worked around with a few lines of code?:** Yes. We could still force the user to use type strings. But it's errorprone. **Is there a reason why this should be in this project and not individually solved?:** Annotations and the entry generating process are both part of the core project.
1.0
Implement property hint string annotations - **Describe the problem or limitation you are having in your project:** With this a user won't have to write error prone property hint strings like this anymore: `0,100000,1000,or_greater` **Describe how this feature / enhancement will help you overcome this problem or limitation:** The usability of the registration system would be much improved when using property hint strings **Show a mock up screenshots/video or a flow diagram explaining how your proposal will work:** We could define annotations like `@HintRange(start = 0, end = 1000, step = 2)` which would provide typesafety **Describe implementation detail for your proposal (in code), if possible:** None to provide atm. **If this enhancement will not be used often, can it be worked around with a few lines of code?:** Yes. We could still force the user to use type strings. But it's errorprone. **Is there a reason why this should be in this project and not individually solved?:** Annotations and the entry generating process are both part of the core project.
process
implement property hint string annotations describe the problem or limitation you are having in your project with this a user won t have to write error prone property hint strings like this anymore or greater describe how this feature enhancement will help you overcome this problem or limitation the usability of the registration system would be much improved when using property hint strings show a mock up screenshots video or a flow diagram explaining how your proposal will work we could define annotations like hintrange start end step which would provide typesafety describe implementation detail for your proposal in code if possible none to provide atm if this enhancement will not be used often can it be worked around with a few lines of code yes we could still force the user to use type strings but it s errorprone is there a reason why this should be in this project and not individually solved annotations and the entry generating process are both part of the core project
1
226,250
17,331,357,608
IssuesEvent
2021-07-28 03:06:54
amzn/selling-partner-api-docs
https://api.github.com/repos/amzn/selling-partner-api-docs
opened
[BUG] Create destination request always response forbidden
bug documentation
### I created Destination to SQS but always response forbidden: **com.amazon.spapi.client.ApiException: Forbidden** I step by step use [https://github.com/amzn/selling-partner-api-docs/blob/main/guides/en-US/use-case-guides/notifications-api-use-case-guide/notifications-use-case-guide-v1.md](url) config my sqs queue 1. select region is policy is : us-east-1 owner: is my iam account name : jona other: set Effect to Allow. Set Principal to 437568002678. Set Actions to SendMessage and GetQueueAttributes. ` { "Version": "2008-10-17", "Id": "__default_policy_ID", "Statement": [ { "Sid": "__owner_statement", "Effect": "Allow", "Principal": { "AWS": "arn:aws:iam::xxxxxxxx:user/jona" }, "Action": "SQS:*", "Resource": "arn:aws:sqs:us-east-1:xxxxxxxx:SQS_Study" }, { "Sid": "437568002678", "Effect": "Allow", "Principal": { "AWS": "arn:aws:iam::437568002678:root" }, "Action": [ "sqs:GetQueueAttributes", "sqs:SendMessage" ], "Resource": "arn:aws:sqs:us-east-1:xxxxxxx:SQS_Study" }, { "Sid": "__receiver_statement", "Effect": "Allow", "Principal": { "AWS": "arn:aws:iam::xxxxxx:role/SellingPartnerRole" }, "Action": [ "SQS:ChangeMessageVisibility", "SQS:DeleteMessage", "SQS:ReceiveMessage" ], "Resource": "arn:aws:sqs:us-east-1:xxxxxx:SQS_Study" } ] } ` and this my java code: ` CreateDestinationRequest body = new CreateDestinationRequest(); body.setName("sqs test"); DestinationResourceSpecification destinationRsn = new DestinationResourceSpecification(); SqsResource sqsResource = new SqsResource(); sqsResource.setArn("arn:aws:sqs:us-east-1:xxxxxxxx:SQS_Study"); destinationRsn.setSqs(sqsResource); body.setResourceSpecification(destinationRsn); AWSAuthenticationCredentials awsAuthenticationCredentials = AWSAuthenticationCredentials.builder() //IAM user的accessKeyId .accessKeyId("xxxxxx") //IAM user的secretKey .secretKey("xxxxxx) .region("us-east-1") .build(); AWSAuthenticationCredentialsProvider awsAuthenticationCredentialsProvider = AWSAuthenticationCredentialsProvider.builder() .roleArn("arn:aws:iam::xxxxxxxx:role/SellingPartnerRole") .roleSessionName("sc-sp-api-2") .build(); LWAAuthorizationCredentials lwaAuthorizationCredentials = LWAAuthorizationCredentials.builder() .clientId("amzn1.application-oa2-client.xxxxxxxx") .clientSecret("xxxxxxxxx") .refreshToken("xxxxxxxxxx") .endpoint("https://api.amazon.com/auth/o2/token") .build(); NotificationsApi api = new NotificationsApi.Builder() .awsAuthenticationCredentials(awsAuthenticationCredentials) .lwaAuthorizationCredentials(lwaAuthorizationCredentials) .awsAuthenticationCredentialsProvider(awsAuthenticationCredentialsProvider) .endpoint("https://sellingpartnerapi-na.amazon.com") .build(); ` ![image](https://user-images.githubusercontent.com/22114034/127257190-d948c290-effb-44de-9a1e-3d2b26c2ea29.png) So i want to know what is wrong with my code or setting
1.0
[BUG] Create destination request always response forbidden - ### I created Destination to SQS but always response forbidden: **com.amazon.spapi.client.ApiException: Forbidden** I step by step use [https://github.com/amzn/selling-partner-api-docs/blob/main/guides/en-US/use-case-guides/notifications-api-use-case-guide/notifications-use-case-guide-v1.md](url) config my sqs queue 1. select region is policy is : us-east-1 owner: is my iam account name : jona other: set Effect to Allow. Set Principal to 437568002678. Set Actions to SendMessage and GetQueueAttributes. ` { "Version": "2008-10-17", "Id": "__default_policy_ID", "Statement": [ { "Sid": "__owner_statement", "Effect": "Allow", "Principal": { "AWS": "arn:aws:iam::xxxxxxxx:user/jona" }, "Action": "SQS:*", "Resource": "arn:aws:sqs:us-east-1:xxxxxxxx:SQS_Study" }, { "Sid": "437568002678", "Effect": "Allow", "Principal": { "AWS": "arn:aws:iam::437568002678:root" }, "Action": [ "sqs:GetQueueAttributes", "sqs:SendMessage" ], "Resource": "arn:aws:sqs:us-east-1:xxxxxxx:SQS_Study" }, { "Sid": "__receiver_statement", "Effect": "Allow", "Principal": { "AWS": "arn:aws:iam::xxxxxx:role/SellingPartnerRole" }, "Action": [ "SQS:ChangeMessageVisibility", "SQS:DeleteMessage", "SQS:ReceiveMessage" ], "Resource": "arn:aws:sqs:us-east-1:xxxxxx:SQS_Study" } ] } ` and this my java code: ` CreateDestinationRequest body = new CreateDestinationRequest(); body.setName("sqs test"); DestinationResourceSpecification destinationRsn = new DestinationResourceSpecification(); SqsResource sqsResource = new SqsResource(); sqsResource.setArn("arn:aws:sqs:us-east-1:xxxxxxxx:SQS_Study"); destinationRsn.setSqs(sqsResource); body.setResourceSpecification(destinationRsn); AWSAuthenticationCredentials awsAuthenticationCredentials = AWSAuthenticationCredentials.builder() //IAM user的accessKeyId .accessKeyId("xxxxxx") //IAM user的secretKey .secretKey("xxxxxx) .region("us-east-1") .build(); AWSAuthenticationCredentialsProvider awsAuthenticationCredentialsProvider = AWSAuthenticationCredentialsProvider.builder() .roleArn("arn:aws:iam::xxxxxxxx:role/SellingPartnerRole") .roleSessionName("sc-sp-api-2") .build(); LWAAuthorizationCredentials lwaAuthorizationCredentials = LWAAuthorizationCredentials.builder() .clientId("amzn1.application-oa2-client.xxxxxxxx") .clientSecret("xxxxxxxxx") .refreshToken("xxxxxxxxxx") .endpoint("https://api.amazon.com/auth/o2/token") .build(); NotificationsApi api = new NotificationsApi.Builder() .awsAuthenticationCredentials(awsAuthenticationCredentials) .lwaAuthorizationCredentials(lwaAuthorizationCredentials) .awsAuthenticationCredentialsProvider(awsAuthenticationCredentialsProvider) .endpoint("https://sellingpartnerapi-na.amazon.com") .build(); ` ![image](https://user-images.githubusercontent.com/22114034/127257190-d948c290-effb-44de-9a1e-3d2b26c2ea29.png) So i want to know what is wrong with my code or setting
non_process
create destination request always response forbidden i created destination to sqs but always response forbidden com amazon spapi client apiexception forbidden i step by step use url config my sqs queue select region is policy is us east owner is my iam account name jona other set effect to allow set principal to set actions to sendmessage and getqueueattributes version id default policy id statement sid owner statement effect allow principal aws arn aws iam xxxxxxxx user jona action sqs resource arn aws sqs us east xxxxxxxx sqs study sid effect allow principal aws arn aws iam root action sqs getqueueattributes sqs sendmessage resource arn aws sqs us east xxxxxxx sqs study sid receiver statement effect allow principal aws arn aws iam xxxxxx role sellingpartnerrole action sqs changemessagevisibility sqs deletemessage sqs receivemessage resource arn aws sqs us east xxxxxx sqs study and this my java code createdestinationrequest body new createdestinationrequest body setname sqs test destinationresourcespecification destinationrsn new destinationresourcespecification sqsresource sqsresource new sqsresource sqsresource setarn arn aws sqs us east xxxxxxxx sqs study destinationrsn setsqs sqsresource body setresourcespecification destinationrsn awsauthenticationcredentials awsauthenticationcredentials awsauthenticationcredentials builder iam user的accesskeyid accesskeyid xxxxxx iam user的secretkey secretkey xxxxxx region us east build awsauthenticationcredentialsprovider awsauthenticationcredentialsprovider awsauthenticationcredentialsprovider builder rolearn arn aws iam xxxxxxxx role sellingpartnerrole rolesessionname sc sp api build lwaauthorizationcredentials lwaauthorizationcredentials lwaauthorizationcredentials builder clientid application client xxxxxxxx clientsecret xxxxxxxxx refreshtoken xxxxxxxxxx endpoint build notificationsapi api new notificationsapi builder awsauthenticationcredentials awsauthenticationcredentials lwaauthorizationcredentials lwaauthorizationcredentials awsauthenticationcredentialsprovider awsauthenticationcredentialsprovider endpoint build so i want to know what is wrong with my code or setting
0
720,514
24,795,643,463
IssuesEvent
2022-10-24 17:00:26
AY2223S1-CS2113-W12-3/tp
https://api.github.com/repos/AY2223S1-CS2113-W12-3/tp
closed
As a user, I want to associate my food intake along with the date
priority.Medium type.Enhancement
so that I can recall what food I consumed on a certain day
1.0
As a user, I want to associate my food intake along with the date - so that I can recall what food I consumed on a certain day
non_process
as a user i want to associate my food intake along with the date so that i can recall what food i consumed on a certain day
0
11,998
14,737,341,944
IssuesEvent
2021-01-07 01:33:50
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
not able to upload billing file for Keener Communications billing group
anc-external anc-ops anc-process anp-1 ant-support
In GitLab by @kdjstudios on May 8, 2018, 11:10 **Submitted by:** Gaylan Garrett <gaylan@keenercom.net> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-05-07-39462 **Server:** External **Client/Site:** Keener **Account:** NA **Issue:** Today is billing day for the Keener Communications billing cycle. I went to upload my file and I do not have that option. I have the option to download file but I did not upload the file yet as I just generated it today. Can you please check into this matter and let me know what to do.
1.0
not able to upload billing file for Keener Communications billing group - In GitLab by @kdjstudios on May 8, 2018, 11:10 **Submitted by:** Gaylan Garrett <gaylan@keenercom.net> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-05-07-39462 **Server:** External **Client/Site:** Keener **Account:** NA **Issue:** Today is billing day for the Keener Communications billing cycle. I went to upload my file and I do not have that option. I have the option to download file but I did not upload the file yet as I just generated it today. Can you please check into this matter and let me know what to do.
process
not able to upload billing file for keener communications billing group in gitlab by kdjstudios on may submitted by gaylan garrett helpdesk server external client site keener account na issue today is billing day for the keener communications billing cycle i went to upload my file and i do not have that option i have the option to download file but i did not upload the file yet as i just generated it today can you please check into this matter and let me know what to do
1
514,255
14,936,436,826
IssuesEvent
2021-01-25 13:23:32
leppott/ContDataQC
https://api.github.com/repos/leppott/ContDataQC
closed
Aggregate Lake Data (different depths)
high priority
**Is your feature request related to a problem? Please describe.** Aggregate data from same location but different depths. **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
Aggregate Lake Data (different depths) - **Is your feature request related to a problem? Please describe.** Aggregate data from same location but different depths. **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
non_process
aggregate lake data different depths is your feature request related to a problem please describe aggregate data from same location but different depths describe the solution you d like a clear and concise description of what you want to happen describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
0
236,759
19,572,663,601
IssuesEvent
2022-01-04 11:53:24
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/maps/embeddable/dashboard·js - maps app embeddable embed in dashboard should apply new container state (time, query, filters) to embeddable
Team:Geo failed-test
A test failed on a tracked branch ``` StaleElementReferenceError: stale element reference: element is not attached to the page document (Session info: headless chrome=94.0.4606.71) at Object.throwDecodedError (/var/lib/buildkite-agent/builds/kb-cigroup-6-60a649a9d0bf3a57/elastic/kibana-hourly/kibana/node_modules/selenium-webdriver/lib/error.js:550:15) at parseHttpResponse (/var/lib/buildkite-agent/builds/kb-cigroup-6-60a649a9d0bf3a57/elastic/kibana-hourly/kibana/node_modules/selenium-webdriver/lib/http.js:565:13) at Executor.execute (/var/lib/buildkite-agent/builds/kb-cigroup-6-60a649a9d0bf3a57/elastic/kibana-hourly/kibana/node_modules/selenium-webdriver/lib/http.js:491:26) at runMicrotasks (<anonymous>) at processTicksAndRejections (internal/process/task_queues.js:95:5) at Task.exec (/var/lib/buildkite-agent/builds/kb-cigroup-6-60a649a9d0bf3a57/elastic/kibana-hourly/kibana/test/functional/services/remote/prevent_parallel_calls.ts:28:20) { remoteStacktrace: '#0 0x55b188eb08f3 <unknown>\n' + '#1 0x55b188995ba8 <unknown>\n' + '#2 0x55b188998a2c <unknown>\n' + '#3 0x55b188998826 <unknown>\n' + '#4 0x55b188998aec <unknown>\n' + '#5 0x55b1889cb6b6 <unknown>\n' + '#6 0x55b1889c0694 <unknown>\n' + '#7 0x55b1889e8e9d <unknown>\n' + '#8 0x55b1889c0593 <unknown>\n' + '#9 0x55b1889e8f7e <unknown>\n' + '#10 0x55b1889fbdac <unknown>\n' + '#11 0x55b1889e8d63 <unknown>\n' + '#12 0x55b1889bf144 <unknown>\n' + '#13 0x55b1889c0135 <unknown>\n' + '#14 0x55b188edfc3e <unknown>\n' + '#15 0x55b188ef56b7 <unknown>\n' + '#16 0x55b188ee0b95 <unknown>\n' + '#17 0x55b188ef6b05 <unknown>\n' + '#18 0x55b188ed52ab <unknown>\n' + '#19 0x55b188f11248 <unknown>\n' + '#20 0x55b188f113c8 <unknown>\n' + '#21 0x55b188f2c33d <unknown>\n' + '#22 0x7f07d056d609 start_thread\n' } ``` First failure: [CI Build - master](https://buildkite.com/elastic/kibana-hourly/builds/945#956df55a-ef31-4f34-a520-dde9e89d013d) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/maps/embeddable/dashboard·js","test.name":"maps app embeddable embed in dashboard should apply new container state (time, query, filters) to embeddable","test.failCount":7}} -->
1.0
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/maps/embeddable/dashboard·js - maps app embeddable embed in dashboard should apply new container state (time, query, filters) to embeddable - A test failed on a tracked branch ``` StaleElementReferenceError: stale element reference: element is not attached to the page document (Session info: headless chrome=94.0.4606.71) at Object.throwDecodedError (/var/lib/buildkite-agent/builds/kb-cigroup-6-60a649a9d0bf3a57/elastic/kibana-hourly/kibana/node_modules/selenium-webdriver/lib/error.js:550:15) at parseHttpResponse (/var/lib/buildkite-agent/builds/kb-cigroup-6-60a649a9d0bf3a57/elastic/kibana-hourly/kibana/node_modules/selenium-webdriver/lib/http.js:565:13) at Executor.execute (/var/lib/buildkite-agent/builds/kb-cigroup-6-60a649a9d0bf3a57/elastic/kibana-hourly/kibana/node_modules/selenium-webdriver/lib/http.js:491:26) at runMicrotasks (<anonymous>) at processTicksAndRejections (internal/process/task_queues.js:95:5) at Task.exec (/var/lib/buildkite-agent/builds/kb-cigroup-6-60a649a9d0bf3a57/elastic/kibana-hourly/kibana/test/functional/services/remote/prevent_parallel_calls.ts:28:20) { remoteStacktrace: '#0 0x55b188eb08f3 <unknown>\n' + '#1 0x55b188995ba8 <unknown>\n' + '#2 0x55b188998a2c <unknown>\n' + '#3 0x55b188998826 <unknown>\n' + '#4 0x55b188998aec <unknown>\n' + '#5 0x55b1889cb6b6 <unknown>\n' + '#6 0x55b1889c0694 <unknown>\n' + '#7 0x55b1889e8e9d <unknown>\n' + '#8 0x55b1889c0593 <unknown>\n' + '#9 0x55b1889e8f7e <unknown>\n' + '#10 0x55b1889fbdac <unknown>\n' + '#11 0x55b1889e8d63 <unknown>\n' + '#12 0x55b1889bf144 <unknown>\n' + '#13 0x55b1889c0135 <unknown>\n' + '#14 0x55b188edfc3e <unknown>\n' + '#15 0x55b188ef56b7 <unknown>\n' + '#16 0x55b188ee0b95 <unknown>\n' + '#17 0x55b188ef6b05 <unknown>\n' + '#18 0x55b188ed52ab <unknown>\n' + '#19 0x55b188f11248 <unknown>\n' + '#20 0x55b188f113c8 <unknown>\n' + '#21 0x55b188f2c33d <unknown>\n' + '#22 0x7f07d056d609 start_thread\n' } ``` First failure: [CI Build - master](https://buildkite.com/elastic/kibana-hourly/builds/945#956df55a-ef31-4f34-a520-dde9e89d013d) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/maps/embeddable/dashboard·js","test.name":"maps app embeddable embed in dashboard should apply new container state (time, query, filters) to embeddable","test.failCount":7}} -->
non_process
failing test chrome x pack ui functional tests x pack test functional apps maps embeddable dashboard·js maps app embeddable embed in dashboard should apply new container state time query filters to embeddable a test failed on a tracked branch staleelementreferenceerror stale element reference element is not attached to the page document session info headless chrome at object throwdecodederror var lib buildkite agent builds kb cigroup elastic kibana hourly kibana node modules selenium webdriver lib error js at parsehttpresponse var lib buildkite agent builds kb cigroup elastic kibana hourly kibana node modules selenium webdriver lib http js at executor execute var lib buildkite agent builds kb cigroup elastic kibana hourly kibana node modules selenium webdriver lib http js at runmicrotasks at processticksandrejections internal process task queues js at task exec var lib buildkite agent builds kb cigroup elastic kibana hourly kibana test functional services remote prevent parallel calls ts remotestacktrace n n n n n n n n n n n n n n n n n n n n n n start thread n first failure
0
686,488
23,493,131,917
IssuesEvent
2022-08-17 20:56:40
googleapis/python-bigtable
https://api.github.com/repos/googleapis/python-bigtable
closed
Upgrade to protobuf version 4?
api: bigtable priority: p2
I'm using another library that pins protobuf at the latest version (4.21.4), and I can't use this library with it because it pins it at <4. https://github.com/googleapis/python-bigtable/blob/main/setup.py#L36 Are there plans to upgrade soon?
1.0
Upgrade to protobuf version 4? - I'm using another library that pins protobuf at the latest version (4.21.4), and I can't use this library with it because it pins it at <4. https://github.com/googleapis/python-bigtable/blob/main/setup.py#L36 Are there plans to upgrade soon?
non_process
upgrade to protobuf version i m using another library that pins protobuf at the latest version and i can t use this library with it because it pins it at are there plans to upgrade soon
0
17,198
23,716,253,180
IssuesEvent
2022-08-30 12:02:53
pnedev/comparePlus
https://api.github.com/repos/pnedev/comparePlus
closed
ARM64 build
fixed compatibility
Hi, I'm running Notepad++ on a Windows 11 VM on a Mac M1 and would like to use this plugin. However, it doesn't show up for me and it seems that's because it doesn't have an ARM build for it. Could you add an ARM build? Thanks!
True
ARM64 build - Hi, I'm running Notepad++ on a Windows 11 VM on a Mac M1 and would like to use this plugin. However, it doesn't show up for me and it seems that's because it doesn't have an ARM build for it. Could you add an ARM build? Thanks!
non_process
build hi i m running notepad on a windows vm on a mac and would like to use this plugin however it doesn t show up for me and it seems that s because it doesn t have an arm build for it could you add an arm build thanks
0
349,194
24,937,434,214
IssuesEvent
2022-10-31 16:08:48
pcaversaccio/create2deployer
https://api.github.com/repos/pcaversaccio/create2deployer
closed
Enforcing front-running prevention
documentation
`Create2` opens the door to front running, hence, enforcing the salt to contain the `msg.sender` or having to pass a signature could be ways to resolve this issue. * `msg.sender` implementation in [Create2Factory from Pr000xy repository](https://github.com/0age/Pr000xy/blob/c293074509c0f0d5fdf54c451ab39b2e00f48b7a/contracts/Create2Factory.sol#L108) * signature implementation in [OZ article](https://blog.openzeppelin.com/getting-the-most-out-of-create2/) It could be a problem to have a malicious address front-running the contract creation for reputation purpose only. Other than that, if I am not mistaken, it seems not to be a problem as you would get the exact same result contract wise. I might be missing a case ? What are your thoughts on this ?
1.0
Enforcing front-running prevention - `Create2` opens the door to front running, hence, enforcing the salt to contain the `msg.sender` or having to pass a signature could be ways to resolve this issue. * `msg.sender` implementation in [Create2Factory from Pr000xy repository](https://github.com/0age/Pr000xy/blob/c293074509c0f0d5fdf54c451ab39b2e00f48b7a/contracts/Create2Factory.sol#L108) * signature implementation in [OZ article](https://blog.openzeppelin.com/getting-the-most-out-of-create2/) It could be a problem to have a malicious address front-running the contract creation for reputation purpose only. Other than that, if I am not mistaken, it seems not to be a problem as you would get the exact same result contract wise. I might be missing a case ? What are your thoughts on this ?
non_process
enforcing front running prevention opens the door to front running hence enforcing the salt to contain the msg sender or having to pass a signature could be ways to resolve this issue msg sender implementation in signature implementation in it could be a problem to have a malicious address front running the contract creation for reputation purpose only other than that if i am not mistaken it seems not to be a problem as you would get the exact same result contract wise i might be missing a case what are your thoughts on this
0
171,174
6,480,520,425
IssuesEvent
2017-08-18 13:37:15
GoogleCloudPlatform/google-cloud-eclipse
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-eclipse
closed
Should we make google-cloud the fourth core library?
high priority question
After App Engine API, Endpoints, and Objectify: https://github.com/GoogleCloudPlatform/google-cloud-java <dependency> <groupId>com.google.cloud</groupId> <artifactId>google-cloud</artifactId> <version>0.9.3-alpha</version> </dependency> 1. Not all release quality yet. 2. Pulls in a lot of dependencies. E.g. a developer might want Datastore but not BigQuery. 3. Could we somehow parse this to offer the individual components in their own library group?
1.0
Should we make google-cloud the fourth core library? - After App Engine API, Endpoints, and Objectify: https://github.com/GoogleCloudPlatform/google-cloud-java <dependency> <groupId>com.google.cloud</groupId> <artifactId>google-cloud</artifactId> <version>0.9.3-alpha</version> </dependency> 1. Not all release quality yet. 2. Pulls in a lot of dependencies. E.g. a developer might want Datastore but not BigQuery. 3. Could we somehow parse this to offer the individual components in their own library group?
non_process
should we make google cloud the fourth core library after app engine api endpoints and objectify com google cloud google cloud alpha not all release quality yet pulls in a lot of dependencies e g a developer might want datastore but not bigquery could we somehow parse this to offer the individual components in their own library group
0
327,775
28,081,836,381
IssuesEvent
2023-03-30 07:03:55
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix random.test_randint
Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4493755053/jobs/7905367871" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4493755053/jobs/7905367871" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4519909128/jobs/7960678214" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details>
1.0
Fix random.test_randint - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4493755053/jobs/7905367871" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4493755053/jobs/7905367871" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4519909128/jobs/7960678214" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details>
non_process
fix random test randint tensorflow img src torch img src numpy img src jax img src not found not found
0
134,764
18,504,337,390
IssuesEvent
2021-10-19 16:46:16
Seagate/cortx-s3server
https://api.github.com/repos/Seagate/cortx-s3server
opened
CVE-2021-22946 (High) detected in curl_7.58.0-2ubuntu3.16_amd64.deb
security vulnerability
## CVE-2021-22946 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>curl_7.58.0-2ubuntu3.16_amd64.deb</b></p></summary> <p>command line tool for transferring data with URL syntax</p> <p>Library home page: <a href="http://security.ubuntu.com/ubuntu/pool/main/c/curl/curl_7.58.0-2ubuntu3.16_amd64.deb">http://security.ubuntu.com/ubuntu/pool/main/c/curl/curl_7.58.0-2ubuntu3.16_amd64.deb</a></p> <p> Dependency Hierarchy: - :x: **curl_7.58.0-2ubuntu3.16_amd64.deb** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-s3server/commit/e7eb2e0abd58bb71075411e5ca632d5713875f3d">e7eb2e0abd58bb71075411e5ca632d5713875f3d</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A user can tell curl >= 7.20.0 and <= 7.78.0 to require a successful upgrade to TLS when speaking to an IMAP, POP3 or FTP server (`--ssl-reqd` on the command line or`CURLOPT_USE_SSL` set to `CURLUSESSL_CONTROL` or `CURLUSESSL_ALL` withlibcurl). This requirement could be bypassed if the server would return a properly crafted but perfectly legitimate response.This flaw would then make curl silently continue its operations **withoutTLS** contrary to the instructions and expectations, exposing possibly sensitive data in clear text over the network. <p>Publish Date: 2021-09-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22946>CVE-2021-22946</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/1334111">https://hackerone.com/reports/1334111</a></p> <p>Release Date: 2021-09-29</p> <p>Fix Resolution: curl-7_79_0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Debian","packageName":"curl","packageVersion":"7.58.0-2ubuntu3.16","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"curl:7.58.0-2ubuntu3.16","isMinimumFixVersionAvailable":true,"minimumFixVersion":"curl-7_79_0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-22946","vulnerabilityDetails":"A user can tell curl \u003e\u003d 7.20.0 and \u003c\u003d 7.78.0 to require a successful upgrade to TLS when speaking to an IMAP, POP3 or FTP server (`--ssl-reqd` on the command line or`CURLOPT_USE_SSL` set to `CURLUSESSL_CONTROL` or `CURLUSESSL_ALL` withlibcurl). This requirement could be bypassed if the server would return a properly crafted but perfectly legitimate response.This flaw would then make curl silently continue its operations **withoutTLS** contrary to the instructions and expectations, exposing possibly sensitive data in clear text over the network.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22946","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-22946 (High) detected in curl_7.58.0-2ubuntu3.16_amd64.deb - ## CVE-2021-22946 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>curl_7.58.0-2ubuntu3.16_amd64.deb</b></p></summary> <p>command line tool for transferring data with URL syntax</p> <p>Library home page: <a href="http://security.ubuntu.com/ubuntu/pool/main/c/curl/curl_7.58.0-2ubuntu3.16_amd64.deb">http://security.ubuntu.com/ubuntu/pool/main/c/curl/curl_7.58.0-2ubuntu3.16_amd64.deb</a></p> <p> Dependency Hierarchy: - :x: **curl_7.58.0-2ubuntu3.16_amd64.deb** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-s3server/commit/e7eb2e0abd58bb71075411e5ca632d5713875f3d">e7eb2e0abd58bb71075411e5ca632d5713875f3d</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A user can tell curl >= 7.20.0 and <= 7.78.0 to require a successful upgrade to TLS when speaking to an IMAP, POP3 or FTP server (`--ssl-reqd` on the command line or`CURLOPT_USE_SSL` set to `CURLUSESSL_CONTROL` or `CURLUSESSL_ALL` withlibcurl). This requirement could be bypassed if the server would return a properly crafted but perfectly legitimate response.This flaw would then make curl silently continue its operations **withoutTLS** contrary to the instructions and expectations, exposing possibly sensitive data in clear text over the network. <p>Publish Date: 2021-09-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22946>CVE-2021-22946</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/1334111">https://hackerone.com/reports/1334111</a></p> <p>Release Date: 2021-09-29</p> <p>Fix Resolution: curl-7_79_0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Debian","packageName":"curl","packageVersion":"7.58.0-2ubuntu3.16","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"curl:7.58.0-2ubuntu3.16","isMinimumFixVersionAvailable":true,"minimumFixVersion":"curl-7_79_0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-22946","vulnerabilityDetails":"A user can tell curl \u003e\u003d 7.20.0 and \u003c\u003d 7.78.0 to require a successful upgrade to TLS when speaking to an IMAP, POP3 or FTP server (`--ssl-reqd` on the command line or`CURLOPT_USE_SSL` set to `CURLUSESSL_CONTROL` or `CURLUSESSL_ALL` withlibcurl). This requirement could be bypassed if the server would return a properly crafted but perfectly legitimate response.This flaw would then make curl silently continue its operations **withoutTLS** contrary to the instructions and expectations, exposing possibly sensitive data in clear text over the network.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22946","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in curl deb cve high severity vulnerability vulnerable library curl deb command line tool for transferring data with url syntax library home page a href dependency hierarchy x curl deb vulnerable library found in head commit a href found in base branch main vulnerability details a user can tell curl and to require a successful upgrade to tls when speaking to an imap or ftp server ssl reqd on the command line or curlopt use ssl set to curlusessl control or curlusessl all withlibcurl this requirement could be bypassed if the server would return a properly crafted but perfectly legitimate response this flaw would then make curl silently continue its operations withouttls contrary to the instructions and expectations exposing possibly sensitive data in clear text over the network publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution curl isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree curl isminimumfixversionavailable true minimumfixversion curl basebranches vulnerabilityidentifier cve vulnerabilitydetails a user can tell curl and to require a successful upgrade to tls when speaking to an imap or ftp server ssl reqd on the command line or curlopt use ssl set to curlusessl control or curlusessl all withlibcurl this requirement could be bypassed if the server would return a properly crafted but perfectly legitimate response this flaw would then make curl silently continue its operations withouttls contrary to the instructions and expectations exposing possibly sensitive data in clear text over the network vulnerabilityurl
0
214,849
7,278,757,322
IssuesEvent
2018-02-22 00:28:08
ClinGen/clincoded
https://api.github.com/repos/ClinGen/clincoded
opened
Provide Definition for Provisional in the VCI
R17 external curator priority: high requires specification variant curation interface
The VCI WG requested a definition for "Provisional" to be displayed in the Evaluation Summary of the VCI. Text TBD.
1.0
Provide Definition for Provisional in the VCI - The VCI WG requested a definition for "Provisional" to be displayed in the Evaluation Summary of the VCI. Text TBD.
non_process
provide definition for provisional in the vci the vci wg requested a definition for provisional to be displayed in the evaluation summary of the vci text tbd
0
14,210
17,109,450,657
IssuesEvent
2021-07-10 02:08:08
pycaret/pycaret
https://api.github.com/repos/pycaret/pycaret
closed
Binning From the pipline having an issue if the feature removed previously by other pre-processing
no-issue-activity preprocessing
The function of binning for numeric features has an issue if a feature was previously removed by a pre-process step. It is suggested to do a check before processing the feature rather than taking what the user input on the `bin_numeric_features`. The updated code demonstrates the concept for how to fix it. `` pycaret/internal/preprocess.py line 1232 if len(self.features_to_discretize) > 0: columns_to_process=[] # place holder for all the features for their binns self.binns = [] for i in self.features_to_discretize: try: # get numbr of binns hist, _ = np.histogram(data[i], bins="sturges") #this line generates an error when `i` is not available from the dataset if len(hist)>2: # adding a fix in case n_hist <2 self.binns.append(len(hist)) get_logger().info(f"Complete binning {i} with {len(hist)} bins") columns_to_process.append(i) else: get_logger().info(f"Skip binning {i} with {len(hist)} bins") except: get_logger().info(f"Unable to process {i} as the column is not available from the obtained dataset, please check if the column previously removed") # how many colums to deal with self.features_to_discretize = columns_to_process ``
1.0
Binning From the pipline having an issue if the feature removed previously by other pre-processing - The function of binning for numeric features has an issue if a feature was previously removed by a pre-process step. It is suggested to do a check before processing the feature rather than taking what the user input on the `bin_numeric_features`. The updated code demonstrates the concept for how to fix it. `` pycaret/internal/preprocess.py line 1232 if len(self.features_to_discretize) > 0: columns_to_process=[] # place holder for all the features for their binns self.binns = [] for i in self.features_to_discretize: try: # get numbr of binns hist, _ = np.histogram(data[i], bins="sturges") #this line generates an error when `i` is not available from the dataset if len(hist)>2: # adding a fix in case n_hist <2 self.binns.append(len(hist)) get_logger().info(f"Complete binning {i} with {len(hist)} bins") columns_to_process.append(i) else: get_logger().info(f"Skip binning {i} with {len(hist)} bins") except: get_logger().info(f"Unable to process {i} as the column is not available from the obtained dataset, please check if the column previously removed") # how many colums to deal with self.features_to_discretize = columns_to_process ``
process
binning from the pipline having an issue if the feature removed previously by other pre processing the function of binning for numeric features has an issue if a feature was previously removed by a pre process step it is suggested to do a check before processing the feature rather than taking what the user input on the bin numeric features the updated code demonstrates the concept for how to fix it pycaret internal preprocess py line if len self features to discretize columns to process place holder for all the features for their binns self binns for i in self features to discretize try get numbr of binns hist np histogram data bins sturges this line generates an error when i is not available from the dataset if len hist adding a fix in case n hist self binns append len hist get logger info f complete binning i with len hist bins columns to process append i else get logger info f skip binning i with len hist bins except get logger info f unable to process i as the column is not available from the obtained dataset please check if the column previously removed how many colums to deal with self features to discretize columns to process
1
39,497
5,097,834,986
IssuesEvent
2017-01-03 22:50:24
MapStory/mapstory
https://api.github.com/repos/MapStory/mapstory
closed
Redesign and reexamine notifications in Composer
design story
- We need to evaluate positioning collisions with loading/notification toasts, legend (currently uses “Toaster”, open to modifying or building custom solution) - Nothing currently exists for this, would be custom built, otherwise no technical feasibility concerns - Might need to re-evaluate Toaster library and how notifications work
1.0
Redesign and reexamine notifications in Composer - - We need to evaluate positioning collisions with loading/notification toasts, legend (currently uses “Toaster”, open to modifying or building custom solution) - Nothing currently exists for this, would be custom built, otherwise no technical feasibility concerns - Might need to re-evaluate Toaster library and how notifications work
non_process
redesign and reexamine notifications in composer we need to evaluate positioning collisions with loading notification toasts legend currently uses “toaster” open to modifying or building custom solution nothing currently exists for this would be custom built otherwise no technical feasibility concerns might need to re evaluate toaster library and how notifications work
0
461,956
13,238,936,664
IssuesEvent
2020-08-19 01:55:15
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
opened
Taint Analysis Not Working with Raw Templates Generated from a Query
Area/Language Priority/Blocker Type/Bug
Following code:- ```ballerina resource function addEmployeeBatch(http:Caller caller, http:Request req, Employee[] payload) returns error? { sql:ParameterizedQuery[] batchQuery = from var emp in payload select `INSERT INTO Employee VALUES (${emp.id}, ${emp.name}, ${emp.age}, ${emp.team})`; _ = check empDB->batchExecute(batchQuery); } ``` The tainted values passed in from "payload", it's data is transferred to "batchQuery" and this compiles without untainiting the input.
1.0
Taint Analysis Not Working with Raw Templates Generated from a Query - Following code:- ```ballerina resource function addEmployeeBatch(http:Caller caller, http:Request req, Employee[] payload) returns error? { sql:ParameterizedQuery[] batchQuery = from var emp in payload select `INSERT INTO Employee VALUES (${emp.id}, ${emp.name}, ${emp.age}, ${emp.team})`; _ = check empDB->batchExecute(batchQuery); } ``` The tainted values passed in from "payload", it's data is transferred to "batchQuery" and this compiles without untainiting the input.
non_process
taint analysis not working with raw templates generated from a query following code ballerina resource function addemployeebatch http caller caller http request req employee payload returns error sql parameterizedquery batchquery from var emp in payload select insert into employee values emp id emp name emp age emp team check empdb batchexecute batchquery the tainted values passed in from payload it s data is transferred to batchquery and this compiles without untainiting the input
0
22,030
30,544,561,094
IssuesEvent
2023-07-20 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 20 Jul 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Source-Free Domain Adaptive Fundus Image Segmentation with Class-Balanced Mean Teacher - **Authors:** Longxiang Tang, Kai Li, Chunming He, Yulun Zhang, Xiu Li - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09973 - **Pdf link:** https://arxiv.org/pdf/2307.09973 - **Abstract** This paper studies source-free domain adaptive fundus image segmentation which aims to adapt a pretrained fundus segmentation model to a target domain using unlabeled images. This is a challenging task because it is highly risky to adapt a model only using unlabeled data. Most existing methods tackle this task mainly by designing techniques to carefully generate pseudo labels from the model's predictions and use the pseudo labels to train the model. While often obtaining positive adaption effects, these methods suffer from two major issues. First, they tend to be fairly unstable - incorrect pseudo labels abruptly emerged may cause a catastrophic impact on the model. Second, they fail to consider the severe class imbalance of fundus images where the foreground (e.g., cup) region is usually very small. This paper aims to address these two issues by proposing the Class-Balanced Mean Teacher (CBMT) model. CBMT addresses the unstable issue by proposing a weak-strong augmented mean teacher learning scheme where only the teacher model generates pseudo labels from weakly augmented images to train a student model that takes strongly augmented images as input. The teacher is updated as the moving average of the instantly trained student, which could be noisy. This prevents the teacher model from being abruptly impacted by incorrect pseudo-labels. For the class imbalance issue, CBMT proposes a novel loss calibration approach to highlight foreground classes according to global statistics. Experiments show that CBMT well addresses these two issues and outperforms existing methods on multiple benchmarks. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### SAMConvex: Fast Discrete Optimization for CT Registration using Self-supervised Anatomical Embedding and Correlation Pyramid - **Authors:** Zi Li, Lin Tian, Tony C. W. Mok, Xiaoyu Bai, Puyang Wang, Jia Ge, Jingren Zhou, Le Lu, Xianghua Ye, Ke Yan, Dakai Jin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09727 - **Pdf link:** https://arxiv.org/pdf/2307.09727 - **Abstract** Estimating displacement vector field via a cost volume computed in the feature space has shown great success in image registration, but it suffers excessive computation burdens. Moreover, existing feature descriptors only extract local features incapable of representing the global semantic information, which is especially important for solving large transformations. To address the discussed issues, we propose SAMConvex, a fast coarse-to-fine discrete optimization method for CT registration that includes a decoupled convex optimization procedure to obtain deformation fields based on a self-supervised anatomical embedding (SAM) feature extractor that captures both local and global information. To be specific, SAMConvex extracts per-voxel features and builds 6D correlation volumes based on SAM features, and iteratively updates a flow field by performing lookups on the correlation volumes with a coarse-to-fine scheme. SAMConvex outperforms the state-of-the-art learning-based methods and optimization-based methods over two inter-patient registration datasets (Abdomen CT and HeadNeck CT) and one intra-patient registration dataset (Lung CT). Moreover, as an optimization-based method, SAMConvex only takes $\sim2$s ($\sim5s$ with instance optimization) for one paired images. ### Density-invariant Features for Distant Point Cloud Registration - **Authors:** Quan Liu, Hongzi Zhu, Yunsong Zhou, Hongyang Li, Shan Chang, Minyi Guo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09788 - **Pdf link:** https://arxiv.org/pdf/2307.09788 - **Abstract** Registration of distant outdoor LiDAR point clouds is crucial to extending the 3D vision of collaborative autonomous vehicles, and yet is challenging due to small overlapping area and a huge disparity between observed point densities. In this paper, we propose Group-wise Contrastive Learning (GCL) scheme to extract density-invariant geometric features to register distant outdoor LiDAR point clouds. We mark through theoretical analysis and experiments that, contrastive positives should be independent and identically distributed (i.i.d.), in order to train densityinvariant feature extractors. We propose upon the conclusion a simple yet effective training scheme to force the feature of multiple point clouds in the same spatial location (referred to as positive groups) to be similar, which naturally avoids the sampling bias introduced by a pair of point clouds to conform with the i.i.d. principle. The resulting fully-convolutional feature extractor is more powerful and density-invariant than state-of-the-art methods, improving the registration recall of distant scenarios on KITTI and nuScenes benchmarks by 40.9% and 26.9%, respectively. The code will be open-sourced. ### LDP: Language-driven Dual-Pixel Image Defocus Deblurring Network - **Authors:** Hao Yang, Liyuan Pan, Yan Yang, Miaomiao Liu, - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09815 - **Pdf link:** https://arxiv.org/pdf/2307.09815 - **Abstract** Recovering sharp images from dual-pixel (DP) pairs with disparity-dependent blur is a challenging task.~Existing blur map-based deblurring methods have demonstrated promising results. In this paper, we propose, to the best of our knowledge, the first framework to introduce the contrastive language-image pre-training framework (CLIP) to achieve accurate blur map estimation from DP pairs unsupervisedly. To this end, we first carefully design text prompts to enable CLIP to understand blur-related geometric prior knowledge from the DP pair. Then, we propose a format to input stereo DP pair to the CLIP without any fine-tuning, where the CLIP is pre-trained on monocular images. Given the estimated blur map, we introduce a blur-prior attention block, a blur-weighting loss and a blur-aware loss to recover the all-in-focus image. Our method achieves state-of-the-art performance in extensive experiments. ### Towards Fair Face Verification: An In-depth Analysis of Demographic Biases - **Authors:** Ioannis Sarridis, Christos Koutlis, Symeon Papadopoulos, Christos Diou - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.10011 - **Pdf link:** https://arxiv.org/pdf/2307.10011 - **Abstract** Deep learning-based person identification and verification systems have remarkably improved in terms of accuracy in recent years; however, such systems, including widely popular cloud-based solutions, have been found to exhibit significant biases related to race, age, and gender, a problem that requires in-depth exploration and solutions. This paper presents an in-depth analysis, with a particular emphasis on the intersectionality of these demographic factors. Intersectional bias refers to the performance discrepancies w.r.t. the different combinations of race, age, and gender groups, an area relatively unexplored in current literature. Furthermore, the reliance of most state-of-the-art approaches on accuracy as the principal evaluation metric often masks significant demographic disparities in performance. To counter this crucial limitation, we incorporate five additional metrics in our quantitative analysis, including disparate impact and mistreatment metrics, which are typically ignored by the relevant fairness-aware approaches. Results on the Racial Faces in-the-Wild (RFW) benchmark indicate pervasive biases in face recognition systems, extending beyond race, with different demographic factors yielding significantly disparate outcomes. In particular, Africans demonstrate an 11.25% lower True Positive Rate (TPR) compared to Caucasians, while only a 3.51% accuracy drop is observed. Even more concerning, the intersections of multiple protected groups, such as African females over 60 years old, demonstrate a +39.89% disparate mistreatment rate compared to the highest Caucasians rate. By shedding light on these biases and their implications, this paper aims to stimulate further research towards developing fairer, more equitable face recognition and verification systems. ### Adversarial Latent Autoencoder with Self-Attention for Structural Image Synthesis - **Authors:** Jiajie Fan, Laure Vuaille, Hao Wang, Thomas Bäck - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computational Engineering, Finance, and Science (cs.CE); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.10166 - **Pdf link:** https://arxiv.org/pdf/2307.10166 - **Abstract** Generative Engineering Design approaches driven by Deep Generative Models (DGM) have been proposed to facilitate industrial engineering processes. In such processes, designs often come in the form of images, such as blueprints, engineering drawings, and CAD models depending on the level of detail. DGMs have been successfully employed for synthesis of natural images, e.g., displaying animals, human faces and landscapes. However, industrial design images are fundamentally different from natural scenes in that they contain rich structural patterns and long-range dependencies, which are challenging for convolution-based DGMs to generate. Moreover, DGM-driven generation process is typically triggered based on random noisy inputs, which outputs unpredictable samples and thus cannot perform an efficient industrial design exploration. We tackle these challenges by proposing a novel model Self-Attention Adversarial Latent Autoencoder (SA-ALAE), which allows generating feasible design images of complex engineering parts. With SA-ALAE, users can not only explore novel variants of an existing design, but also control the generation process by operating in latent space. The potential of SA-ALAE is shown by generating engineering blueprints in a real automotive design task. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### Adversarial Bayesian Augmentation for Single-Source Domain Generalization - **Authors:** Sheng Cheng, Tejas Gokhale, Yezhou Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09520 - **Pdf link:** https://arxiv.org/pdf/2307.09520 - **Abstract** Generalizing to unseen image domains is a challenging problem primarily due to the lack of diverse training data, inaccessible target data, and the large domain shift that may exist in many real-world settings. As such data augmentation is a critical component of domain generalization methods that seek to address this problem. We present Adversarial Bayesian Augmentation (ABA), a novel algorithm that learns to generate image augmentations in the challenging single-source domain generalization setting. ABA draws on the strengths of adversarial learning and Bayesian neural networks to guide the generation of diverse data augmentations -- these synthesized image domains aid the classifier in generalizing to unseen domains. We demonstrate the strength of ABA on several types of domain shift including style shift, subpopulation shift, and shift in the medical imaging setting. ABA outperforms all previous state-of-the-art methods, including pre-specified augmentations, pixel-based and convolutional-based augmentations. ### Transient Neural Radiance Fields for Lidar View Synthesis and 3D Reconstruction - **Authors:** Anagh Malik, Parsa Mirdehghan, Sotiris Nousias, Kiriakos N. Kutulakos, David B. Lindell - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.09555 - **Pdf link:** https://arxiv.org/pdf/2307.09555 - **Abstract** Neural radiance fields (NeRFs) have become a ubiquitous tool for modeling scene appearance and geometry from multiview imagery. Recent work has also begun to explore how to use additional supervision from lidar or depth sensor measurements in the NeRF framework. However, previous lidar-supervised NeRFs focus on rendering conventional camera imagery and use lidar-derived point cloud data as auxiliary supervision; thus, they fail to incorporate the underlying image formation model of the lidar. Here, we propose a novel method for rendering transient NeRFs that take as input the raw, time-resolved photon count histograms measured by a single-photon lidar system, and we seek to render such histograms from novel views. Different from conventional NeRFs, the approach relies on a time-resolved version of the volume rendering equation to render the lidar measurements and capture transient light transport phenomena at picosecond timescales. We evaluate our method on a first-of-its-kind dataset of simulated and captured transient multiview scans from a prototype single-photon lidar. Overall, our work brings NeRFs to a new dimension of imaging at transient timescales, newly enabling rendering of transient imagery from novel views. Additionally, we show that our approach recovers improved geometry and conventional appearance compared to point cloud-based supervision when training on few input viewpoints. Transient NeRFs may be especially useful for applications which seek to simulate raw lidar measurements for downstream tasks in autonomous driving, robotics, and remote sensing. ### Watch out Venomous Snake Species: A Solution to SnakeCLEF2023 - **Authors:** Feiran Hu, Peng Wang, Yangyang Li, Chenlong Duan, Zijian Zhu, Fei Wang, Faen Zhang, Yong Li, Xiu-Shen Wei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09748 - **Pdf link:** https://arxiv.org/pdf/2307.09748 - **Abstract** The SnakeCLEF2023 competition aims to the development of advanced algorithms for snake species identification through the analysis of images and accompanying metadata. This paper presents a method leveraging utilization of both images and metadata. Modern CNN models and strong data augmentation are utilized to learn better representation of images. To relieve the challenge of long-tailed distribution, seesaw loss is utilized in our method. We also design a light model to calculate prior probabilities using metadata features extracted from CLIP in post processing stage. Besides, we attach more importance to venomous species by assigning venomous species labels to some examples that model is uncertain about. Our method achieves 91.31% score of the final metric combined of F1 and other metrics on private leaderboard, which is the 1st place among the participators. The code is available at https://github.com/xiaoxsparraw/CLEF2023. ### Adversarial Latent Autoencoder with Self-Attention for Structural Image Synthesis - **Authors:** Jiajie Fan, Laure Vuaille, Hao Wang, Thomas Bäck - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computational Engineering, Finance, and Science (cs.CE); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.10166 - **Pdf link:** https://arxiv.org/pdf/2307.10166 - **Abstract** Generative Engineering Design approaches driven by Deep Generative Models (DGM) have been proposed to facilitate industrial engineering processes. In such processes, designs often come in the form of images, such as blueprints, engineering drawings, and CAD models depending on the level of detail. DGMs have been successfully employed for synthesis of natural images, e.g., displaying animals, human faces and landscapes. However, industrial design images are fundamentally different from natural scenes in that they contain rich structural patterns and long-range dependencies, which are challenging for convolution-based DGMs to generate. Moreover, DGM-driven generation process is typically triggered based on random noisy inputs, which outputs unpredictable samples and thus cannot perform an efficient industrial design exploration. We tackle these challenges by proposing a novel model Self-Attention Adversarial Latent Autoencoder (SA-ALAE), which allows generating feasible design images of complex engineering parts. With SA-ALAE, users can not only explore novel variants of an existing design, but also control the generation process by operating in latent space. The potential of SA-ALAE is shown by generating engineering blueprints in a real automotive design task. ## Keyword: raw image There is no result
2.0
New submissions for Thu, 20 Jul 23 - ## Keyword: events ### Source-Free Domain Adaptive Fundus Image Segmentation with Class-Balanced Mean Teacher - **Authors:** Longxiang Tang, Kai Li, Chunming He, Yulun Zhang, Xiu Li - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09973 - **Pdf link:** https://arxiv.org/pdf/2307.09973 - **Abstract** This paper studies source-free domain adaptive fundus image segmentation which aims to adapt a pretrained fundus segmentation model to a target domain using unlabeled images. This is a challenging task because it is highly risky to adapt a model only using unlabeled data. Most existing methods tackle this task mainly by designing techniques to carefully generate pseudo labels from the model's predictions and use the pseudo labels to train the model. While often obtaining positive adaption effects, these methods suffer from two major issues. First, they tend to be fairly unstable - incorrect pseudo labels abruptly emerged may cause a catastrophic impact on the model. Second, they fail to consider the severe class imbalance of fundus images where the foreground (e.g., cup) region is usually very small. This paper aims to address these two issues by proposing the Class-Balanced Mean Teacher (CBMT) model. CBMT addresses the unstable issue by proposing a weak-strong augmented mean teacher learning scheme where only the teacher model generates pseudo labels from weakly augmented images to train a student model that takes strongly augmented images as input. The teacher is updated as the moving average of the instantly trained student, which could be noisy. This prevents the teacher model from being abruptly impacted by incorrect pseudo-labels. For the class imbalance issue, CBMT proposes a novel loss calibration approach to highlight foreground classes according to global statistics. Experiments show that CBMT well addresses these two issues and outperforms existing methods on multiple benchmarks. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### SAMConvex: Fast Discrete Optimization for CT Registration using Self-supervised Anatomical Embedding and Correlation Pyramid - **Authors:** Zi Li, Lin Tian, Tony C. W. Mok, Xiaoyu Bai, Puyang Wang, Jia Ge, Jingren Zhou, Le Lu, Xianghua Ye, Ke Yan, Dakai Jin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09727 - **Pdf link:** https://arxiv.org/pdf/2307.09727 - **Abstract** Estimating displacement vector field via a cost volume computed in the feature space has shown great success in image registration, but it suffers excessive computation burdens. Moreover, existing feature descriptors only extract local features incapable of representing the global semantic information, which is especially important for solving large transformations. To address the discussed issues, we propose SAMConvex, a fast coarse-to-fine discrete optimization method for CT registration that includes a decoupled convex optimization procedure to obtain deformation fields based on a self-supervised anatomical embedding (SAM) feature extractor that captures both local and global information. To be specific, SAMConvex extracts per-voxel features and builds 6D correlation volumes based on SAM features, and iteratively updates a flow field by performing lookups on the correlation volumes with a coarse-to-fine scheme. SAMConvex outperforms the state-of-the-art learning-based methods and optimization-based methods over two inter-patient registration datasets (Abdomen CT and HeadNeck CT) and one intra-patient registration dataset (Lung CT). Moreover, as an optimization-based method, SAMConvex only takes $\sim2$s ($\sim5s$ with instance optimization) for one paired images. ### Density-invariant Features for Distant Point Cloud Registration - **Authors:** Quan Liu, Hongzi Zhu, Yunsong Zhou, Hongyang Li, Shan Chang, Minyi Guo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09788 - **Pdf link:** https://arxiv.org/pdf/2307.09788 - **Abstract** Registration of distant outdoor LiDAR point clouds is crucial to extending the 3D vision of collaborative autonomous vehicles, and yet is challenging due to small overlapping area and a huge disparity between observed point densities. In this paper, we propose Group-wise Contrastive Learning (GCL) scheme to extract density-invariant geometric features to register distant outdoor LiDAR point clouds. We mark through theoretical analysis and experiments that, contrastive positives should be independent and identically distributed (i.i.d.), in order to train densityinvariant feature extractors. We propose upon the conclusion a simple yet effective training scheme to force the feature of multiple point clouds in the same spatial location (referred to as positive groups) to be similar, which naturally avoids the sampling bias introduced by a pair of point clouds to conform with the i.i.d. principle. The resulting fully-convolutional feature extractor is more powerful and density-invariant than state-of-the-art methods, improving the registration recall of distant scenarios on KITTI and nuScenes benchmarks by 40.9% and 26.9%, respectively. The code will be open-sourced. ### LDP: Language-driven Dual-Pixel Image Defocus Deblurring Network - **Authors:** Hao Yang, Liyuan Pan, Yan Yang, Miaomiao Liu, - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09815 - **Pdf link:** https://arxiv.org/pdf/2307.09815 - **Abstract** Recovering sharp images from dual-pixel (DP) pairs with disparity-dependent blur is a challenging task.~Existing blur map-based deblurring methods have demonstrated promising results. In this paper, we propose, to the best of our knowledge, the first framework to introduce the contrastive language-image pre-training framework (CLIP) to achieve accurate blur map estimation from DP pairs unsupervisedly. To this end, we first carefully design text prompts to enable CLIP to understand blur-related geometric prior knowledge from the DP pair. Then, we propose a format to input stereo DP pair to the CLIP without any fine-tuning, where the CLIP is pre-trained on monocular images. Given the estimated blur map, we introduce a blur-prior attention block, a blur-weighting loss and a blur-aware loss to recover the all-in-focus image. Our method achieves state-of-the-art performance in extensive experiments. ### Towards Fair Face Verification: An In-depth Analysis of Demographic Biases - **Authors:** Ioannis Sarridis, Christos Koutlis, Symeon Papadopoulos, Christos Diou - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.10011 - **Pdf link:** https://arxiv.org/pdf/2307.10011 - **Abstract** Deep learning-based person identification and verification systems have remarkably improved in terms of accuracy in recent years; however, such systems, including widely popular cloud-based solutions, have been found to exhibit significant biases related to race, age, and gender, a problem that requires in-depth exploration and solutions. This paper presents an in-depth analysis, with a particular emphasis on the intersectionality of these demographic factors. Intersectional bias refers to the performance discrepancies w.r.t. the different combinations of race, age, and gender groups, an area relatively unexplored in current literature. Furthermore, the reliance of most state-of-the-art approaches on accuracy as the principal evaluation metric often masks significant demographic disparities in performance. To counter this crucial limitation, we incorporate five additional metrics in our quantitative analysis, including disparate impact and mistreatment metrics, which are typically ignored by the relevant fairness-aware approaches. Results on the Racial Faces in-the-Wild (RFW) benchmark indicate pervasive biases in face recognition systems, extending beyond race, with different demographic factors yielding significantly disparate outcomes. In particular, Africans demonstrate an 11.25% lower True Positive Rate (TPR) compared to Caucasians, while only a 3.51% accuracy drop is observed. Even more concerning, the intersections of multiple protected groups, such as African females over 60 years old, demonstrate a +39.89% disparate mistreatment rate compared to the highest Caucasians rate. By shedding light on these biases and their implications, this paper aims to stimulate further research towards developing fairer, more equitable face recognition and verification systems. ### Adversarial Latent Autoencoder with Self-Attention for Structural Image Synthesis - **Authors:** Jiajie Fan, Laure Vuaille, Hao Wang, Thomas Bäck - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computational Engineering, Finance, and Science (cs.CE); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.10166 - **Pdf link:** https://arxiv.org/pdf/2307.10166 - **Abstract** Generative Engineering Design approaches driven by Deep Generative Models (DGM) have been proposed to facilitate industrial engineering processes. In such processes, designs often come in the form of images, such as blueprints, engineering drawings, and CAD models depending on the level of detail. DGMs have been successfully employed for synthesis of natural images, e.g., displaying animals, human faces and landscapes. However, industrial design images are fundamentally different from natural scenes in that they contain rich structural patterns and long-range dependencies, which are challenging for convolution-based DGMs to generate. Moreover, DGM-driven generation process is typically triggered based on random noisy inputs, which outputs unpredictable samples and thus cannot perform an efficient industrial design exploration. We tackle these challenges by proposing a novel model Self-Attention Adversarial Latent Autoencoder (SA-ALAE), which allows generating feasible design images of complex engineering parts. With SA-ALAE, users can not only explore novel variants of an existing design, but also control the generation process by operating in latent space. The potential of SA-ALAE is shown by generating engineering blueprints in a real automotive design task. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### Adversarial Bayesian Augmentation for Single-Source Domain Generalization - **Authors:** Sheng Cheng, Tejas Gokhale, Yezhou Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09520 - **Pdf link:** https://arxiv.org/pdf/2307.09520 - **Abstract** Generalizing to unseen image domains is a challenging problem primarily due to the lack of diverse training data, inaccessible target data, and the large domain shift that may exist in many real-world settings. As such data augmentation is a critical component of domain generalization methods that seek to address this problem. We present Adversarial Bayesian Augmentation (ABA), a novel algorithm that learns to generate image augmentations in the challenging single-source domain generalization setting. ABA draws on the strengths of adversarial learning and Bayesian neural networks to guide the generation of diverse data augmentations -- these synthesized image domains aid the classifier in generalizing to unseen domains. We demonstrate the strength of ABA on several types of domain shift including style shift, subpopulation shift, and shift in the medical imaging setting. ABA outperforms all previous state-of-the-art methods, including pre-specified augmentations, pixel-based and convolutional-based augmentations. ### Transient Neural Radiance Fields for Lidar View Synthesis and 3D Reconstruction - **Authors:** Anagh Malik, Parsa Mirdehghan, Sotiris Nousias, Kiriakos N. Kutulakos, David B. Lindell - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.09555 - **Pdf link:** https://arxiv.org/pdf/2307.09555 - **Abstract** Neural radiance fields (NeRFs) have become a ubiquitous tool for modeling scene appearance and geometry from multiview imagery. Recent work has also begun to explore how to use additional supervision from lidar or depth sensor measurements in the NeRF framework. However, previous lidar-supervised NeRFs focus on rendering conventional camera imagery and use lidar-derived point cloud data as auxiliary supervision; thus, they fail to incorporate the underlying image formation model of the lidar. Here, we propose a novel method for rendering transient NeRFs that take as input the raw, time-resolved photon count histograms measured by a single-photon lidar system, and we seek to render such histograms from novel views. Different from conventional NeRFs, the approach relies on a time-resolved version of the volume rendering equation to render the lidar measurements and capture transient light transport phenomena at picosecond timescales. We evaluate our method on a first-of-its-kind dataset of simulated and captured transient multiview scans from a prototype single-photon lidar. Overall, our work brings NeRFs to a new dimension of imaging at transient timescales, newly enabling rendering of transient imagery from novel views. Additionally, we show that our approach recovers improved geometry and conventional appearance compared to point cloud-based supervision when training on few input viewpoints. Transient NeRFs may be especially useful for applications which seek to simulate raw lidar measurements for downstream tasks in autonomous driving, robotics, and remote sensing. ### Watch out Venomous Snake Species: A Solution to SnakeCLEF2023 - **Authors:** Feiran Hu, Peng Wang, Yangyang Li, Chenlong Duan, Zijian Zhu, Fei Wang, Faen Zhang, Yong Li, Xiu-Shen Wei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09748 - **Pdf link:** https://arxiv.org/pdf/2307.09748 - **Abstract** The SnakeCLEF2023 competition aims to the development of advanced algorithms for snake species identification through the analysis of images and accompanying metadata. This paper presents a method leveraging utilization of both images and metadata. Modern CNN models and strong data augmentation are utilized to learn better representation of images. To relieve the challenge of long-tailed distribution, seesaw loss is utilized in our method. We also design a light model to calculate prior probabilities using metadata features extracted from CLIP in post processing stage. Besides, we attach more importance to venomous species by assigning venomous species labels to some examples that model is uncertain about. Our method achieves 91.31% score of the final metric combined of F1 and other metrics on private leaderboard, which is the 1st place among the participators. The code is available at https://github.com/xiaoxsparraw/CLEF2023. ### Adversarial Latent Autoencoder with Self-Attention for Structural Image Synthesis - **Authors:** Jiajie Fan, Laure Vuaille, Hao Wang, Thomas Bäck - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computational Engineering, Finance, and Science (cs.CE); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.10166 - **Pdf link:** https://arxiv.org/pdf/2307.10166 - **Abstract** Generative Engineering Design approaches driven by Deep Generative Models (DGM) have been proposed to facilitate industrial engineering processes. In such processes, designs often come in the form of images, such as blueprints, engineering drawings, and CAD models depending on the level of detail. DGMs have been successfully employed for synthesis of natural images, e.g., displaying animals, human faces and landscapes. However, industrial design images are fundamentally different from natural scenes in that they contain rich structural patterns and long-range dependencies, which are challenging for convolution-based DGMs to generate. Moreover, DGM-driven generation process is typically triggered based on random noisy inputs, which outputs unpredictable samples and thus cannot perform an efficient industrial design exploration. We tackle these challenges by proposing a novel model Self-Attention Adversarial Latent Autoencoder (SA-ALAE), which allows generating feasible design images of complex engineering parts. With SA-ALAE, users can not only explore novel variants of an existing design, but also control the generation process by operating in latent space. The potential of SA-ALAE is shown by generating engineering blueprints in a real automotive design task. ## Keyword: raw image There is no result
process
new submissions for thu jul keyword events source free domain adaptive fundus image segmentation with class balanced mean teacher authors longxiang tang kai li chunming he yulun zhang xiu li subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract this paper studies source free domain adaptive fundus image segmentation which aims to adapt a pretrained fundus segmentation model to a target domain using unlabeled images this is a challenging task because it is highly risky to adapt a model only using unlabeled data most existing methods tackle this task mainly by designing techniques to carefully generate pseudo labels from the model s predictions and use the pseudo labels to train the model while often obtaining positive adaption effects these methods suffer from two major issues first they tend to be fairly unstable incorrect pseudo labels abruptly emerged may cause a catastrophic impact on the model second they fail to consider the severe class imbalance of fundus images where the foreground e g cup region is usually very small this paper aims to address these two issues by proposing the class balanced mean teacher cbmt model cbmt addresses the unstable issue by proposing a weak strong augmented mean teacher learning scheme where only the teacher model generates pseudo labels from weakly augmented images to train a student model that takes strongly augmented images as input the teacher is updated as the moving average of the instantly trained student which could be noisy this prevents the teacher model from being abruptly impacted by incorrect pseudo labels for the class imbalance issue cbmt proposes a novel loss calibration approach to highlight foreground classes according to global statistics experiments show that cbmt well addresses these two issues and outperforms existing methods on multiple benchmarks keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp samconvex fast discrete optimization for ct registration using self supervised anatomical embedding and correlation pyramid authors zi li lin tian tony c w mok xiaoyu bai puyang wang jia ge jingren zhou le lu xianghua ye ke yan dakai jin subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract estimating displacement vector field via a cost volume computed in the feature space has shown great success in image registration but it suffers excessive computation burdens moreover existing feature descriptors only extract local features incapable of representing the global semantic information which is especially important for solving large transformations to address the discussed issues we propose samconvex a fast coarse to fine discrete optimization method for ct registration that includes a decoupled convex optimization procedure to obtain deformation fields based on a self supervised anatomical embedding sam feature extractor that captures both local and global information to be specific samconvex extracts per voxel features and builds correlation volumes based on sam features and iteratively updates a flow field by performing lookups on the correlation volumes with a coarse to fine scheme samconvex outperforms the state of the art learning based methods and optimization based methods over two inter patient registration datasets abdomen ct and headneck ct and one intra patient registration dataset lung ct moreover as an optimization based method samconvex only takes s with instance optimization for one paired images density invariant features for distant point cloud registration authors quan liu hongzi zhu yunsong zhou hongyang li shan chang minyi guo subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract registration of distant outdoor lidar point clouds is crucial to extending the vision of collaborative autonomous vehicles and yet is challenging due to small overlapping area and a huge disparity between observed point densities in this paper we propose group wise contrastive learning gcl scheme to extract density invariant geometric features to register distant outdoor lidar point clouds we mark through theoretical analysis and experiments that contrastive positives should be independent and identically distributed i i d in order to train densityinvariant feature extractors we propose upon the conclusion a simple yet effective training scheme to force the feature of multiple point clouds in the same spatial location referred to as positive groups to be similar which naturally avoids the sampling bias introduced by a pair of point clouds to conform with the i i d principle the resulting fully convolutional feature extractor is more powerful and density invariant than state of the art methods improving the registration recall of distant scenarios on kitti and nuscenes benchmarks by and respectively the code will be open sourced ldp language driven dual pixel image defocus deblurring network authors hao yang liyuan pan yan yang miaomiao liu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract recovering sharp images from dual pixel dp pairs with disparity dependent blur is a challenging task existing blur map based deblurring methods have demonstrated promising results in this paper we propose to the best of our knowledge the first framework to introduce the contrastive language image pre training framework clip to achieve accurate blur map estimation from dp pairs unsupervisedly to this end we first carefully design text prompts to enable clip to understand blur related geometric prior knowledge from the dp pair then we propose a format to input stereo dp pair to the clip without any fine tuning where the clip is pre trained on monocular images given the estimated blur map we introduce a blur prior attention block a blur weighting loss and a blur aware loss to recover the all in focus image our method achieves state of the art performance in extensive experiments towards fair face verification an in depth analysis of demographic biases authors ioannis sarridis christos koutlis symeon papadopoulos christos diou subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deep learning based person identification and verification systems have remarkably improved in terms of accuracy in recent years however such systems including widely popular cloud based solutions have been found to exhibit significant biases related to race age and gender a problem that requires in depth exploration and solutions this paper presents an in depth analysis with a particular emphasis on the intersectionality of these demographic factors intersectional bias refers to the performance discrepancies w r t the different combinations of race age and gender groups an area relatively unexplored in current literature furthermore the reliance of most state of the art approaches on accuracy as the principal evaluation metric often masks significant demographic disparities in performance to counter this crucial limitation we incorporate five additional metrics in our quantitative analysis including disparate impact and mistreatment metrics which are typically ignored by the relevant fairness aware approaches results on the racial faces in the wild rfw benchmark indicate pervasive biases in face recognition systems extending beyond race with different demographic factors yielding significantly disparate outcomes in particular africans demonstrate an lower true positive rate tpr compared to caucasians while only a accuracy drop is observed even more concerning the intersections of multiple protected groups such as african females over years old demonstrate a disparate mistreatment rate compared to the highest caucasians rate by shedding light on these biases and their implications this paper aims to stimulate further research towards developing fairer more equitable face recognition and verification systems adversarial latent autoencoder with self attention for structural image synthesis authors jiajie fan laure vuaille hao wang thomas bäck subjects computer vision and pattern recognition cs cv computational engineering finance and science cs ce image and video processing eess iv arxiv link pdf link abstract generative engineering design approaches driven by deep generative models dgm have been proposed to facilitate industrial engineering processes in such processes designs often come in the form of images such as blueprints engineering drawings and cad models depending on the level of detail dgms have been successfully employed for synthesis of natural images e g displaying animals human faces and landscapes however industrial design images are fundamentally different from natural scenes in that they contain rich structural patterns and long range dependencies which are challenging for convolution based dgms to generate moreover dgm driven generation process is typically triggered based on random noisy inputs which outputs unpredictable samples and thus cannot perform an efficient industrial design exploration we tackle these challenges by proposing a novel model self attention adversarial latent autoencoder sa alae which allows generating feasible design images of complex engineering parts with sa alae users can not only explore novel variants of an existing design but also control the generation process by operating in latent space the potential of sa alae is shown by generating engineering blueprints in a real automotive design task keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw adversarial bayesian augmentation for single source domain generalization authors sheng cheng tejas gokhale yezhou yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract generalizing to unseen image domains is a challenging problem primarily due to the lack of diverse training data inaccessible target data and the large domain shift that may exist in many real world settings as such data augmentation is a critical component of domain generalization methods that seek to address this problem we present adversarial bayesian augmentation aba a novel algorithm that learns to generate image augmentations in the challenging single source domain generalization setting aba draws on the strengths of adversarial learning and bayesian neural networks to guide the generation of diverse data augmentations these synthesized image domains aid the classifier in generalizing to unseen domains we demonstrate the strength of aba on several types of domain shift including style shift subpopulation shift and shift in the medical imaging setting aba outperforms all previous state of the art methods including pre specified augmentations pixel based and convolutional based augmentations transient neural radiance fields for lidar view synthesis and reconstruction authors anagh malik parsa mirdehghan sotiris nousias kiriakos n kutulakos david b lindell subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract neural radiance fields nerfs have become a ubiquitous tool for modeling scene appearance and geometry from multiview imagery recent work has also begun to explore how to use additional supervision from lidar or depth sensor measurements in the nerf framework however previous lidar supervised nerfs focus on rendering conventional camera imagery and use lidar derived point cloud data as auxiliary supervision thus they fail to incorporate the underlying image formation model of the lidar here we propose a novel method for rendering transient nerfs that take as input the raw time resolved photon count histograms measured by a single photon lidar system and we seek to render such histograms from novel views different from conventional nerfs the approach relies on a time resolved version of the volume rendering equation to render the lidar measurements and capture transient light transport phenomena at picosecond timescales we evaluate our method on a first of its kind dataset of simulated and captured transient multiview scans from a prototype single photon lidar overall our work brings nerfs to a new dimension of imaging at transient timescales newly enabling rendering of transient imagery from novel views additionally we show that our approach recovers improved geometry and conventional appearance compared to point cloud based supervision when training on few input viewpoints transient nerfs may be especially useful for applications which seek to simulate raw lidar measurements for downstream tasks in autonomous driving robotics and remote sensing watch out venomous snake species a solution to authors feiran hu peng wang yangyang li chenlong duan zijian zhu fei wang faen zhang yong li xiu shen wei subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the competition aims to the development of advanced algorithms for snake species identification through the analysis of images and accompanying metadata this paper presents a method leveraging utilization of both images and metadata modern cnn models and strong data augmentation are utilized to learn better representation of images to relieve the challenge of long tailed distribution seesaw loss is utilized in our method we also design a light model to calculate prior probabilities using metadata features extracted from clip in post processing stage besides we attach more importance to venomous species by assigning venomous species labels to some examples that model is uncertain about our method achieves score of the final metric combined of and other metrics on private leaderboard which is the place among the participators the code is available at adversarial latent autoencoder with self attention for structural image synthesis authors jiajie fan laure vuaille hao wang thomas bäck subjects computer vision and pattern recognition cs cv computational engineering finance and science cs ce image and video processing eess iv arxiv link pdf link abstract generative engineering design approaches driven by deep generative models dgm have been proposed to facilitate industrial engineering processes in such processes designs often come in the form of images such as blueprints engineering drawings and cad models depending on the level of detail dgms have been successfully employed for synthesis of natural images e g displaying animals human faces and landscapes however industrial design images are fundamentally different from natural scenes in that they contain rich structural patterns and long range dependencies which are challenging for convolution based dgms to generate moreover dgm driven generation process is typically triggered based on random noisy inputs which outputs unpredictable samples and thus cannot perform an efficient industrial design exploration we tackle these challenges by proposing a novel model self attention adversarial latent autoencoder sa alae which allows generating feasible design images of complex engineering parts with sa alae users can not only explore novel variants of an existing design but also control the generation process by operating in latent space the potential of sa alae is shown by generating engineering blueprints in a real automotive design task keyword raw image there is no result
1
21,924
30,446,557,972
IssuesEvent
2023-07-15 18:48:17
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1b15 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b15", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:205", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmp9lmfpi_y/pyutils" } }```
1.0
pyutils 0.0.1b15 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b15", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:205", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmp9lmfpi_y/pyutils" } }```
process
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pytils python utils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp y pyutils
1
15,855
20,033,047,189
IssuesEvent
2022-02-02 08:55:53
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Support `process.chdir` in worker thread
wontfix feature request process worker
### What is the problem this feature will solve? Some test runners like [AVA](https://github.com/avajs/ava) run tests files in separate worker threads, for improved performance and context isolation. It's common in tests to use `process.chdir()` to change the directory to be able to test how a certain thing works when running from a different place, or maybe you just want it to run in a temporary directory. ### What is the feature you are proposing to solve the problem? I'm aware `process.chdir()` works per process and not per thread, but maybe it would be possible to expose a `process.withCwd(tempDirectory, () => {})` method, where the CWD change is only available inside the given synchronous callback? ### What alternatives have you considered? > multi-process model + shared memory. That lets you support SharedArrayBuffer without the headaches of the multi-thread model. > - https://github.com/nodejs/node/issues/13143#issuecomment-303041478
1.0
Support `process.chdir` in worker thread - ### What is the problem this feature will solve? Some test runners like [AVA](https://github.com/avajs/ava) run tests files in separate worker threads, for improved performance and context isolation. It's common in tests to use `process.chdir()` to change the directory to be able to test how a certain thing works when running from a different place, or maybe you just want it to run in a temporary directory. ### What is the feature you are proposing to solve the problem? I'm aware `process.chdir()` works per process and not per thread, but maybe it would be possible to expose a `process.withCwd(tempDirectory, () => {})` method, where the CWD change is only available inside the given synchronous callback? ### What alternatives have you considered? > multi-process model + shared memory. That lets you support SharedArrayBuffer without the headaches of the multi-thread model. > - https://github.com/nodejs/node/issues/13143#issuecomment-303041478
process
support process chdir in worker thread what is the problem this feature will solve some test runners like run tests files in separate worker threads for improved performance and context isolation it s common in tests to use process chdir to change the directory to be able to test how a certain thing works when running from a different place or maybe you just want it to run in a temporary directory what is the feature you are proposing to solve the problem i m aware process chdir works per process and not per thread but maybe it would be possible to expose a process withcwd tempdirectory method where the cwd change is only available inside the given synchronous callback what alternatives have you considered multi process model shared memory that lets you support sharedarraybuffer without the headaches of the multi thread model
1
3,766
6,736,988,700
IssuesEvent
2017-10-19 07:39:21
jimbrown75/Permit-Vision-Enhancements
https://api.github.com/repos/jimbrown75/Permit-Vision-Enhancements
closed
ICC: Add "Squeeze Off" to Isolation Methods
2 - Medium Priority enhancement Should Fix Take Forward Verified by PTW Process Lead
The Isolation Method "Squeeze off" is not available in V8 Enterprise Build. It was added to v7 as an Isolation Method to support work requirements for QGC. We will be adding QGC to the Enterprise build in 2018, so we need to have this added to v8. Isolation Type: **Mechanical/Process** Isolation Method: **Squeeze Off** Isolated State: **Squeezed Off** De-Isolated State: **Open** ![squeeze off](https://user-images.githubusercontent.com/23561839/30821370-5f5f6212-a1e2-11e7-82fd-415fbd5d109c.PNG)
1.0
ICC: Add "Squeeze Off" to Isolation Methods - The Isolation Method "Squeeze off" is not available in V8 Enterprise Build. It was added to v7 as an Isolation Method to support work requirements for QGC. We will be adding QGC to the Enterprise build in 2018, so we need to have this added to v8. Isolation Type: **Mechanical/Process** Isolation Method: **Squeeze Off** Isolated State: **Squeezed Off** De-Isolated State: **Open** ![squeeze off](https://user-images.githubusercontent.com/23561839/30821370-5f5f6212-a1e2-11e7-82fd-415fbd5d109c.PNG)
process
icc add squeeze off to isolation methods the isolation method squeeze off is not available in enterprise build it was added to as an isolation method to support work requirements for qgc we will be adding qgc to the enterprise build in so we need to have this added to isolation type mechanical process isolation method squeeze off isolated state squeezed off de isolated state open
1
12,847
15,228,614,591
IssuesEvent
2021-02-18 11:45:36
Figma-Linux/figma-linux
https://api.github.com/repos/Figma-Linux/figma-linux
opened
Current color palette resets after sync themes
Kind: Renderer Process Priority: High Status: To Do Type: Bug
* App version: 0.7.1 * The output of `screenfetch`: -- * Type of installed package (Snap, AppImage, deb, rpm, pacman): any **Bug description** If set any theme, and then click on the sync button, after that color palette of current theme will be resets to default palette for Desktop App UI. **Screenshots** Before click: ![screenshot_2021-02-18T14:43:09](https://user-images.githubusercontent.com/20404541/108352358-c7ed2380-71f7-11eb-90c9-826ea6197cb8.jpg) After click: ![screenshot_2021-02-18T14:43:18](https://user-images.githubusercontent.com/20404541/108352367-cae81400-71f7-11eb-81dc-d4b79e94266b.jpg)
1.0
Current color palette resets after sync themes - * App version: 0.7.1 * The output of `screenfetch`: -- * Type of installed package (Snap, AppImage, deb, rpm, pacman): any **Bug description** If set any theme, and then click on the sync button, after that color palette of current theme will be resets to default palette for Desktop App UI. **Screenshots** Before click: ![screenshot_2021-02-18T14:43:09](https://user-images.githubusercontent.com/20404541/108352358-c7ed2380-71f7-11eb-90c9-826ea6197cb8.jpg) After click: ![screenshot_2021-02-18T14:43:18](https://user-images.githubusercontent.com/20404541/108352367-cae81400-71f7-11eb-81dc-d4b79e94266b.jpg)
process
current color palette resets after sync themes app version the output of screenfetch type of installed package snap appimage deb rpm pacman any bug description if set any theme and then click on the sync button after that color palette of current theme will be resets to default palette for desktop app ui screenshots before click after click
1
652,053
21,520,098,908
IssuesEvent
2022-04-28 13:32:32
desktop/desktop
https://api.github.com/repos/desktop/desktop
closed
Can't merge conflicts
bug priority-2
### The problem I am trying to confirm merge conflicts but the program's option is way below my screen to confirm. I am stuck ### Release version 3.0.0 (x64) ### Operating system Windows 11 ### Steps to reproduce the behavior _No response_ ### Log files _No response_ ### Screenshots ![Screenshot (214)](https://user-images.githubusercontent.com/68997177/165659522-91c02926-7f74-45ae-b51b-5571a3b05eaf.png) ### Additional context _No response_
1.0
Can't merge conflicts - ### The problem I am trying to confirm merge conflicts but the program's option is way below my screen to confirm. I am stuck ### Release version 3.0.0 (x64) ### Operating system Windows 11 ### Steps to reproduce the behavior _No response_ ### Log files _No response_ ### Screenshots ![Screenshot (214)](https://user-images.githubusercontent.com/68997177/165659522-91c02926-7f74-45ae-b51b-5571a3b05eaf.png) ### Additional context _No response_
non_process
can t merge conflicts the problem i am trying to confirm merge conflicts but the program s option is way below my screen to confirm i am stuck release version operating system windows steps to reproduce the behavior no response log files no response screenshots additional context no response
0
197,887
22,614,064,951
IssuesEvent
2022-06-29 19:59:44
bcgov/cloud-pathfinder
https://api.github.com/repos/bcgov/cloud-pathfinder
opened
To Get Terraform Cloud STRA & SoAR Signed Off
Security
**Describe the issue** Complete TFC STRA and SoAR v1.0 and get signed off **Additional context** Add any other context, attachments or screenshots **Definition of done** - Complete STRA and SoAR - Share with PO, Nick and Tech Team - Incorporate review comment - Get STRA and SoAR number from IMB - Submit for Approval
True
To Get Terraform Cloud STRA & SoAR Signed Off - **Describe the issue** Complete TFC STRA and SoAR v1.0 and get signed off **Additional context** Add any other context, attachments or screenshots **Definition of done** - Complete STRA and SoAR - Share with PO, Nick and Tech Team - Incorporate review comment - Get STRA and SoAR number from IMB - Submit for Approval
non_process
to get terraform cloud stra soar signed off describe the issue complete tfc stra and soar and get signed off additional context add any other context attachments or screenshots definition of done complete stra and soar share with po nick and tech team incorporate review comment get stra and soar number from imb submit for approval
0
11,982
14,737,103,959
IssuesEvent
2021-01-07 00:52:55
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
123-E0394 Subout.com - Delete payment
anc-ops anc-process anp-important ant-bug ant-support
In GitLab by @kdjstudios on Apr 17, 2018, 09:34 **Submitted by:** "Kimberly Gagner" <kimberly.gagner@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-17-35638/conversation **Server:** Internal **Client/Site:** 123 **Account:** 123-E0394 **Issue:** Our client Subout.com made a credit card payment from the SAB online payment portal and made a typo in the payment amount. The payment made on April 16, 2018, in the amount of $43,497.00 was reversed and should be removed from SAB (I do not have the ability to delete the payment).
1.0
123-E0394 Subout.com - Delete payment - In GitLab by @kdjstudios on Apr 17, 2018, 09:34 **Submitted by:** "Kimberly Gagner" <kimberly.gagner@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-17-35638/conversation **Server:** Internal **Client/Site:** 123 **Account:** 123-E0394 **Issue:** Our client Subout.com made a credit card payment from the SAB online payment portal and made a typo in the payment amount. The payment made on April 16, 2018, in the amount of $43,497.00 was reversed and should be removed from SAB (I do not have the ability to delete the payment).
process
subout com delete payment in gitlab by kdjstudios on apr submitted by kimberly gagner helpdesk server internal client site account issue our client subout com made a credit card payment from the sab online payment portal and made a typo in the payment amount the payment made on april in the amount of was reversed and should be removed from sab i do not have the ability to delete the payment
1
248,037
26,772,569,264
IssuesEvent
2023-01-31 15:03:44
lukebrogan-mend/WebGoat
https://api.github.com/repos/lukebrogan-mend/WebGoat
opened
spring-boot-starter-undertow-2.7.1.jar: 2 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-boot-starter-undertow-2.7.1.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/undertow/undertow-core/2.2.18.Final/undertow-core-2.2.18.Final.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/WebGoat/commit/19074b5312ffc4968df5adb4e31c3b0d7f44ef61">19074b5312ffc4968df5adb4e31c3b0d7f44ef61</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (spring-boot-starter-undertow version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-0084](https://www.mend.io/vulnerability-database/CVE-2022-0084) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | xnio-api-3.8.7.Final.jar | Transitive | N/A* | &#10060; | | [CVE-2022-2053](https://www.mend.io/vulnerability-database/CVE-2022-2053) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | undertow-core-2.2.18.Final.jar | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-0084</summary> ### Vulnerable Library - <b>xnio-api-3.8.7.Final.jar</b></p> <p>The API JAR of the XNIO project</p> <p>Library home page: <a href="http://www.jboss.org/xnio">http://www.jboss.org/xnio</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jboss/xnio/xnio-api/3.8.7.Final/xnio-api-3.8.7.Final.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-undertow-2.7.1.jar (Root Library) - undertow-core-2.2.18.Final.jar - :x: **xnio-api-3.8.7.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/WebGoat/commit/19074b5312ffc4968df5adb4e31c3b0d7f44ef61">19074b5312ffc4968df5adb4e31c3b0d7f44ef61</a></p> <p>Found in base branch: <b>develop</b></p> </p> <p></p> ### Vulnerability Details <p> A flaw was found in XNIO, specifically in the notifyReadClosed method. The issue revealed this method was logging a message to another expected end. This flaw allows an attacker to send flawed requests to a server, possibly causing log contention-related performance concerns or an unwanted disk fill-up. <p>Publish Date: 2022-08-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0084>CVE-2022-0084</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-08-26</p> <p>Fix Resolution: org.jboss.xnio:xnio-api:3.8.8.Final</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-2053</summary> ### Vulnerable Library - <b>undertow-core-2.2.18.Final.jar</b></p> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/undertow/undertow-core/2.2.18.Final/undertow-core-2.2.18.Final.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-undertow-2.7.1.jar (Root Library) - :x: **undertow-core-2.2.18.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/WebGoat/commit/19074b5312ffc4968df5adb4e31c3b0d7f44ef61">19074b5312ffc4968df5adb4e31c3b0d7f44ef61</a></p> <p>Found in base branch: <b>develop</b></p> </p> <p></p> ### Vulnerability Details <p> When a POST request comes through AJP and the request exceeds the max-post-size limit (maxEntitySize), Undertow's AjpServerRequestConduit implementation closes a connection without sending any response to the client/proxy. This behavior results in that a front-end proxy marking the backend worker (application server) as an error state and not forward requests to the worker for a while. In mod_cluster, this continues until the next STATUS request (10 seconds intervals) from the application server updates the server state. So, in the worst case, it can result in "All workers are in error state" and mod_cluster responds "503 Service Unavailable" for a while (up to 10 seconds). In mod_proxy_balancer, it does not forward requests to the worker until the "retry" timeout passes. However, luckily, mod_proxy_balancer has "forcerecovery" setting (On by default; this parameter can force the immediate recovery of all workers without considering the retry parameter of the workers if all workers of a balancer are in error state.). So, unlike mod_cluster, mod_proxy_balancer does not result in responding "503 Service Unavailable". An attacker could use this behavior to send a malicious request and trigger server errors, resulting in DoS (denial of service). This flaw was fixed in Undertow 2.2.19.Final, Undertow 2.3.0.Alpha2. <p>Publish Date: 2022-08-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-2053>CVE-2022-2053</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-95rf-557x-44g5">https://github.com/advisories/GHSA-95rf-557x-44g5</a></p> <p>Release Date: 2022-08-05</p> <p>Fix Resolution: io.undertow:undertow-core:2.2.19.Final</p> </p> <p></p> </details>
True
spring-boot-starter-undertow-2.7.1.jar: 2 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-boot-starter-undertow-2.7.1.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/undertow/undertow-core/2.2.18.Final/undertow-core-2.2.18.Final.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/WebGoat/commit/19074b5312ffc4968df5adb4e31c3b0d7f44ef61">19074b5312ffc4968df5adb4e31c3b0d7f44ef61</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (spring-boot-starter-undertow version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-0084](https://www.mend.io/vulnerability-database/CVE-2022-0084) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | xnio-api-3.8.7.Final.jar | Transitive | N/A* | &#10060; | | [CVE-2022-2053](https://www.mend.io/vulnerability-database/CVE-2022-2053) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | undertow-core-2.2.18.Final.jar | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-0084</summary> ### Vulnerable Library - <b>xnio-api-3.8.7.Final.jar</b></p> <p>The API JAR of the XNIO project</p> <p>Library home page: <a href="http://www.jboss.org/xnio">http://www.jboss.org/xnio</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jboss/xnio/xnio-api/3.8.7.Final/xnio-api-3.8.7.Final.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-undertow-2.7.1.jar (Root Library) - undertow-core-2.2.18.Final.jar - :x: **xnio-api-3.8.7.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/WebGoat/commit/19074b5312ffc4968df5adb4e31c3b0d7f44ef61">19074b5312ffc4968df5adb4e31c3b0d7f44ef61</a></p> <p>Found in base branch: <b>develop</b></p> </p> <p></p> ### Vulnerability Details <p> A flaw was found in XNIO, specifically in the notifyReadClosed method. The issue revealed this method was logging a message to another expected end. This flaw allows an attacker to send flawed requests to a server, possibly causing log contention-related performance concerns or an unwanted disk fill-up. <p>Publish Date: 2022-08-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0084>CVE-2022-0084</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-08-26</p> <p>Fix Resolution: org.jboss.xnio:xnio-api:3.8.8.Final</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-2053</summary> ### Vulnerable Library - <b>undertow-core-2.2.18.Final.jar</b></p> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/undertow/undertow-core/2.2.18.Final/undertow-core-2.2.18.Final.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-undertow-2.7.1.jar (Root Library) - :x: **undertow-core-2.2.18.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/WebGoat/commit/19074b5312ffc4968df5adb4e31c3b0d7f44ef61">19074b5312ffc4968df5adb4e31c3b0d7f44ef61</a></p> <p>Found in base branch: <b>develop</b></p> </p> <p></p> ### Vulnerability Details <p> When a POST request comes through AJP and the request exceeds the max-post-size limit (maxEntitySize), Undertow's AjpServerRequestConduit implementation closes a connection without sending any response to the client/proxy. This behavior results in that a front-end proxy marking the backend worker (application server) as an error state and not forward requests to the worker for a while. In mod_cluster, this continues until the next STATUS request (10 seconds intervals) from the application server updates the server state. So, in the worst case, it can result in "All workers are in error state" and mod_cluster responds "503 Service Unavailable" for a while (up to 10 seconds). In mod_proxy_balancer, it does not forward requests to the worker until the "retry" timeout passes. However, luckily, mod_proxy_balancer has "forcerecovery" setting (On by default; this parameter can force the immediate recovery of all workers without considering the retry parameter of the workers if all workers of a balancer are in error state.). So, unlike mod_cluster, mod_proxy_balancer does not result in responding "503 Service Unavailable". An attacker could use this behavior to send a malicious request and trigger server errors, resulting in DoS (denial of service). This flaw was fixed in Undertow 2.2.19.Final, Undertow 2.3.0.Alpha2. <p>Publish Date: 2022-08-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-2053>CVE-2022-2053</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-95rf-557x-44g5">https://github.com/advisories/GHSA-95rf-557x-44g5</a></p> <p>Release Date: 2022-08-05</p> <p>Fix Resolution: io.undertow:undertow-core:2.2.19.Final</p> </p> <p></p> </details>
non_process
spring boot starter undertow jar vulnerabilities highest severity is vulnerable library spring boot starter undertow jar path to dependency file pom xml path to vulnerable library home wss scanner repository io undertow undertow core final undertow core final jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in spring boot starter undertow version remediation available high xnio api final jar transitive n a high undertow core final jar transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library xnio api final jar the api jar of the xnio project library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org jboss xnio xnio api final xnio api final jar dependency hierarchy spring boot starter undertow jar root library undertow core final jar x xnio api final jar vulnerable library found in head commit a href found in base branch develop vulnerability details a flaw was found in xnio specifically in the notifyreadclosed method the issue revealed this method was logging a message to another expected end this flaw allows an attacker to send flawed requests to a server possibly causing log contention related performance concerns or an unwanted disk fill up publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution org jboss xnio xnio api final cve vulnerable library undertow core final jar path to dependency file pom xml path to vulnerable library home wss scanner repository io undertow undertow core final undertow core final jar dependency hierarchy spring boot starter undertow jar root library x undertow core final jar vulnerable library found in head commit a href found in base branch develop vulnerability details when a post request comes through ajp and the request exceeds the max post size limit maxentitysize undertow s ajpserverrequestconduit implementation closes a connection without sending any response to the client proxy this behavior results in that a front end proxy marking the backend worker application server as an error state and not forward requests to the worker for a while in mod cluster this continues until the next status request seconds intervals from the application server updates the server state so in the worst case it can result in all workers are in error state and mod cluster responds service unavailable for a while up to seconds in mod proxy balancer it does not forward requests to the worker until the retry timeout passes however luckily mod proxy balancer has forcerecovery setting on by default this parameter can force the immediate recovery of all workers without considering the retry parameter of the workers if all workers of a balancer are in error state so unlike mod cluster mod proxy balancer does not result in responding service unavailable an attacker could use this behavior to send a malicious request and trigger server errors resulting in dos denial of service this flaw was fixed in undertow final undertow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io undertow undertow core final
0
293,568
8,997,711,504
IssuesEvent
2019-02-02 15:08:11
TailorMadeSoftwareHouse/AGEstateAgentsTenancyManagementSystem
https://api.github.com/repos/TailorMadeSoftwareHouse/AGEstateAgentsTenancyManagementSystem
closed
Add new fields to the landlords form.
High priority Ready for development
We need to add the following fields into the Add & Update Landlords Form: Admin Fee: £ Set Up fee: £ Inventory fee: £ Gas Cert Fee: £ EPC Fee: £ Tenant Reference Fee: £ Inventory Check out Fee: £ Electric Certificate Fee: £ Legionella Risk Assessment Fee: £
1.0
Add new fields to the landlords form. - We need to add the following fields into the Add & Update Landlords Form: Admin Fee: £ Set Up fee: £ Inventory fee: £ Gas Cert Fee: £ EPC Fee: £ Tenant Reference Fee: £ Inventory Check out Fee: £ Electric Certificate Fee: £ Legionella Risk Assessment Fee: £
non_process
add new fields to the landlords form we need to add the following fields into the add update landlords form admin fee £ set up fee £ inventory fee £ gas cert fee £ epc fee £ tenant reference fee £ inventory check out fee £ electric certificate fee £ legionella risk assessment fee £
0
139,148
18,840,233,556
IssuesEvent
2021-11-11 08:41:20
rammatzkvosky/Home-assignment
https://api.github.com/repos/rammatzkvosky/Home-assignment
opened
CVE-2019-14892 (High) detected in jackson-databind-2.8.7.jar
security vulnerability
## CVE-2019-14892 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.7.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Home-assignment/pom.xml</p> <p>Path to vulnerable library: itory/com/fasterxml/jackson/core/jackson-databind/2.8.7/jackson-databind-2.8.7.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/Home-assignment/commit/994eb8c8e34df13ffb792b62ec4c8e5c05738834">994eb8c8e34df13ffb792b62ec4c8e5c05738834</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2462">https://github.com/FasterXML/jackson-databind/issues/2462</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.7","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-14892","vulnerabilityDetails":"A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-14892 (High) detected in jackson-databind-2.8.7.jar - ## CVE-2019-14892 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.7.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Home-assignment/pom.xml</p> <p>Path to vulnerable library: itory/com/fasterxml/jackson/core/jackson-databind/2.8.7/jackson-databind-2.8.7.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/Home-assignment/commit/994eb8c8e34df13ffb792b62ec4c8e5c05738834">994eb8c8e34df13ffb792b62ec4c8e5c05738834</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2462">https://github.com/FasterXML/jackson-databind/issues/2462</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.7","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-14892","vulnerabilityDetails":"A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file home assignment pom xml path to vulnerable library itory com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code vulnerabilityurl
0
10,121
13,044,162,257
IssuesEvent
2020-07-29 03:47:31
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `StringStringTimeDiff` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `StringStringTimeDiff` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `StringStringTimeDiff` from TiDB - ## Description Port the scalar function `StringStringTimeDiff` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function stringstringtimediff from tidb description port the scalar function stringstringtimediff from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
1
73,503
24,663,659,121
IssuesEvent
2022-10-18 08:41:25
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Component: Chips
defect
### Describe the bug Firefox bug Provided we use p-chips with max set to 5. After reaching the fifth value and add this value by pressing `enter` the `p-focus` css class does not get removed from the `ul` element In Chrome and Edge Chromium it works as aspected. ### Environment Running Angular v14 and primeNG v14 ### Reproducer https://stackblitz.com/edit/primeng-chips-demo-hbcyme?file=src/app/app.component.html ### Angular version 14 ### PrimeNG version 14 ### Build / Runtime Angular CLI App ### Language TypeScript ### Node version (for AoT issues node --version) 14 ### Browser(s) Firefox 105 ### Steps to reproduce the behavior Add max attribute to the chips component, keep adding values and hit enter untill max is reached. In FF the class `p-focus` is still bound to the ul element while in Chrome end Edge the class has been removed. The example provided here is running on angular v13 and primeng v13, but it has the same behavior on v14 ### Expected behavior After reaching the max and hitting `enter` that the `p-focus` class has been removed from the `ul` element
1.0
Component: Chips - ### Describe the bug Firefox bug Provided we use p-chips with max set to 5. After reaching the fifth value and add this value by pressing `enter` the `p-focus` css class does not get removed from the `ul` element In Chrome and Edge Chromium it works as aspected. ### Environment Running Angular v14 and primeNG v14 ### Reproducer https://stackblitz.com/edit/primeng-chips-demo-hbcyme?file=src/app/app.component.html ### Angular version 14 ### PrimeNG version 14 ### Build / Runtime Angular CLI App ### Language TypeScript ### Node version (for AoT issues node --version) 14 ### Browser(s) Firefox 105 ### Steps to reproduce the behavior Add max attribute to the chips component, keep adding values and hit enter untill max is reached. In FF the class `p-focus` is still bound to the ul element while in Chrome end Edge the class has been removed. The example provided here is running on angular v13 and primeng v13, but it has the same behavior on v14 ### Expected behavior After reaching the max and hitting `enter` that the `p-focus` class has been removed from the `ul` element
non_process
component chips describe the bug firefox bug provided we use p chips with max set to after reaching the fifth value and add this value by pressing enter the p focus css class does not get removed from the ul element in chrome and edge chromium it works as aspected environment running angular and primeng reproducer angular version primeng version build runtime angular cli app language typescript node version for aot issues node version browser s firefox steps to reproduce the behavior add max attribute to the chips component keep adding values and hit enter untill max is reached in ff the class p focus is still bound to the ul element while in chrome end edge the class has been removed the example provided here is running on angular and primeng but it has the same behavior on expected behavior after reaching the max and hitting enter that the p focus class has been removed from the ul element
0
22,474
31,389,007,791
IssuesEvent
2023-08-26 05:07:49
serai-dex/serai
https://api.github.com/repos/serai-dex/serai
opened
ProcessorMessages have extraneous/suboptimized fields
improvement processor coordinator
As the processor and coordinator are e2e tested, certain fields have gone unused/been found suboptimal and worthy of revisiting. While we need to revisit all of them, I'd immediately like to flag `key`. `key` isn't needed for batches since only a single key will be used for batches at any given time (per network). It is needed for sign, if we sign TXs from multiple multisigs at once (such as while transitioning). Instead of key, we should use the much more minimal `Session`. That'd allow reverting https://github.com/serai-dex/serai/commit/f249e20028e8dca2218ea240cc20ed40791014a5 and reduce disk space usage of messages a good percent.
1.0
ProcessorMessages have extraneous/suboptimized fields - As the processor and coordinator are e2e tested, certain fields have gone unused/been found suboptimal and worthy of revisiting. While we need to revisit all of them, I'd immediately like to flag `key`. `key` isn't needed for batches since only a single key will be used for batches at any given time (per network). It is needed for sign, if we sign TXs from multiple multisigs at once (such as while transitioning). Instead of key, we should use the much more minimal `Session`. That'd allow reverting https://github.com/serai-dex/serai/commit/f249e20028e8dca2218ea240cc20ed40791014a5 and reduce disk space usage of messages a good percent.
process
processormessages have extraneous suboptimized fields as the processor and coordinator are tested certain fields have gone unused been found suboptimal and worthy of revisiting while we need to revisit all of them i d immediately like to flag key key isn t needed for batches since only a single key will be used for batches at any given time per network it is needed for sign if we sign txs from multiple multisigs at once such as while transitioning instead of key we should use the much more minimal session that d allow reverting and reduce disk space usage of messages a good percent
1
674,098
23,039,092,074
IssuesEvent
2022-07-22 23:31:45
meerk40t/meerk40t
https://api.github.com/repos/meerk40t/meerk40t
closed
Animations for Widgets/Scene
Type: Enhancement Status: Accepted Priority: Low
While it hasn't mattered much yet, the original code I wrote in android for scene and widget stuff had animations and I never ported any of that stuff over. Doing so would allow widgets access to animations. Basically they call `start()` this adds them into the timed elements list in scene which performs ticks in its own thread (this would likely be in `kernel.threaded()`) this thread adds and removes anything that are in the add-remove lists (in a synchronized fashion) and then performs the ticks on the list of animated widgets. These widgets are expected to call stop() when they are done but otherwise animate by updating their current position (where they are drawn), until they reach their goal position (where they stop). There's not much known need for this yet, but some animations like moving the laser head position over several ticks rather than instantly can help explain to the user what's going on. There's also some things like direction of marching ants to show the directionality of a curve that be useful in laser software. Widgets would need to implement: ```java public void start() { getScene().addTimedElement(this); } public void stop() { getScene().removeTimedElement(this); } void tick(); void onStartMain(); void onStartThreaded(); void onStopThreaded(); void onStopMain(); ``` Scene would need to permit widget animations. ```java @Override public void run() { ArrayList<TimedElement> runningElements = new ArrayList<>(); halting = false; while (true) { if (!addTimed.isEmpty()) { synchronized (SceneView.this) { for (int i = 0; i < addTimed.size(); i++) { TimedElement element = addTimed.get(i); if (!runningElements.contains(element)) { element.onStartThreaded(); runningElements.add(element); } } addTimed.clear(); } } if (!removeTimed.isEmpty()) { synchronized (SceneView.this) { for (int i = 0; i < removeTimed.size(); i++) { TimedElement element = removeTimed.get(i); element.onStopThreaded(); runningElements.remove(element); notifyStoppedElement(element); } removeTimed.clear(); goInvalidate(); } } if (halting) { synchronized (SceneView.this) { for (TimedElement widget : runningElements) { widget.onStopThreaded(); notifyStoppedElement(widget); } runningElements.clear(); goInvalidate(); halting = false; } } if (runningElements.isEmpty()) { break; } long timeBetweenTicks = System.currentTimeMillis() - lastTick; long waitTick = tick - timeBetweenTicks; try { Thread.sleep(Math.max(waitTick, TICK_REQUIRED)); } catch (InterruptedException e) { break; } lastTick = System.currentTimeMillis(); synchronized (SceneView.this) { for (TimedElement widget : runningElements) { widget.tick(); } } //not forced full revalidate, just partial if requested. goInvalidate(); } getActivity().runOnUiThread(new Runnable() { @Override public void run() { running = false; if (!addTimed.isEmpty()) startIfNeeded(); } }); } private void notifyStoppedElement(@NonNull final TimedElement element) { getActivity().runOnUiThread(new Runnable() { @Override public void run() { element.onStopMain(); goInvalidate(); } }); } public void addTimedElement(@NonNull TimedElement widget) { synchronized (SceneView.this) { if (!addTimed.contains(widget)) { widget.onStartMain(); addTimed.add(widget); } } startIfNeeded(); } public void removeTimedElement(TimedElement widget) { synchronized (SceneView.this) { if (running) { if (!removeTimed.contains(widget)) { removeTimed.add(widget); } } } } private void startIfNeeded() { synchronized (SceneView.this) { if (running) return; running = true; Thread thread = new Thread(this); thread.start(); } } private void halt() { if (running) { halting = true; } } ```
1.0
Animations for Widgets/Scene - While it hasn't mattered much yet, the original code I wrote in android for scene and widget stuff had animations and I never ported any of that stuff over. Doing so would allow widgets access to animations. Basically they call `start()` this adds them into the timed elements list in scene which performs ticks in its own thread (this would likely be in `kernel.threaded()`) this thread adds and removes anything that are in the add-remove lists (in a synchronized fashion) and then performs the ticks on the list of animated widgets. These widgets are expected to call stop() when they are done but otherwise animate by updating their current position (where they are drawn), until they reach their goal position (where they stop). There's not much known need for this yet, but some animations like moving the laser head position over several ticks rather than instantly can help explain to the user what's going on. There's also some things like direction of marching ants to show the directionality of a curve that be useful in laser software. Widgets would need to implement: ```java public void start() { getScene().addTimedElement(this); } public void stop() { getScene().removeTimedElement(this); } void tick(); void onStartMain(); void onStartThreaded(); void onStopThreaded(); void onStopMain(); ``` Scene would need to permit widget animations. ```java @Override public void run() { ArrayList<TimedElement> runningElements = new ArrayList<>(); halting = false; while (true) { if (!addTimed.isEmpty()) { synchronized (SceneView.this) { for (int i = 0; i < addTimed.size(); i++) { TimedElement element = addTimed.get(i); if (!runningElements.contains(element)) { element.onStartThreaded(); runningElements.add(element); } } addTimed.clear(); } } if (!removeTimed.isEmpty()) { synchronized (SceneView.this) { for (int i = 0; i < removeTimed.size(); i++) { TimedElement element = removeTimed.get(i); element.onStopThreaded(); runningElements.remove(element); notifyStoppedElement(element); } removeTimed.clear(); goInvalidate(); } } if (halting) { synchronized (SceneView.this) { for (TimedElement widget : runningElements) { widget.onStopThreaded(); notifyStoppedElement(widget); } runningElements.clear(); goInvalidate(); halting = false; } } if (runningElements.isEmpty()) { break; } long timeBetweenTicks = System.currentTimeMillis() - lastTick; long waitTick = tick - timeBetweenTicks; try { Thread.sleep(Math.max(waitTick, TICK_REQUIRED)); } catch (InterruptedException e) { break; } lastTick = System.currentTimeMillis(); synchronized (SceneView.this) { for (TimedElement widget : runningElements) { widget.tick(); } } //not forced full revalidate, just partial if requested. goInvalidate(); } getActivity().runOnUiThread(new Runnable() { @Override public void run() { running = false; if (!addTimed.isEmpty()) startIfNeeded(); } }); } private void notifyStoppedElement(@NonNull final TimedElement element) { getActivity().runOnUiThread(new Runnable() { @Override public void run() { element.onStopMain(); goInvalidate(); } }); } public void addTimedElement(@NonNull TimedElement widget) { synchronized (SceneView.this) { if (!addTimed.contains(widget)) { widget.onStartMain(); addTimed.add(widget); } } startIfNeeded(); } public void removeTimedElement(TimedElement widget) { synchronized (SceneView.this) { if (running) { if (!removeTimed.contains(widget)) { removeTimed.add(widget); } } } } private void startIfNeeded() { synchronized (SceneView.this) { if (running) return; running = true; Thread thread = new Thread(this); thread.start(); } } private void halt() { if (running) { halting = true; } } ```
non_process
animations for widgets scene while it hasn t mattered much yet the original code i wrote in android for scene and widget stuff had animations and i never ported any of that stuff over doing so would allow widgets access to animations basically they call start this adds them into the timed elements list in scene which performs ticks in its own thread this would likely be in kernel threaded this thread adds and removes anything that are in the add remove lists in a synchronized fashion and then performs the ticks on the list of animated widgets these widgets are expected to call stop when they are done but otherwise animate by updating their current position where they are drawn until they reach their goal position where they stop there s not much known need for this yet but some animations like moving the laser head position over several ticks rather than instantly can help explain to the user what s going on there s also some things like direction of marching ants to show the directionality of a curve that be useful in laser software widgets would need to implement java public void start getscene addtimedelement this public void stop getscene removetimedelement this void tick void onstartmain void onstartthreaded void onstopthreaded void onstopmain scene would need to permit widget animations java override public void run arraylist runningelements new arraylist halting false while true if addtimed isempty synchronized sceneview this for int i i addtimed size i timedelement element addtimed get i if runningelements contains element element onstartthreaded runningelements add element addtimed clear if removetimed isempty synchronized sceneview this for int i i removetimed size i timedelement element removetimed get i element onstopthreaded runningelements remove element notifystoppedelement element removetimed clear goinvalidate if halting synchronized sceneview this for timedelement widget runningelements widget onstopthreaded notifystoppedelement widget runningelements clear goinvalidate halting false if runningelements isempty break long timebetweenticks system currenttimemillis lasttick long waittick tick timebetweenticks try thread sleep math max waittick tick required catch interruptedexception e break lasttick system currenttimemillis synchronized sceneview this for timedelement widget runningelements widget tick not forced full revalidate just partial if requested goinvalidate getactivity runonuithread new runnable override public void run running false if addtimed isempty startifneeded private void notifystoppedelement nonnull final timedelement element getactivity runonuithread new runnable override public void run element onstopmain goinvalidate public void addtimedelement nonnull timedelement widget synchronized sceneview this if addtimed contains widget widget onstartmain addtimed add widget startifneeded public void removetimedelement timedelement widget synchronized sceneview this if running if removetimed contains widget removetimed add widget private void startifneeded synchronized sceneview this if running return running true thread thread new thread this thread start private void halt if running halting true
0
10,773
3,423,649,243
IssuesEvent
2015-12-09 08:16:29
roniRamon/big-project
https://api.github.com/repos/roniRamon/big-project
closed
Wording for tutorial (FS)
Documentation Adjustment
Please check and update wording for Welcome message, About page heading and tutorial if needed. For example, when and where should we tell the user about the python and should we warn him in the beginning about the final fight? Update FS accordingly.
1.0
Wording for tutorial (FS) - Please check and update wording for Welcome message, About page heading and tutorial if needed. For example, when and where should we tell the user about the python and should we warn him in the beginning about the final fight? Update FS accordingly.
non_process
wording for tutorial fs please check and update wording for welcome message about page heading and tutorial if needed for example when and where should we tell the user about the python and should we warn him in the beginning about the final fight update fs accordingly
0
22,406
31,142,291,837
IssuesEvent
2023-08-16 01:44:44
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky test: Timeout of 50ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves.
OS: windows process: flaky test topic: flake ❄️ stage: flake "topic: done()" stale
### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41309/workflows/91b07182-06e5-4d6c-88aa-8861ac4c660f/jobs/1710135 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/network/test/integration/connect_spec.ts#L14 ### Analysis <img width="1192" alt="Screen Shot 2022-08-05 at 12 34 23 PM" src="https://user-images.githubusercontent.com/26726429/183148315-5100912a-c49e-4037-b537-433da8c77775.png"> ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
1.0
Flaky test: Timeout of 50ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. - ### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41309/workflows/91b07182-06e5-4d6c-88aa-8861ac4c660f/jobs/1710135 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/network/test/integration/connect_spec.ts#L14 ### Analysis <img width="1192" alt="Screen Shot 2022-08-05 at 12 34 23 PM" src="https://user-images.githubusercontent.com/26726429/183148315-5100912a-c49e-4037-b537-433da8c77775.png"> ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
process
flaky test timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at pm src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
1
13,222
15,690,854,185
IssuesEvent
2021-03-25 17:10:49
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Upgrade to Helm 3
P3 enhancement process
**Problem** Our charts are still using the Helm 2 chart format. Helm 2 is deprecated. Luckily we use Helm 3 to actually deploy. **Solution** - [Migrate](https://helm.sh/docs/topics/v2_v3_migration/) chart config to helm 3 - Migrate Marketplace to v3 - Test migrated chart **Alternatives** **Additional Context**
1.0
Upgrade to Helm 3 - **Problem** Our charts are still using the Helm 2 chart format. Helm 2 is deprecated. Luckily we use Helm 3 to actually deploy. **Solution** - [Migrate](https://helm.sh/docs/topics/v2_v3_migration/) chart config to helm 3 - Migrate Marketplace to v3 - Test migrated chart **Alternatives** **Additional Context**
process
upgrade to helm problem our charts are still using the helm chart format helm is deprecated luckily we use helm to actually deploy solution chart config to helm migrate marketplace to test migrated chart alternatives additional context
1
1,294
3,829,970,485
IssuesEvent
2016-03-31 13:01:33
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
OTR GO:1990209 negative regulation by symbiont of RNA levels in host
Other term-related request question RNA processes
To whom it may concern, I am a student in the CACAO class at Texas A&M University. I was looking for a term in which a symbiont negatively regulates host transcription, and I found the following two GO terms: GO:0052026 modulation by symbiont of host transcription - any process in which an organism modulates the frequency, rate or extent of its host's transcription. The host is defined as the larger of the organisms involved in a symbiotic interaction. GO:1990209 negative regulation by symbiont of RNA levels in host - any process in which an organism stops, prevents, or reduces the frequency, rate or extent of the RNA levels in the host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction Given the existence of the term "modulation by symbiont of host transcription," I don't quite understand how "negative regulation by symbiont of RNA levels in host" fits in. The latter term seems to be somewhat vague - based on my knowledge, a lot of processes could cause an organism to stop/prevent/reduce the frequency/rate/extent of RNA levels, including negative regulation of transcription. I talked to Dr. Jim Hu and Suzi Aleksander about this, and they thought that the latter GO term sounds more like a phenotype or an artefact of an assay than a biological process or function. I am therefore interested in hearing your opinion on this GO term. Thanks, Lorna
1.0
OTR GO:1990209 negative regulation by symbiont of RNA levels in host - To whom it may concern, I am a student in the CACAO class at Texas A&M University. I was looking for a term in which a symbiont negatively regulates host transcription, and I found the following two GO terms: GO:0052026 modulation by symbiont of host transcription - any process in which an organism modulates the frequency, rate or extent of its host's transcription. The host is defined as the larger of the organisms involved in a symbiotic interaction. GO:1990209 negative regulation by symbiont of RNA levels in host - any process in which an organism stops, prevents, or reduces the frequency, rate or extent of the RNA levels in the host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction Given the existence of the term "modulation by symbiont of host transcription," I don't quite understand how "negative regulation by symbiont of RNA levels in host" fits in. The latter term seems to be somewhat vague - based on my knowledge, a lot of processes could cause an organism to stop/prevent/reduce the frequency/rate/extent of RNA levels, including negative regulation of transcription. I talked to Dr. Jim Hu and Suzi Aleksander about this, and they thought that the latter GO term sounds more like a phenotype or an artefact of an assay than a biological process or function. I am therefore interested in hearing your opinion on this GO term. Thanks, Lorna
process
otr go negative regulation by symbiont of rna levels in host to whom it may concern i am a student in the cacao class at texas a m university i was looking for a term in which a symbiont negatively regulates host transcription and i found the following two go terms go modulation by symbiont of host transcription any process in which an organism modulates the frequency rate or extent of its host s transcription the host is defined as the larger of the organisms involved in a symbiotic interaction go negative regulation by symbiont of rna levels in host any process in which an organism stops prevents or reduces the frequency rate or extent of the rna levels in the host organism the host is defined as the larger of the organisms involved in a symbiotic interaction given the existence of the term modulation by symbiont of host transcription i don t quite understand how negative regulation by symbiont of rna levels in host fits in the latter term seems to be somewhat vague based on my knowledge a lot of processes could cause an organism to stop prevent reduce the frequency rate extent of rna levels including negative regulation of transcription i talked to dr jim hu and suzi aleksander about this and they thought that the latter go term sounds more like a phenotype or an artefact of an assay than a biological process or function i am therefore interested in hearing your opinion on this go term thanks lorna
1
232,722
18,904,094,466
IssuesEvent
2021-11-16 06:47:30
HaxeFoundation/haxe
https://api.github.com/repos/HaxeFoundation/haxe
closed
unnecessary counters are added to the var names inside local functions
haxe-foundation-partner test-needed
this is minor and probably was there for a long time, but we might want to fix this someday: ```haxe class Test { static function main() { var f = function(a) { var b = a; return b; } var f2 = function(a) { var b = a; return b; } } } ``` generates: ```js Test.main = function() { var f = function(a) { var b = a; return b; }; var f2 = function(a1) { var b1 = a1; return b1; }; }; ``` it can get a bit ridiculous when macro-generating a lot of code, e.g.: ```js description1346.addInjectionMethod(function(target1083,injector1063) { target1083.campaignModel = injector1063.getInstanceForMapping("de.innogames.strategycity.campaignmap.model.CampaignModel"); target1083.userDataModel = injector1063.getInstanceForMapping("de.innogames.strategycity.main.model.UserDataModel"); }); ```
1.0
unnecessary counters are added to the var names inside local functions - this is minor and probably was there for a long time, but we might want to fix this someday: ```haxe class Test { static function main() { var f = function(a) { var b = a; return b; } var f2 = function(a) { var b = a; return b; } } } ``` generates: ```js Test.main = function() { var f = function(a) { var b = a; return b; }; var f2 = function(a1) { var b1 = a1; return b1; }; }; ``` it can get a bit ridiculous when macro-generating a lot of code, e.g.: ```js description1346.addInjectionMethod(function(target1083,injector1063) { target1083.campaignModel = injector1063.getInstanceForMapping("de.innogames.strategycity.campaignmap.model.CampaignModel"); target1083.userDataModel = injector1063.getInstanceForMapping("de.innogames.strategycity.main.model.UserDataModel"); }); ```
non_process
unnecessary counters are added to the var names inside local functions this is minor and probably was there for a long time but we might want to fix this someday haxe class test static function main var f function a var b a return b var function a var b a return b generates js test main function var f function a var b a return b var function var return it can get a bit ridiculous when macro generating a lot of code e g js addinjectionmethod function campaignmodel getinstanceformapping de innogames strategycity campaignmap model campaignmodel userdatamodel getinstanceformapping de innogames strategycity main model userdatamodel
0
4,302
7,196,601,692
IssuesEvent
2018-02-05 04:12:24
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
@font-face files omitted when building with parcel build index.html
#Bug CSS Preprocessing
```javascript import { h, app } from 'hyperapp'; import 'normalize.css'; import 'typeface-varela-round'; import './index.css'; ``` This works fine when running `parcel index.html` but it does not work using `parcel build index.html` Somehow, the typeface file gets omitted. The only way it'll work is if I manually copy the @font-face declaration to another css file that already contains some other css.
1.0
@font-face files omitted when building with parcel build index.html - ```javascript import { h, app } from 'hyperapp'; import 'normalize.css'; import 'typeface-varela-round'; import './index.css'; ``` This works fine when running `parcel index.html` but it does not work using `parcel build index.html` Somehow, the typeface file gets omitted. The only way it'll work is if I manually copy the @font-face declaration to another css file that already contains some other css.
process
font face files omitted when building with parcel build index html javascript import h app from hyperapp import normalize css import typeface varela round import index css this works fine when running parcel index html but it does not work using parcel build index html somehow the typeface file gets omitted the only way it ll work is if i manually copy the font face declaration to another css file that already contains some other css
1
95,081
10,865,586,060
IssuesEvent
2019-11-14 19:19:31
open-contracting/infrastructure
https://api.github.com/repos/open-contracting/infrastructure
closed
Link to OCDS publication policy guidance and add package metadata to worked example
bug documentation
@yolile noticed that the publication policy isn't mentioned in the documentation (though it's present in the schema). Since this is important metadata, we should mention it somewhere (likely linking to OCDS documentation to avoid duplication of content).
1.0
Link to OCDS publication policy guidance and add package metadata to worked example - @yolile noticed that the publication policy isn't mentioned in the documentation (though it's present in the schema). Since this is important metadata, we should mention it somewhere (likely linking to OCDS documentation to avoid duplication of content).
non_process
link to ocds publication policy guidance and add package metadata to worked example yolile noticed that the publication policy isn t mentioned in the documentation though it s present in the schema since this is important metadata we should mention it somewhere likely linking to ocds documentation to avoid duplication of content
0
6,860
9,997,402,682
IssuesEvent
2019-07-12 04:15:38
material-components/material-components-ios
https://api.github.com/repos/material-components/material-components-ios
closed
[Tabs] Add `image`, `selectedImage`, `selected` APIs to MDCTabBarViewItemView
[Tabs] type:Process
In order to support KVO changes to UITabBarItem's `selectedImage` API correctly, the item view of the tab bar will need to manage its imageview's image internally. When setting `UITabBarItem.image`, the `selectedImage` property is also set (though can be overridden). As a result, when trying to observe changes to `image` it's not possible to determine whether the selected image view's image should be updated. * You cannot check for a `nil` value on `UITabBarItem.selectedImage` because `selectedImage` returns a copy of `image`. * You cannot compare if `UITabBarItem.image == UITabBarItem.selectedImage` because the copy in `selectedImage` is not the same object as `image`. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/136004466](http://b/136004466)
1.0
[Tabs] Add `image`, `selectedImage`, `selected` APIs to MDCTabBarViewItemView - In order to support KVO changes to UITabBarItem's `selectedImage` API correctly, the item view of the tab bar will need to manage its imageview's image internally. When setting `UITabBarItem.image`, the `selectedImage` property is also set (though can be overridden). As a result, when trying to observe changes to `image` it's not possible to determine whether the selected image view's image should be updated. * You cannot check for a `nil` value on `UITabBarItem.selectedImage` because `selectedImage` returns a copy of `image`. * You cannot compare if `UITabBarItem.image == UITabBarItem.selectedImage` because the copy in `selectedImage` is not the same object as `image`. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/136004466](http://b/136004466)
process
add image selectedimage selected apis to mdctabbarviewitemview in order to support kvo changes to uitabbaritem s selectedimage api correctly the item view of the tab bar will need to manage its imageview s image internally when setting uitabbaritem image the selectedimage property is also set though can be overridden as a result when trying to observe changes to image it s not possible to determine whether the selected image view s image should be updated you cannot check for a nil value on uitabbaritem selectedimage because selectedimage returns a copy of image you cannot compare if uitabbaritem image uitabbaritem selectedimage because the copy in selectedimage is not the same object as image internal data associated internal bug
1
894
18,726,607,045
IssuesEvent
2021-11-03 16:53:28
planningcenter/developers
https://api.github.com/repos/planningcenter/developers
closed
Overdue Card Count at Workflow Endpoint
feature request People
**Related Product** Which product is this feature related to? People **Is your feature request related to a problem? Please describe.** I'm looking to build a dashboard view of roughly 20 different workflows for some members of our team. Ideally, this dashboard would show each workflow, along with the number of cards for each workflow where the card's "overdue" attribute is True. Currently, I believe I could hit the "/cards" endpoint and loop through each individual card for each workflow, but if it's easy enough to implement on your end, just having a "total_overdue_card_count" attribute, similar to the existing "total_ready_card_count" at the /workflows endpoint would be a big help in terms of efficiency. **Describe the solution you'd like** A "total_overdue_card_count" attribute added to workflow objects. **Describe alternatives you've considered** Alternatively, being able to filter or query the "/cards" endpoint for only overdue cards could be helpful as well. Honestly, that would be a nice feature to have even if the "total_overdue_card_count" attribute gets added, as then we could really efficiently link to the specific cards that are overdue without looping through all cards. **Additional context** Here is a screenshot of the dashboard as it stands so far. For each of the squares, I will be adding an overdue count underneath the workflow name. <img width="1428" alt="Screen Shot 2021-11-03 at 12 09 09 PM" src="https://user-images.githubusercontent.com/89470894/140100134-f04c6db6-ceac-4bef-911b-c1ee967ff08e.png"> ## I have.. - [x] Reviewed the documentation found at https://developer.planning.center/docs - [x] Searched for previous issues asking for this feature request - [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.) - [x] Reviewed my issue for completeness
1.0
Overdue Card Count at Workflow Endpoint - **Related Product** Which product is this feature related to? People **Is your feature request related to a problem? Please describe.** I'm looking to build a dashboard view of roughly 20 different workflows for some members of our team. Ideally, this dashboard would show each workflow, along with the number of cards for each workflow where the card's "overdue" attribute is True. Currently, I believe I could hit the "/cards" endpoint and loop through each individual card for each workflow, but if it's easy enough to implement on your end, just having a "total_overdue_card_count" attribute, similar to the existing "total_ready_card_count" at the /workflows endpoint would be a big help in terms of efficiency. **Describe the solution you'd like** A "total_overdue_card_count" attribute added to workflow objects. **Describe alternatives you've considered** Alternatively, being able to filter or query the "/cards" endpoint for only overdue cards could be helpful as well. Honestly, that would be a nice feature to have even if the "total_overdue_card_count" attribute gets added, as then we could really efficiently link to the specific cards that are overdue without looping through all cards. **Additional context** Here is a screenshot of the dashboard as it stands so far. For each of the squares, I will be adding an overdue count underneath the workflow name. <img width="1428" alt="Screen Shot 2021-11-03 at 12 09 09 PM" src="https://user-images.githubusercontent.com/89470894/140100134-f04c6db6-ceac-4bef-911b-c1ee967ff08e.png"> ## I have.. - [x] Reviewed the documentation found at https://developer.planning.center/docs - [x] Searched for previous issues asking for this feature request - [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.) - [x] Reviewed my issue for completeness
non_process
overdue card count at workflow endpoint related product which product is this feature related to people is your feature request related to a problem please describe i m looking to build a dashboard view of roughly different workflows for some members of our team ideally this dashboard would show each workflow along with the number of cards for each workflow where the card s overdue attribute is true currently i believe i could hit the cards endpoint and loop through each individual card for each workflow but if it s easy enough to implement on your end just having a total overdue card count attribute similar to the existing total ready card count at the workflows endpoint would be a big help in terms of efficiency describe the solution you d like a total overdue card count attribute added to workflow objects describe alternatives you ve considered alternatively being able to filter or query the cards endpoint for only overdue cards could be helpful as well honestly that would be a nice feature to have even if the total overdue card count attribute gets added as then we could really efficiently link to the specific cards that are overdue without looping through all cards additional context here is a screenshot of the dashboard as it stands so far for each of the squares i will be adding an overdue count underneath the workflow name img width alt screen shot at pm src i have reviewed the documentation found at searched for previous issues asking for this feature request removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness
0
429,342
30,037,625,940
IssuesEvent
2023-06-27 13:42:13
pyg-team/pytorch_geometric
https://api.github.com/repos/pyg-team/pytorch_geometric
closed
Broken link in the GNN Cheat sheet
documentation
### 📚 Describe the documentation issue In the cheatsheet [https://pytorch-geometric.readthedocs.io/en/latest/cheatsheet/gnn_cheatsheet.html](https://pytorch-geometric.readthedocs.io/en/latest/cheatsheet/gnn_cheatsheet.html). The paper link for SimpleConv points to a non-existant page. There should not be invalid links on the page. [SimpleConv](https://pytorch-geometric.readthedocs.io/en/latest/generated/torch_geometric.nn.conv.SimpleConv.html#torch_geometric.nn.conv.SimpleConv) ([Paper](https://pytorch-geometric.readthedocs.io/en/latest/cheatsheet/None)) ### Suggest a potential alternative/fix I see the code does ``` not torch_geometric.nn.conv.utils.processes_point_clouds(cls) %} * - :class:`~torch_geometric.nn.conv.{{ cls }}` (`Paper <{{ torch_geometric.nn.conv.utils.paper_link(cls) }}>`__) - {% if torch_geometric.nn.conv.utils.supports_sparse_tensor(cls) %}✓{% endif %} - {% if ``` If there is a valid appropriate paper - then we should point to that; if not then I suggest having a new document in this repository as the target for this link. The document should describe what SimpleConv does and why there is not paper for it; I assume because it is a very simple example.
1.0
Broken link in the GNN Cheat sheet - ### 📚 Describe the documentation issue In the cheatsheet [https://pytorch-geometric.readthedocs.io/en/latest/cheatsheet/gnn_cheatsheet.html](https://pytorch-geometric.readthedocs.io/en/latest/cheatsheet/gnn_cheatsheet.html). The paper link for SimpleConv points to a non-existant page. There should not be invalid links on the page. [SimpleConv](https://pytorch-geometric.readthedocs.io/en/latest/generated/torch_geometric.nn.conv.SimpleConv.html#torch_geometric.nn.conv.SimpleConv) ([Paper](https://pytorch-geometric.readthedocs.io/en/latest/cheatsheet/None)) ### Suggest a potential alternative/fix I see the code does ``` not torch_geometric.nn.conv.utils.processes_point_clouds(cls) %} * - :class:`~torch_geometric.nn.conv.{{ cls }}` (`Paper <{{ torch_geometric.nn.conv.utils.paper_link(cls) }}>`__) - {% if torch_geometric.nn.conv.utils.supports_sparse_tensor(cls) %}✓{% endif %} - {% if ``` If there is a valid appropriate paper - then we should point to that; if not then I suggest having a new document in this repository as the target for this link. The document should describe what SimpleConv does and why there is not paper for it; I assume because it is a very simple example.
non_process
broken link in the gnn cheat sheet 📚 describe the documentation issue in the cheatsheet the paper link for simpleconv points to a non existant page there should not be invalid links on the page suggest a potential alternative fix i see the code does not torch geometric nn conv utils processes point clouds cls class torch geometric nn conv cls paper if torch geometric nn conv utils supports sparse tensor cls ✓ endif if if there is a valid appropriate paper then we should point to that if not then i suggest having a new document in this repository as the target for this link the document should describe what simpleconv does and why there is not paper for it i assume because it is a very simple example
0
20,937
27,795,304,843
IssuesEvent
2023-03-17 12:05:18
q191201771/lal
https://api.github.com/repos/q191201771/lal
closed
记录的 FLV 的 FPS 从 60 更改为 62.5 - 严重问题 #Bug
#Bug *In process
流式传输 60 FPS 的示例视频文件会将其录制文件的 FPS 更改为 62.5 FPS OBS 或 WireCast 流媒体软件也存在同样的问题 我在下面附上了 lal 录音文件的输入和输出. 我认为这是一个严重的问题 ```shell ffmpeg -i /home/abr/frame_counter_4k_60fps.flv -c copy -f flv rtmp://127.0.0.1:1935/app/frame_counter_stream ``` ### Input file: ``` $ ffmpeg -i frame_counter_4k_60fps.flv Input #0, flv, from 'frame_counter_4k_60fps.flv': Metadata: major_brand : isom minor_version : 512 compatible_brands: isomiso2avc1mp41 encoder : Lavf59.16.100 Duration: 00:01:00.04, start: 0.019000, bitrate: 658 kb/s Stream #0:0: Video: h264, yuv420p(progressive), 3840x2160 [SAR 1:1 DAR 16:9], 640 kb/s, 60 fps, 60 tbr, 1k tbn Stream #0:1: Audio: aac, 48000 Hz, stereo, fltp, 2 kb/s ``` ### Recorded file: ``` $ ffmpeg -i lal_record/flv/frame_counter_stream-1657775570.flv Input #0, flv, from 'frame_counter_stream-1657775570.flv': Duration: 00:00:59.98, start: 0.000000, bitrate: 659 kb/s Stream #0:0: Data: none Stream #0:1: Video: h264, yuv420p(progressive), 3840x2160 [SAR 1:1 DAR 16:9], 62.50 fps, 60 tbr, 1k tbn Stream #0:2: Audio: aac, 48000 Hz, stereo, fltp ``` **English:** Streaming a sample video file at 60 FPS changes the FPS of the recorded file to 62.5 FPS. The same problem is with OBS or WireCast streaming software. I have attached the input and output from lal recording files below **This is a critical problem** [input_frame_counter_4k_60fps.zip](https://github.com/q191201771/lal/files/9108723/input_frame_counter_4k_60fps.zip) [output_lal_recorded_file.zip](https://github.com/q191201771/lal/files/9108815/output_lal_recorded_file.zip) [another_sample_file.mp4](https://twitch-event-engineering-public.s3.amazonaws.com/sync-footage/Sync-Footage-V1-H264.mp4) **#Bug**
1.0
记录的 FLV 的 FPS 从 60 更改为 62.5 - 严重问题 #Bug - 流式传输 60 FPS 的示例视频文件会将其录制文件的 FPS 更改为 62.5 FPS OBS 或 WireCast 流媒体软件也存在同样的问题 我在下面附上了 lal 录音文件的输入和输出. 我认为这是一个严重的问题 ```shell ffmpeg -i /home/abr/frame_counter_4k_60fps.flv -c copy -f flv rtmp://127.0.0.1:1935/app/frame_counter_stream ``` ### Input file: ``` $ ffmpeg -i frame_counter_4k_60fps.flv Input #0, flv, from 'frame_counter_4k_60fps.flv': Metadata: major_brand : isom minor_version : 512 compatible_brands: isomiso2avc1mp41 encoder : Lavf59.16.100 Duration: 00:01:00.04, start: 0.019000, bitrate: 658 kb/s Stream #0:0: Video: h264, yuv420p(progressive), 3840x2160 [SAR 1:1 DAR 16:9], 640 kb/s, 60 fps, 60 tbr, 1k tbn Stream #0:1: Audio: aac, 48000 Hz, stereo, fltp, 2 kb/s ``` ### Recorded file: ``` $ ffmpeg -i lal_record/flv/frame_counter_stream-1657775570.flv Input #0, flv, from 'frame_counter_stream-1657775570.flv': Duration: 00:00:59.98, start: 0.000000, bitrate: 659 kb/s Stream #0:0: Data: none Stream #0:1: Video: h264, yuv420p(progressive), 3840x2160 [SAR 1:1 DAR 16:9], 62.50 fps, 60 tbr, 1k tbn Stream #0:2: Audio: aac, 48000 Hz, stereo, fltp ``` **English:** Streaming a sample video file at 60 FPS changes the FPS of the recorded file to 62.5 FPS. The same problem is with OBS or WireCast streaming software. I have attached the input and output from lal recording files below **This is a critical problem** [input_frame_counter_4k_60fps.zip](https://github.com/q191201771/lal/files/9108723/input_frame_counter_4k_60fps.zip) [output_lal_recorded_file.zip](https://github.com/q191201771/lal/files/9108815/output_lal_recorded_file.zip) [another_sample_file.mp4](https://twitch-event-engineering-public.s3.amazonaws.com/sync-footage/Sync-Footage-V1-H264.mp4) **#Bug**
process
记录的 flv 的 fps 从 更改为 严重问题 bug 流式传输 fps 的示例视频文件会将其录制文件的 fps 更改为 fps obs 或 wirecast 流媒体软件也存在同样的问题 我在下面附上了 lal 录音文件的输入和输出 我认为这是一个严重的问题 shell ffmpeg i home abr frame counter flv c copy f flv rtmp app frame counter stream input file ffmpeg i frame counter flv input flv from frame counter flv metadata major brand isom minor version compatible brands encoder duration start bitrate kb s stream video progressive kb s fps tbr tbn stream audio aac hz stereo fltp kb s recorded file ffmpeg i lal record flv frame counter stream flv input flv from frame counter stream flv duration start bitrate kb s stream data none stream video progressive fps tbr tbn stream audio aac hz stereo fltp english streaming a sample video file at fps changes the fps of the recorded file to fps the same problem is with obs or wirecast streaming software i have attached the input and output from lal recording files below this is a critical problem bug
1
12,722
15,093,827,441
IssuesEvent
2021-02-07 02:47:14
Maximus5/ConEmu
https://api.github.com/repos/Maximus5/ConEmu
closed
Prefix all error messages with "ConEmu: "
processes
### Versions ConEmu build: latest OS version: irrelevant Used shell version: Far ### Problem description When something goes wrong, various error messages are displayed, e.g.: https://github.com/Maximus5/ConEmu/blob/ab9520df14f135abc07e0d8c6d2f01bfd38208e9/src/ConEmuCD/ConsoleMain.cpp#L2214 \- which is a good thing. However, there's no indication where these messages are coming from, which can be confusing for the users, who often aren't even aware that ConEmu is involved when something is executed ([example](https://forum.farmanager.com/viewtopic.php?p=164667#p164667)). Consider decorating such error messages with a recognizable prefix. This would help the users to pinpoint the root cause and direct their bug reports accordingly.
1.0
Prefix all error messages with "ConEmu: " - ### Versions ConEmu build: latest OS version: irrelevant Used shell version: Far ### Problem description When something goes wrong, various error messages are displayed, e.g.: https://github.com/Maximus5/ConEmu/blob/ab9520df14f135abc07e0d8c6d2f01bfd38208e9/src/ConEmuCD/ConsoleMain.cpp#L2214 \- which is a good thing. However, there's no indication where these messages are coming from, which can be confusing for the users, who often aren't even aware that ConEmu is involved when something is executed ([example](https://forum.farmanager.com/viewtopic.php?p=164667#p164667)). Consider decorating such error messages with a recognizable prefix. This would help the users to pinpoint the root cause and direct their bug reports accordingly.
process
prefix all error messages with conemu versions conemu build latest os version irrelevant used shell version far problem description when something goes wrong various error messages are displayed e g which is a good thing however there s no indication where these messages are coming from which can be confusing for the users who often aren t even aware that conemu is involved when something is executed consider decorating such error messages with a recognizable prefix this would help the users to pinpoint the root cause and direct their bug reports accordingly
1
494,736
14,264,448,456
IssuesEvent
2020-11-20 15:46:09
rism-ch/verovio
https://api.github.com/repos/rism-ch/verovio
opened
slur + @place="between" interaction
enhancement low priority
When dynamics (and probably other element types) are positioned with `@place="between"`, slurs will aggressively push the dynamics away from the between position: <img width="374" alt="Screen Shot 2020-11-20 at 7 39 30 AM" src="https://user-images.githubusercontent.com/3487289/99818957-b49cc000-2b03-11eb-9efa-5928ac28675a.png"> Probably due to the bounding box (or estimated bounding box) around the slur being too large. MEI test data: ```xml <?xml version="1.0" encoding="UTF-8"?> <?xml-model href="https://music-encoding.org/schema/4.0.0/mei-all.rng" type="application/xml" schematypens="http://relaxng.org/ns/structure/1.0"?> <?xml-model href="https://music-encoding.org/schema/4.0.0/mei-all.rng" type="application/xml" schematypens="http://purl.oclc.org/dsdl/schematron"?> <mei xmlns="http://www.music-encoding.org/ns/mei" meiversion="4.0.0"> <meiHead> <fileDesc> <titleStmt> <title /> </titleStmt> <pubStmt /> </fileDesc> <encodingDesc> <appInfo> <application isodate="2020-11-20T07:39:43" version="3.1.0-dev-94a4985-dirty"> <name>Verovio</name> <p>Transcoded from Humdrum</p> </application> </appInfo> </encodingDesc> <workList> <work> <title /> </work> </workList> </meiHead> <music> <body> <mdiv xml:id="mdiv-0000001044307350"> <score xml:id="score-0000000995153518"> <scoreDef xml:id="scoredef-0000001440342042"> <staffGrp xml:id="staffgrp-0000000014002306" symbol="brace" bar.thru="true"> <staffDef xml:id="staffdef-0000001154934699" n="1" lines="5"> <clef xml:id="clef-L3F2" shape="G" line="2" /> <meterSig xml:id="metersig-L4F2" count="3" unit="4" /> </staffDef> <staffDef xml:id="staffdef-0000000801964291" n="2" lines="5"> <clef xml:id="clef-L3F1" shape="F" line="4" /> <meterSig xml:id="metersig-L4F1" count="3" unit="4" /> </staffDef> </staffGrp> </scoreDef> <section xml:id="section-L1F1"> <measure xml:id="measure-L1"> <staff xml:id="staff-0000000497425680" n="1"> <layer xml:id="layer-L1F2N1" n="1"> <mRest xml:id="mrest-L6F2" visible="false" /> </layer> </staff> <staff xml:id="staff-0000001135985012" n="2"> <layer xml:id="layer-L1F1N1" n="1"> <note xml:id="note-L6F1" dur="4" oct="3" pname="c" accid.ges="n" /> <note xml:id="note-L7F1" dur="4" oct="3" pname="d" accid.ges="n" /> <note xml:id="note-L8F1" dur="4" oct="3" pname="e" accid.ges="n" /> </layer> </staff> <dynam xml:id="dynam-L6F3" place="between" staff="1 2" tstamp="1.000000">f</dynam> <dynam xml:id="dynam-L7F3" place="between" staff="1 2" tstamp="2.000000">p</dynam> <dynam xml:id="dynam-L8F3" place="between" staff="1 2" tstamp="3.000000">f</dynam> </measure> <measure xml:id="measure-L9"> <staff xml:id="staff-L9F2N1" n="1"> <layer xml:id="layer-L9F2N1" n="1"> <mRest xml:id="mrest-L10F2" visible="false" /> </layer> </staff> <staff xml:id="staff-L9F1N1" n="2"> <layer xml:id="layer-L9F1N1" n="1"> <note xml:id="note-L10F1" dur="4" oct="3" pname="c" accid.ges="n" /> <note xml:id="note-L11F1" dur="4" oct="3" pname="d" accid.ges="n" /> <note xml:id="note-L12F1" dur="4" oct="3" pname="e" accid.ges="n" /> </layer> </staff> <dynam xml:id="dynam-L10F3" place="between" staff="1 2" tstamp="1.000000">f</dynam> <dynam xml:id="dynam-L11F3" place="between" staff="1 2" tstamp="2.000000">p</dynam> <dynam xml:id="dynam-L12F3" place="between" staff="1 2" tstamp="3.000000">f</dynam> <slur xml:id="slur-L10F1-L12F1" staff="2" startid="#note-L10F1" endid="#note-L12F1" /> </measure> </section> </score> </mdiv> </body> </music> </mei> ```
1.0
slur + @place="between" interaction - When dynamics (and probably other element types) are positioned with `@place="between"`, slurs will aggressively push the dynamics away from the between position: <img width="374" alt="Screen Shot 2020-11-20 at 7 39 30 AM" src="https://user-images.githubusercontent.com/3487289/99818957-b49cc000-2b03-11eb-9efa-5928ac28675a.png"> Probably due to the bounding box (or estimated bounding box) around the slur being too large. MEI test data: ```xml <?xml version="1.0" encoding="UTF-8"?> <?xml-model href="https://music-encoding.org/schema/4.0.0/mei-all.rng" type="application/xml" schematypens="http://relaxng.org/ns/structure/1.0"?> <?xml-model href="https://music-encoding.org/schema/4.0.0/mei-all.rng" type="application/xml" schematypens="http://purl.oclc.org/dsdl/schematron"?> <mei xmlns="http://www.music-encoding.org/ns/mei" meiversion="4.0.0"> <meiHead> <fileDesc> <titleStmt> <title /> </titleStmt> <pubStmt /> </fileDesc> <encodingDesc> <appInfo> <application isodate="2020-11-20T07:39:43" version="3.1.0-dev-94a4985-dirty"> <name>Verovio</name> <p>Transcoded from Humdrum</p> </application> </appInfo> </encodingDesc> <workList> <work> <title /> </work> </workList> </meiHead> <music> <body> <mdiv xml:id="mdiv-0000001044307350"> <score xml:id="score-0000000995153518"> <scoreDef xml:id="scoredef-0000001440342042"> <staffGrp xml:id="staffgrp-0000000014002306" symbol="brace" bar.thru="true"> <staffDef xml:id="staffdef-0000001154934699" n="1" lines="5"> <clef xml:id="clef-L3F2" shape="G" line="2" /> <meterSig xml:id="metersig-L4F2" count="3" unit="4" /> </staffDef> <staffDef xml:id="staffdef-0000000801964291" n="2" lines="5"> <clef xml:id="clef-L3F1" shape="F" line="4" /> <meterSig xml:id="metersig-L4F1" count="3" unit="4" /> </staffDef> </staffGrp> </scoreDef> <section xml:id="section-L1F1"> <measure xml:id="measure-L1"> <staff xml:id="staff-0000000497425680" n="1"> <layer xml:id="layer-L1F2N1" n="1"> <mRest xml:id="mrest-L6F2" visible="false" /> </layer> </staff> <staff xml:id="staff-0000001135985012" n="2"> <layer xml:id="layer-L1F1N1" n="1"> <note xml:id="note-L6F1" dur="4" oct="3" pname="c" accid.ges="n" /> <note xml:id="note-L7F1" dur="4" oct="3" pname="d" accid.ges="n" /> <note xml:id="note-L8F1" dur="4" oct="3" pname="e" accid.ges="n" /> </layer> </staff> <dynam xml:id="dynam-L6F3" place="between" staff="1 2" tstamp="1.000000">f</dynam> <dynam xml:id="dynam-L7F3" place="between" staff="1 2" tstamp="2.000000">p</dynam> <dynam xml:id="dynam-L8F3" place="between" staff="1 2" tstamp="3.000000">f</dynam> </measure> <measure xml:id="measure-L9"> <staff xml:id="staff-L9F2N1" n="1"> <layer xml:id="layer-L9F2N1" n="1"> <mRest xml:id="mrest-L10F2" visible="false" /> </layer> </staff> <staff xml:id="staff-L9F1N1" n="2"> <layer xml:id="layer-L9F1N1" n="1"> <note xml:id="note-L10F1" dur="4" oct="3" pname="c" accid.ges="n" /> <note xml:id="note-L11F1" dur="4" oct="3" pname="d" accid.ges="n" /> <note xml:id="note-L12F1" dur="4" oct="3" pname="e" accid.ges="n" /> </layer> </staff> <dynam xml:id="dynam-L10F3" place="between" staff="1 2" tstamp="1.000000">f</dynam> <dynam xml:id="dynam-L11F3" place="between" staff="1 2" tstamp="2.000000">p</dynam> <dynam xml:id="dynam-L12F3" place="between" staff="1 2" tstamp="3.000000">f</dynam> <slur xml:id="slur-L10F1-L12F1" staff="2" startid="#note-L10F1" endid="#note-L12F1" /> </measure> </section> </score> </mdiv> </body> </music> </mei> ```
non_process
slur place between interaction when dynamics and probably other element types are positioned with place between slurs will aggressively push the dynamics away from the between position img width alt screen shot at am src probably due to the bounding box or estimated bounding box around the slur being too large mei test data xml xml model href type application xml schematypens xml model href type application xml schematypens verovio transcoded from humdrum f p f f p f
0
395,286
27,063,968,763
IssuesEvent
2023-02-13 22:15:28
dagster-io/dagster
https://api.github.com/repos/dagster-io/dagster
opened
Config schema should be visible in API docs
documentation
### What's the issue or suggestion? See: https://docs.dagster.io/_apidocs/jobs#dagster.job We used to have config schema rendering here. ### Additional information _No response_ ### Message from the maintainers Impacted by this issue? Give it a 👍! We factor engagement into prioritization.
1.0
Config schema should be visible in API docs - ### What's the issue or suggestion? See: https://docs.dagster.io/_apidocs/jobs#dagster.job We used to have config schema rendering here. ### Additional information _No response_ ### Message from the maintainers Impacted by this issue? Give it a 👍! We factor engagement into prioritization.
non_process
config schema should be visible in api docs what s the issue or suggestion see we used to have config schema rendering here additional information no response message from the maintainers impacted by this issue give it a 👍 we factor engagement into prioritization
0
442,471
30,837,390,545
IssuesEvent
2023-08-02 08:21:31
epinio/docs
https://api.github.com/repos/epinio/docs
opened
Epic: documentation review, restructure, validate and enhance
area/documentation
Documentation work plan, keeping track of docs issues. - Review - Restructure - Validate and enhance existing content
1.0
Epic: documentation review, restructure, validate and enhance - Documentation work plan, keeping track of docs issues. - Review - Restructure - Validate and enhance existing content
non_process
epic documentation review restructure validate and enhance documentation work plan keeping track of docs issues review restructure validate and enhance existing content
0
529,321
15,386,424,793
IssuesEvent
2021-03-03 08:11:57
davidsaulrodriguez/FitnessPlusPlus
https://api.github.com/repos/davidsaulrodriguez/FitnessPlusPlus
opened
🐞 - No HTML Routes In Place
backend bug urgent priority
**Describe the bug** The server starts just fine and handles static files, properly, as it should. There are no HTML routes setup yet. This is critical. Let's get the routes setup and working so that we can easily link back and fourth to the desired pages described throughout the app's navigation links. **To Reproduce** Steps to reproduce the behavior: 1. Go to 'http://localhost:3000' || 'http://localhost:4000' 2. Click on any navigable link 3. See **GET** request error. **Expected behavior** When a user clicks on a link, they expect to see the page that that link is attempting to navigate them to. **Desktop (please complete the following information):** - OS: KDE Neon 5.21.1 (Ubuntu LTS 20.04) - Browser: Mozilla Firefox - Version: 86.0+
1.0
🐞 - No HTML Routes In Place - **Describe the bug** The server starts just fine and handles static files, properly, as it should. There are no HTML routes setup yet. This is critical. Let's get the routes setup and working so that we can easily link back and fourth to the desired pages described throughout the app's navigation links. **To Reproduce** Steps to reproduce the behavior: 1. Go to 'http://localhost:3000' || 'http://localhost:4000' 2. Click on any navigable link 3. See **GET** request error. **Expected behavior** When a user clicks on a link, they expect to see the page that that link is attempting to navigate them to. **Desktop (please complete the following information):** - OS: KDE Neon 5.21.1 (Ubuntu LTS 20.04) - Browser: Mozilla Firefox - Version: 86.0+
non_process
🐞 no html routes in place describe the bug the server starts just fine and handles static files properly as it should there are no html routes setup yet this is critical let s get the routes setup and working so that we can easily link back and fourth to the desired pages described throughout the app s navigation links to reproduce steps to reproduce the behavior go to click on any navigable link see get request error expected behavior when a user clicks on a link they expect to see the page that that link is attempting to navigate them to desktop please complete the following information os kde neon ubuntu lts browser mozilla firefox version
0
200,653
7,010,205,809
IssuesEvent
2017-12-19 22:11:32
rconstanzo/karma
https://api.github.com/repos/rconstanzo/karma
closed
Creating a karma~ with no buffer should throw no error message
enhancement priority::low
Creating a [karma~] object with no associated buffer throws up an error of: `karma~: will not load properly without associated buffer~ name!` This should be in line with current [groove~] behavior which lets you make a bufferless [groove~] object with no errors.
1.0
Creating a karma~ with no buffer should throw no error message - Creating a [karma~] object with no associated buffer throws up an error of: `karma~: will not load properly without associated buffer~ name!` This should be in line with current [groove~] behavior which lets you make a bufferless [groove~] object with no errors.
non_process
creating a karma with no buffer should throw no error message creating a object with no associated buffer throws up an error of karma will not load properly without associated buffer name this should be in line with current behavior which lets you make a bufferless object with no errors
0
744,383
25,940,796,977
IssuesEvent
2022-12-16 18:10:56
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
Address logging v3.2.2 issues in samples
priority: p2 type: bug samples
Based on #8167, there has been an introduction with a new format of output that's messing with the expected output for the samples in `functions/v2/http_logging`. Few options that Daniel suggested: ``` The easiest solution for testing is probably to use the existing internal flag to disable instrumentation for the tests here: google.cloud.logging_v2._instrumentation_emitted = True. I think changing the tests is also acceptable if that ends up being easier though. ```
1.0
Address logging v3.2.2 issues in samples - Based on #8167, there has been an introduction with a new format of output that's messing with the expected output for the samples in `functions/v2/http_logging`. Few options that Daniel suggested: ``` The easiest solution for testing is probably to use the existing internal flag to disable instrumentation for the tests here: google.cloud.logging_v2._instrumentation_emitted = True. I think changing the tests is also acceptable if that ends up being easier though. ```
non_process
address logging issues in samples based on there has been an introduction with a new format of output that s messing with the expected output for the samples in functions http logging few options that daniel suggested the easiest solution for testing is probably to use the existing internal flag to disable instrumentation for the tests here google cloud logging instrumentation emitted true i think changing the tests is also acceptable if that ends up being easier though
0
168,106
6,362,140,642
IssuesEvent
2017-07-31 14:25:07
GoogleCloudPlatform/google-cloud-dotnet
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-dotnet
closed
Push Google.Cloud.DevTools.Common 1.0 or untag it
Priority: P0
(This is a "middle of the night" worry that I need to address first thing on Monday morning; it's a placeholder to avoid me forgetting.) We've got a release tag for Google.Cloud.DevTools.Common v1.0, but no corresponding nuget package. We should only be bumping version numbers as part of the release process - so the time between "tag created" and "package released" should be very small. There have been commits to the source code since the tag - we need to review those and see whether or not they *should* be in a release of the package. If not, and if everything else is ready, we can build and release 1.0. Otherwise, we should remove the 1.0 tag while we work out when we *do* want to release 1.0.
1.0
Push Google.Cloud.DevTools.Common 1.0 or untag it - (This is a "middle of the night" worry that I need to address first thing on Monday morning; it's a placeholder to avoid me forgetting.) We've got a release tag for Google.Cloud.DevTools.Common v1.0, but no corresponding nuget package. We should only be bumping version numbers as part of the release process - so the time between "tag created" and "package released" should be very small. There have been commits to the source code since the tag - we need to review those and see whether or not they *should* be in a release of the package. If not, and if everything else is ready, we can build and release 1.0. Otherwise, we should remove the 1.0 tag while we work out when we *do* want to release 1.0.
non_process
push google cloud devtools common or untag it this is a middle of the night worry that i need to address first thing on monday morning it s a placeholder to avoid me forgetting we ve got a release tag for google cloud devtools common but no corresponding nuget package we should only be bumping version numbers as part of the release process so the time between tag created and package released should be very small there have been commits to the source code since the tag we need to review those and see whether or not they should be in a release of the package if not and if everything else is ready we can build and release otherwise we should remove the tag while we work out when we do want to release
0
219,091
24,443,551,089
IssuesEvent
2022-10-06 16:06:36
MatBenfield/news
https://api.github.com/repos/MatBenfield/news
closed
[SecurityWeek] SCADA Systems Involved in Many Breaches Suffered by US Ports, Terminals
SecurityWeek Stale
**Law firm Jones Walker has published the results of a survey focusing on the cybersecurity preparedness of ports and terminals in the United States.** [read more](https://www.securityweek.com/scada-systems-involved-many-breaches-suffered-us-ports-terminals) <https://www.securityweek.com/scada-systems-involved-many-breaches-suffered-us-ports-terminals>
True
[SecurityWeek] SCADA Systems Involved in Many Breaches Suffered by US Ports, Terminals - **Law firm Jones Walker has published the results of a survey focusing on the cybersecurity preparedness of ports and terminals in the United States.** [read more](https://www.securityweek.com/scada-systems-involved-many-breaches-suffered-us-ports-terminals) <https://www.securityweek.com/scada-systems-involved-many-breaches-suffered-us-ports-terminals>
non_process
scada systems involved in many breaches suffered by us ports terminals law firm jones walker has published the results of a survey focusing on the cybersecurity preparedness of ports and terminals in the united states
0
14,617
17,760,703,030
IssuesEvent
2021-08-29 16:36:11
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Add maxBuffer to spawn options
question child_process
Hello, All the other child process functions(exec, execSync, spawnSync) have a `maxBuffer` options except `spawn`, why is that? can it be fixed? thank you in advance
1.0
Add maxBuffer to spawn options - Hello, All the other child process functions(exec, execSync, spawnSync) have a `maxBuffer` options except `spawn`, why is that? can it be fixed? thank you in advance
process
add maxbuffer to spawn options hello all the other child process functions exec execsync spawnsync have a maxbuffer options except spawn why is that can it be fixed thank you in advance
1
13,718
16,483,256,071
IssuesEvent
2021-05-24 14:29:48
paul-buerkner/brms
https://api.github.com/repos/paul-buerkner/brms
closed
Change default `cores` for `posterior_predict` on Windows
efficiency post-processing
On Windows, setting up new R processes is much more expensive than on Linux as the new processes need to pass data and setup. This makes running `posterior_predict` in parallel mostly very slow. I would suggest that the default `cores` argument should be `1` on Windows and `getOption("mc.cores", 1)` elsewhere, as I really need to have `mc.cores` high for fitting, but it really does not pay off for `posterior_predict`. In fact, the higher I set `mc.cores`, the slower I get predictions and the slowdown is _substantial_. Here's an example timing from a recent project of mine: ``` > system.time({print(pp_check(fit_auc1, "stat_grouped", stat = "sd", group = "type", nsamples = 4000, binwidth = 20, cores = 1))}) user system elapsed 2.23 0.07 2.39 Warning message: The following arguments were unrecognized and ignored: cores > system.time({print(pp_check(fit_auc1, "stat_grouped", stat = "sd", group = "type", nsamples = 4000, binwidth = 20, cores = 4))}) user system elapsed 2.15 0.11 13.78 Warning message: The following arguments were unrecognized and ignored: cores > system.time({print(pp_check(fit_auc1, "stat_grouped", stat = "sd", group = "type", nsamples = 4000, binwidth = 20, cores = 12))}) user system elapsed 2.35 0.36 35.88 Warning message: The following arguments were unrecognized and ignored: cores ``` Additionally, `pp_check` should not pass the `cores` argument to the `ppc_XXX` functions to avoid the warning... As usual, happy to file a pull request if you agree with the overall idea.
1.0
Change default `cores` for `posterior_predict` on Windows - On Windows, setting up new R processes is much more expensive than on Linux as the new processes need to pass data and setup. This makes running `posterior_predict` in parallel mostly very slow. I would suggest that the default `cores` argument should be `1` on Windows and `getOption("mc.cores", 1)` elsewhere, as I really need to have `mc.cores` high for fitting, but it really does not pay off for `posterior_predict`. In fact, the higher I set `mc.cores`, the slower I get predictions and the slowdown is _substantial_. Here's an example timing from a recent project of mine: ``` > system.time({print(pp_check(fit_auc1, "stat_grouped", stat = "sd", group = "type", nsamples = 4000, binwidth = 20, cores = 1))}) user system elapsed 2.23 0.07 2.39 Warning message: The following arguments were unrecognized and ignored: cores > system.time({print(pp_check(fit_auc1, "stat_grouped", stat = "sd", group = "type", nsamples = 4000, binwidth = 20, cores = 4))}) user system elapsed 2.15 0.11 13.78 Warning message: The following arguments were unrecognized and ignored: cores > system.time({print(pp_check(fit_auc1, "stat_grouped", stat = "sd", group = "type", nsamples = 4000, binwidth = 20, cores = 12))}) user system elapsed 2.35 0.36 35.88 Warning message: The following arguments were unrecognized and ignored: cores ``` Additionally, `pp_check` should not pass the `cores` argument to the `ppc_XXX` functions to avoid the warning... As usual, happy to file a pull request if you agree with the overall idea.
process
change default cores for posterior predict on windows on windows setting up new r processes is much more expensive than on linux as the new processes need to pass data and setup this makes running posterior predict in parallel mostly very slow i would suggest that the default cores argument should be on windows and getoption mc cores elsewhere as i really need to have mc cores high for fitting but it really does not pay off for posterior predict in fact the higher i set mc cores the slower i get predictions and the slowdown is substantial here s an example timing from a recent project of mine system time print pp check fit stat grouped stat sd group type nsamples binwidth cores user system elapsed warning message the following arguments were unrecognized and ignored cores system time print pp check fit stat grouped stat sd group type nsamples binwidth cores user system elapsed warning message the following arguments were unrecognized and ignored cores system time print pp check fit stat grouped stat sd group type nsamples binwidth cores user system elapsed warning message the following arguments were unrecognized and ignored cores additionally pp check should not pass the cores argument to the ppc xxx functions to avoid the warning as usual happy to file a pull request if you agree with the overall idea
1
21,730
30,242,939,579
IssuesEvent
2023-07-06 14:36:10
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Pty host lazy spawning doesn't work on remote
bug remote terminal-process
1. Open OSS 2. Open new test resolver window 3. Close non-test resolver window 4. Close all terminals 5. Exit application 6. Open OSS 7. Open process manager pty host should not be under remote agent: ![image](https://github.com/microsoft/vscode/assets/2193314/16d9b32b-6a7a-4ebf-aa93-c2b910c237d4)
1.0
Pty host lazy spawning doesn't work on remote - 1. Open OSS 2. Open new test resolver window 3. Close non-test resolver window 4. Close all terminals 5. Exit application 6. Open OSS 7. Open process manager pty host should not be under remote agent: ![image](https://github.com/microsoft/vscode/assets/2193314/16d9b32b-6a7a-4ebf-aa93-c2b910c237d4)
process
pty host lazy spawning doesn t work on remote open oss open new test resolver window close non test resolver window close all terminals exit application open oss open process manager pty host should not be under remote agent
1
2,329
5,142,555,371
IssuesEvent
2017-01-12 13:41:02
jimbrown75/Permit-Vision-Enhancements
https://api.github.com/repos/jimbrown75/Permit-Vision-Enhancements
opened
Planned Start and end dates
High Priority Must Fix Verified by PTW Process Lead
Planned start and end dates. (Now called Validity start end end) but we need to remove any logic that forces you to limit the validity to the maximum lifetime validity of the permit. If there was no logic then this could be the actually planned dates of the work itself instead of the validity of the permit. it would also allow work order dates to be transferred in and would help with planning.
1.0
Planned Start and end dates - Planned start and end dates. (Now called Validity start end end) but we need to remove any logic that forces you to limit the validity to the maximum lifetime validity of the permit. If there was no logic then this could be the actually planned dates of the work itself instead of the validity of the permit. it would also allow work order dates to be transferred in and would help with planning.
process
planned start and end dates planned start and end dates now called validity start end end but we need to remove any logic that forces you to limit the validity to the maximum lifetime validity of the permit if there was no logic then this could be the actually planned dates of the work itself instead of the validity of the permit it would also allow work order dates to be transferred in and would help with planning
1
331,513
28,966,103,839
IssuesEvent
2023-05-10 08:01:53
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
JCache TCK test org.jsr107.tck.event.CacheListenerTest got stuck
Team: Core Type: Test-Failure Source: Internal Module: ICache
_master_ Issue was introduced by commit https://github.com/hazelcast/hazelcast/commit/4c8057d2470b4a628e13793a1cce6a8a1f0c6268 Failed on JCache TCK tests: https://jenkins.hazelcast.com/view/Official%20Builds/job/JCache-1.1-TCK-master-client-OS/548/console JCache TCK tests are using https://github.com/hazelcast/jcache-tck/tree/v1.1.0/master/client/OS <details><summary>Standard output:</summary> ``` 04:22:19 Running org.jsr107.tck.event.CacheListenerTest 04:22:19 May 10, 2023 2:22:19 AM org.jsr107.tck.support.Server createServerSocket 04:22:19 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Removed cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM org.jsr107.tck.support.Server createServerSocket 04:22:19 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Removed cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM org.jsr107.tck.support.Server createServerSocket 04:22:19 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM org.jsr107.tck.support.Client <init> 04:22:19 INFO: Starting org.jsr107.tck.support.Client client connecting to server at address:/172.17.0.11 port:10011 04:22:20 May 10, 2023 2:22:20 AM org.jsr107.tck.event.CacheListenerTest testFilteredListener 04:22:20 INFO: Sooty 04:22:20 May 10, 2023 2:22:20 AM org.jsr107.tck.event.CacheListenerTest testFilteredListener 04:22:20 INFO: Fred 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.cache.impl.CacheService 04:22:20 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Removed cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:20 May 10, 2023 2:22:20 AM org.jsr107.tck.support.Server createServerSocket 04:22:20 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.cache.impl.CacheService 04:22:20 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.cache.impl.CacheService 04:22:20 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Removed cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.spi.impl.proxyservice.impl.operations.DistributedObjectDestroyOperation 04:22:20 SEVERE: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] CacheEntryListenerCount decremented to a value below zero! New value: -1 04:22:20 java.lang.AssertionError: CacheEntryListenerCount decremented to a value below zero! New value: -1 04:22:20 at com.hazelcast.cache.impl.CacheContext.decreaseCacheEntryListenerCount(CacheContext.java:67) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:131) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:76) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListenerInternal(EventServiceSegment.java:122) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListener(EventServiceSegment.java:99) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.removeRegistrations(EventServiceSegment.java:206) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.deregisterAllLocalListeners(EventServiceImpl.java:403) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deregisterAllListener(AbstractCacheService.java:679) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deleteCache(AbstractCacheService.java:366) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.destroyDistributedObject(AbstractCacheService.java:279) 04:22:20 at com.hazelcast.internal.services.RemoteService.destroyDistributedObject(RemoteService.java:68) 04:22:20 at com.hazelcast.spi.impl.proxyservice.impl.ProxyServiceImpl.destroyLocalDistributedObject(ProxyServiceImpl.java:193) 04:22:20 at com.hazelcast.spi.impl.proxyservice.impl.operations.DistributedObjectDestroyOperation.run(DistributedObjectDestroyOperation.java:45) 04:22:20 at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:299) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:219) 04:22:20 at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.run(OperationExecutorImpl.java:454) 04:22:20 at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.runOrExecute(OperationExecutorImpl.java:511) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvokeLocal(Invocation.java:618) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvoke(Invocation.java:597) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke0(Invocation.java:558) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke(Invocation.java:241) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.InvocationBuilderImpl.invoke(InvocationBuilderImpl.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMultiTargetMessageTask.processInternal(AbstractMultiTargetMessageTask.java:70) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractAsyncMessageTask.processMessage(AbstractAsyncMessageTask.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:169) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:132) 04:22:20 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 04:22:20 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 04:22:20 at java.lang.Thread.run(Thread.java:750) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:111) 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.client.impl.protocol.task.DestroyProxyMessageTask 04:22:20 WARNING: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Error while destroying a proxy. 04:22:20 java.lang.AssertionError: CacheEntryListenerCount decremented to a value below zero! New value: -1 04:22:20 at com.hazelcast.cache.impl.CacheContext.decreaseCacheEntryListenerCount(CacheContext.java:67) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:131) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:76) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListenerInternal(EventServiceSegment.java:122) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListener(EventServiceSegment.java:99) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.removeRegistrations(EventServiceSegment.java:206) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.deregisterAllLocalListeners(EventServiceImpl.java:403) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deregisterAllListener(AbstractCacheService.java:679) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deleteCache(AbstractCacheService.java:366) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.destroyDistributedObject(AbstractCacheService.java:279) 04:22:20 at com.hazelcast.internal.services.RemoteService.destroyDistributedObject(RemoteService.java:68) 04:22:20 at com.hazelcast.spi.impl.proxyservice.impl.ProxyServiceImpl.destroyLocalDistributedObject(ProxyServiceImpl.java:193) 04:22:20 at com.hazelcast.spi.impl.proxyservice.impl.operations.DistributedObjectDestroyOperation.run(DistributedObjectDestroyOperation.java:45) 04:22:20 at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:299) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:219) 04:22:20 at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.run(OperationExecutorImpl.java:454) 04:22:20 at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.runOrExecute(OperationExecutorImpl.java:511) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvokeLocal(Invocation.java:618) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvoke(Invocation.java:597) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke0(Invocation.java:558) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke(Invocation.java:241) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.InvocationBuilderImpl.invoke(InvocationBuilderImpl.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMultiTargetMessageTask.processInternal(AbstractMultiTargetMessageTask.java:70) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractAsyncMessageTask.processMessage(AbstractAsyncMessageTask.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:169) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:132) 04:22:20 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 04:22:20 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 04:22:20 at java.lang.Thread.run(Thread.java:750) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:111) 04:22:20 May 10, 2023 2:22:20 AM org.jsr107.tck.support.Server createServerSocket 04:22:20 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.cache.impl.CacheService 04:22:20 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.client.impl.spi.ClientListenerService 04:22:20 WARNING: _hzinstance_jcache_shared [dev] [5.3.0-SNAPSHOT] Deregistration of listener with ID e38f3700-efdb-4548-ba50-e2cabae28b5f has failed for address [172.17.0.11]:5701 04:22:20 java.lang.AssertionError: CacheEntryListenerCount decremented to a value below zero! New value: -1 04:22:20 at com.hazelcast.cache.impl.CacheContext.decreaseCacheEntryListenerCount(CacheContext.java:67) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:131) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:76) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListenerInternal(EventServiceSegment.java:122) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListener(EventServiceSegment.java:99) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.removeRegistration(EventServiceSegment.java:189) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.deregisterListenerAsync(EventServiceImpl.java:386) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deregisterListenerAsync(AbstractCacheService.java:644) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheRemoveEntryListenerMessageTask.deRegisterListener(CacheRemoveEntryListenerMessageTask.java:52) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractRemoveListenerMessageTask.processInternal(AbstractRemoveListenerMessageTask.java:42) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractAsyncMessageTask.processMessage(AbstractAsyncMessageTask.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:169) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:132) 04:22:20 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 04:22:20 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 04:22:20 at java.lang.Thread.run(Thread.java:750) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:111) 05:21:06 Build timed out (after 60 minutes). Marking the build as failed. ``` </details>
1.0
JCache TCK test org.jsr107.tck.event.CacheListenerTest got stuck - _master_ Issue was introduced by commit https://github.com/hazelcast/hazelcast/commit/4c8057d2470b4a628e13793a1cce6a8a1f0c6268 Failed on JCache TCK tests: https://jenkins.hazelcast.com/view/Official%20Builds/job/JCache-1.1-TCK-master-client-OS/548/console JCache TCK tests are using https://github.com/hazelcast/jcache-tck/tree/v1.1.0/master/client/OS <details><summary>Standard output:</summary> ``` 04:22:19 Running org.jsr107.tck.event.CacheListenerTest 04:22:19 May 10, 2023 2:22:19 AM org.jsr107.tck.support.Server createServerSocket 04:22:19 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Removed cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM org.jsr107.tck.support.Server createServerSocket 04:22:19 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Removed cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM org.jsr107.tck.support.Server createServerSocket 04:22:19 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:19 May 10, 2023 2:22:19 AM com.hazelcast.cache.impl.CacheService 04:22:19 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:19 May 10, 2023 2:22:19 AM org.jsr107.tck.support.Client <init> 04:22:19 INFO: Starting org.jsr107.tck.support.Client client connecting to server at address:/172.17.0.11 port:10011 04:22:20 May 10, 2023 2:22:20 AM org.jsr107.tck.event.CacheListenerTest testFilteredListener 04:22:20 INFO: Sooty 04:22:20 May 10, 2023 2:22:20 AM org.jsr107.tck.event.CacheListenerTest testFilteredListener 04:22:20 INFO: Fred 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.cache.impl.CacheService 04:22:20 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Removed cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:20 May 10, 2023 2:22:20 AM org.jsr107.tck.support.Server createServerSocket 04:22:20 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.cache.impl.CacheService 04:22:20 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.cache.impl.CacheService 04:22:20 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Removed cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.spi.impl.proxyservice.impl.operations.DistributedObjectDestroyOperation 04:22:20 SEVERE: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] CacheEntryListenerCount decremented to a value below zero! New value: -1 04:22:20 java.lang.AssertionError: CacheEntryListenerCount decremented to a value below zero! New value: -1 04:22:20 at com.hazelcast.cache.impl.CacheContext.decreaseCacheEntryListenerCount(CacheContext.java:67) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:131) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:76) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListenerInternal(EventServiceSegment.java:122) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListener(EventServiceSegment.java:99) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.removeRegistrations(EventServiceSegment.java:206) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.deregisterAllLocalListeners(EventServiceImpl.java:403) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deregisterAllListener(AbstractCacheService.java:679) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deleteCache(AbstractCacheService.java:366) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.destroyDistributedObject(AbstractCacheService.java:279) 04:22:20 at com.hazelcast.internal.services.RemoteService.destroyDistributedObject(RemoteService.java:68) 04:22:20 at com.hazelcast.spi.impl.proxyservice.impl.ProxyServiceImpl.destroyLocalDistributedObject(ProxyServiceImpl.java:193) 04:22:20 at com.hazelcast.spi.impl.proxyservice.impl.operations.DistributedObjectDestroyOperation.run(DistributedObjectDestroyOperation.java:45) 04:22:20 at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:299) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:219) 04:22:20 at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.run(OperationExecutorImpl.java:454) 04:22:20 at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.runOrExecute(OperationExecutorImpl.java:511) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvokeLocal(Invocation.java:618) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvoke(Invocation.java:597) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke0(Invocation.java:558) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke(Invocation.java:241) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.InvocationBuilderImpl.invoke(InvocationBuilderImpl.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMultiTargetMessageTask.processInternal(AbstractMultiTargetMessageTask.java:70) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractAsyncMessageTask.processMessage(AbstractAsyncMessageTask.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:169) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:132) 04:22:20 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 04:22:20 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 04:22:20 at java.lang.Thread.run(Thread.java:750) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:111) 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.client.impl.protocol.task.DestroyProxyMessageTask 04:22:20 WARNING: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Error while destroying a proxy. 04:22:20 java.lang.AssertionError: CacheEntryListenerCount decremented to a value below zero! New value: -1 04:22:20 at com.hazelcast.cache.impl.CacheContext.decreaseCacheEntryListenerCount(CacheContext.java:67) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:131) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:76) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListenerInternal(EventServiceSegment.java:122) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListener(EventServiceSegment.java:99) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.removeRegistrations(EventServiceSegment.java:206) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.deregisterAllLocalListeners(EventServiceImpl.java:403) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deregisterAllListener(AbstractCacheService.java:679) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deleteCache(AbstractCacheService.java:366) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.destroyDistributedObject(AbstractCacheService.java:279) 04:22:20 at com.hazelcast.internal.services.RemoteService.destroyDistributedObject(RemoteService.java:68) 04:22:20 at com.hazelcast.spi.impl.proxyservice.impl.ProxyServiceImpl.destroyLocalDistributedObject(ProxyServiceImpl.java:193) 04:22:20 at com.hazelcast.spi.impl.proxyservice.impl.operations.DistributedObjectDestroyOperation.run(DistributedObjectDestroyOperation.java:45) 04:22:20 at com.hazelcast.spi.impl.operationservice.Operation.call(Operation.java:192) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(OperationRunnerImpl.java:299) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:270) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:219) 04:22:20 at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.run(OperationExecutorImpl.java:454) 04:22:20 at com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl.runOrExecute(OperationExecutorImpl.java:511) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvokeLocal(Invocation.java:618) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.doInvoke(Invocation.java:597) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke0(Invocation.java:558) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.Invocation.invoke(Invocation.java:241) 04:22:20 at com.hazelcast.spi.impl.operationservice.impl.InvocationBuilderImpl.invoke(InvocationBuilderImpl.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMultiTargetMessageTask.processInternal(AbstractMultiTargetMessageTask.java:70) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractAsyncMessageTask.processMessage(AbstractAsyncMessageTask.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:169) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:132) 04:22:20 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 04:22:20 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 04:22:20 at java.lang.Thread.run(Thread.java:750) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:111) 04:22:20 May 10, 2023 2:22:20 AM org.jsr107.tck.support.Server createServerSocket 04:22:20 INFO: Starting org.jsr107.tck.event.CacheEntryListenerServer server at address:/172.17.0.11 port:10011 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.cache.impl.CacheService 04:22:20 INFO: [172.17.0.11]:5701 [dev] [5.3.0-SNAPSHOT] Added cache config: CacheConfig{name='org.jsr107.tck.event.CacheListenerTest', managerPrefix='/hz/', inMemoryFormat=BINARY, backupCount=1, hotRestart=HotRestartConfig{enabled=false, fsync=false}, dataPersistenceConfig=DataPersistenceConfig{enabled=false, fsync=false}, wanReplicationRef=null, merkleTreeConfig=MerkleTreeConfig{enabled=null, depth=10}} 04:22:20 May 10, 2023 2:22:20 AM com.hazelcast.client.impl.spi.ClientListenerService 04:22:20 WARNING: _hzinstance_jcache_shared [dev] [5.3.0-SNAPSHOT] Deregistration of listener with ID e38f3700-efdb-4548-ba50-e2cabae28b5f has failed for address [172.17.0.11]:5701 04:22:20 java.lang.AssertionError: CacheEntryListenerCount decremented to a value below zero! New value: -1 04:22:20 at com.hazelcast.cache.impl.CacheContext.decreaseCacheEntryListenerCount(CacheContext.java:67) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:131) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheAddEntryListenerMessageTask$CacheEntryListener.onDeregister(CacheAddEntryListenerMessageTask.java:76) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListenerInternal(EventServiceSegment.java:122) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.pingNotifiableEventListener(EventServiceSegment.java:99) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceSegment.removeRegistration(EventServiceSegment.java:189) 04:22:20 at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.deregisterListenerAsync(EventServiceImpl.java:386) 04:22:20 at com.hazelcast.cache.impl.AbstractCacheService.deregisterListenerAsync(AbstractCacheService.java:644) 04:22:20 at com.hazelcast.client.impl.protocol.task.cache.CacheRemoveEntryListenerMessageTask.deRegisterListener(CacheRemoveEntryListenerMessageTask.java:52) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractRemoveListenerMessageTask.processInternal(AbstractRemoveListenerMessageTask.java:42) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractAsyncMessageTask.processMessage(AbstractAsyncMessageTask.java:71) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.initializeAndProcessMessage(AbstractMessageTask.java:169) 04:22:20 at com.hazelcast.client.impl.protocol.task.AbstractMessageTask.run(AbstractMessageTask.java:132) 04:22:20 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 04:22:20 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 04:22:20 at java.lang.Thread.run(Thread.java:750) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) 04:22:20 at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:111) 05:21:06 Build timed out (after 60 minutes). Marking the build as failed. ``` </details>
non_process
jcache tck test org tck event cachelistenertest got stuck master issue was introduced by commit failed on jcache tck tests jcache tck tests are using standard output running org tck event cachelistenertest may am org tck support server createserversocket info starting org tck event cacheentrylistenerserver server at address port may am com hazelcast cache impl cacheservice info added cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am com hazelcast cache impl cacheservice info removed cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am org tck support server createserversocket info starting org tck event cacheentrylistenerserver server at address port may am com hazelcast cache impl cacheservice info added cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am com hazelcast cache impl cacheservice info removed cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am org tck support server createserversocket info starting org tck event cacheentrylistenerserver server at address port may am com hazelcast cache impl cacheservice info added cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am org tck support client info starting org tck support client client connecting to server at address port may am org tck event cachelistenertest testfilteredlistener info sooty may am org tck event cachelistenertest testfilteredlistener info fred may am com hazelcast cache impl cacheservice info removed cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am org tck support server createserversocket info starting org tck event cacheentrylistenerserver server at address port may am com hazelcast cache impl cacheservice info added cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am com hazelcast cache impl cacheservice info removed cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am com hazelcast spi impl proxyservice impl operations distributedobjectdestroyoperation severe cacheentrylistenercount decremented to a value below zero new value java lang assertionerror cacheentrylistenercount decremented to a value below zero new value at com hazelcast cache impl cachecontext decreasecacheentrylistenercount cachecontext java at com hazelcast client impl protocol task cache cacheaddentrylistenermessagetask cacheentrylistener onderegister cacheaddentrylistenermessagetask java at com hazelcast client impl protocol task cache cacheaddentrylistenermessagetask cacheentrylistener onderegister cacheaddentrylistenermessagetask java at com hazelcast spi impl eventservice impl eventservicesegment pingnotifiableeventlistenerinternal eventservicesegment java at com hazelcast spi impl eventservice impl eventservicesegment pingnotifiableeventlistener eventservicesegment java at com hazelcast spi impl eventservice impl eventservicesegment removeregistrations eventservicesegment java at com hazelcast spi impl eventservice impl eventserviceimpl deregisteralllocallisteners eventserviceimpl java at com hazelcast cache impl abstractcacheservice deregisteralllistener abstractcacheservice java at com hazelcast cache impl abstractcacheservice deletecache abstractcacheservice java at com hazelcast cache impl abstractcacheservice destroydistributedobject abstractcacheservice java at com hazelcast internal services remoteservice destroydistributedobject remoteservice java at com hazelcast spi impl proxyservice impl proxyserviceimpl destroylocaldistributedobject proxyserviceimpl java at com hazelcast spi impl proxyservice impl operations distributedobjectdestroyoperation run distributedobjectdestroyoperation java at com hazelcast spi impl operationservice operation call operation java at com hazelcast spi impl operationservice impl operationrunnerimpl call operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationexecutor impl operationexecutorimpl run operationexecutorimpl java at com hazelcast spi impl operationexecutor impl operationexecutorimpl runorexecute operationexecutorimpl java at com hazelcast spi impl operationservice impl invocation doinvokelocal invocation java at com hazelcast spi impl operationservice impl invocation doinvoke invocation java at com hazelcast spi impl operationservice impl invocation invocation java at com hazelcast spi impl operationservice impl invocation invoke invocation java at com hazelcast spi impl operationservice impl invocationbuilderimpl invoke invocationbuilderimpl java at com hazelcast client impl protocol task abstractmultitargetmessagetask processinternal abstractmultitargetmessagetask java at com hazelcast client impl protocol task abstractasyncmessagetask processmessage abstractasyncmessagetask java at com hazelcast client impl protocol task abstractmessagetask initializeandprocessmessage abstractmessagetask java at com hazelcast client impl protocol task abstractmessagetask run abstractmessagetask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast internal util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java may am com hazelcast client impl protocol task destroyproxymessagetask warning error while destroying a proxy java lang assertionerror cacheentrylistenercount decremented to a value below zero new value at com hazelcast cache impl cachecontext decreasecacheentrylistenercount cachecontext java at com hazelcast client impl protocol task cache cacheaddentrylistenermessagetask cacheentrylistener onderegister cacheaddentrylistenermessagetask java at com hazelcast client impl protocol task cache cacheaddentrylistenermessagetask cacheentrylistener onderegister cacheaddentrylistenermessagetask java at com hazelcast spi impl eventservice impl eventservicesegment pingnotifiableeventlistenerinternal eventservicesegment java at com hazelcast spi impl eventservice impl eventservicesegment pingnotifiableeventlistener eventservicesegment java at com hazelcast spi impl eventservice impl eventservicesegment removeregistrations eventservicesegment java at com hazelcast spi impl eventservice impl eventserviceimpl deregisteralllocallisteners eventserviceimpl java at com hazelcast cache impl abstractcacheservice deregisteralllistener abstractcacheservice java at com hazelcast cache impl abstractcacheservice deletecache abstractcacheservice java at com hazelcast cache impl abstractcacheservice destroydistributedobject abstractcacheservice java at com hazelcast internal services remoteservice destroydistributedobject remoteservice java at com hazelcast spi impl proxyservice impl proxyserviceimpl destroylocaldistributedobject proxyserviceimpl java at com hazelcast spi impl proxyservice impl operations distributedobjectdestroyoperation run distributedobjectdestroyoperation java at com hazelcast spi impl operationservice operation call operation java at com hazelcast spi impl operationservice impl operationrunnerimpl call operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationexecutor impl operationexecutorimpl run operationexecutorimpl java at com hazelcast spi impl operationexecutor impl operationexecutorimpl runorexecute operationexecutorimpl java at com hazelcast spi impl operationservice impl invocation doinvokelocal invocation java at com hazelcast spi impl operationservice impl invocation doinvoke invocation java at com hazelcast spi impl operationservice impl invocation invocation java at com hazelcast spi impl operationservice impl invocation invoke invocation java at com hazelcast spi impl operationservice impl invocationbuilderimpl invoke invocationbuilderimpl java at com hazelcast client impl protocol task abstractmultitargetmessagetask processinternal abstractmultitargetmessagetask java at com hazelcast client impl protocol task abstractasyncmessagetask processmessage abstractasyncmessagetask java at com hazelcast client impl protocol task abstractmessagetask initializeandprocessmessage abstractmessagetask java at com hazelcast client impl protocol task abstractmessagetask run abstractmessagetask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast internal util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java may am org tck support server createserversocket info starting org tck event cacheentrylistenerserver server at address port may am com hazelcast cache impl cacheservice info added cache config cacheconfig name org tck event cachelistenertest managerprefix hz inmemoryformat binary backupcount hotrestart hotrestartconfig enabled false fsync false datapersistenceconfig datapersistenceconfig enabled false fsync false wanreplicationref null merkletreeconfig merkletreeconfig enabled null depth may am com hazelcast client impl spi clientlistenerservice warning hzinstance jcache shared deregistration of listener with id efdb has failed for address java lang assertionerror cacheentrylistenercount decremented to a value below zero new value at com hazelcast cache impl cachecontext decreasecacheentrylistenercount cachecontext java at com hazelcast client impl protocol task cache cacheaddentrylistenermessagetask cacheentrylistener onderegister cacheaddentrylistenermessagetask java at com hazelcast client impl protocol task cache cacheaddentrylistenermessagetask cacheentrylistener onderegister cacheaddentrylistenermessagetask java at com hazelcast spi impl eventservice impl eventservicesegment pingnotifiableeventlistenerinternal eventservicesegment java at com hazelcast spi impl eventservice impl eventservicesegment pingnotifiableeventlistener eventservicesegment java at com hazelcast spi impl eventservice impl eventservicesegment removeregistration eventservicesegment java at com hazelcast spi impl eventservice impl eventserviceimpl deregisterlistenerasync eventserviceimpl java at com hazelcast cache impl abstractcacheservice deregisterlistenerasync abstractcacheservice java at com hazelcast client impl protocol task cache cacheremoveentrylistenermessagetask deregisterlistener cacheremoveentrylistenermessagetask java at com hazelcast client impl protocol task abstractremovelistenermessagetask processinternal abstractremovelistenermessagetask java at com hazelcast client impl protocol task abstractasyncmessagetask processmessage abstractasyncmessagetask java at com hazelcast client impl protocol task abstractmessagetask initializeandprocessmessage abstractmessagetask java at com hazelcast client impl protocol task abstractmessagetask run abstractmessagetask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast internal util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java build timed out after minutes marking the build as failed
0
443,322
12,792,548,904
IssuesEvent
2020-07-02 01:40:39
lardemua/atom
https://api.github.com/repos/lardemua/atom
closed
Merge general hand eye with calibrate
High Priority enhancement help wanted
Should do this to advance in the general approach. Defined with @eupedrosa that the best way is to try to put in calibrate the stuff from hand eye, i.e. to use the calibrate as base.
1.0
Merge general hand eye with calibrate - Should do this to advance in the general approach. Defined with @eupedrosa that the best way is to try to put in calibrate the stuff from hand eye, i.e. to use the calibrate as base.
non_process
merge general hand eye with calibrate should do this to advance in the general approach defined with eupedrosa that the best way is to try to put in calibrate the stuff from hand eye i e to use the calibrate as base
0
6,588
9,663,631,477
IssuesEvent
2019-05-21 01:36:52
NottingHack/hms2
https://api.github.com/repos/NottingHack/hms2
closed
Ability to update the account linked to a user
Process
Joint accounts Need a way to do this after details have been approved and send out some email as needed posible automated way? user can request to have there account linked?
1.0
Ability to update the account linked to a user - Joint accounts Need a way to do this after details have been approved and send out some email as needed posible automated way? user can request to have there account linked?
process
ability to update the account linked to a user joint accounts need a way to do this after details have been approved and send out some email as needed posible automated way user can request to have there account linked
1
16,733
21,898,164,271
IssuesEvent
2022-05-20 10:41:54
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
closed
Grok patterns break processing
processing to-verify
<!--- Provide a general summary of the issue in the Title above --> I am using a 3 node graylog cluster, with a 3 node ES cluster. All 6 nodes are working properly and are healthy indexing between 5 and 10G messages per day. I am trying to process logs from amavisd-new (spamassassin/av interface for mailers). I am adding a new grok patterns extractor on my syslog input (~125 msg/s). The new extractor only matches 1 or 2 messages per second: The extractor: ``` Condition Will only attempt to run if the message includes the string spam-tag Configuration grok_pattern: %{SYSLOGHOST}%{SYSLOGPROG}: %{AMAVIS_SA_RESULT} named_captures_only: ``` The Grok patterns: ``` AMAVIS_SA_RESULT \(%{DATA:amavis_thread}\) spam-tag, <(?:%{DATA:amavis_sender})?> -> <(?:%{EMAILADDRESS:amavis_recipient})?>(%{AMAVIS_RECIPIENT_LIST})?, %{AMAVIS_YES_NO:amavis_sa_spamtag}, score=(%{DATA:amavis_sa_score}|x) tagged_above=%{NOTSPACE} required=%{NOTSPACE} (%{WORD} )?tests=(\[(%{AMAVIS_SA_TEST_LIST})?\]) EMAILADDRESS %{EMAILLOCALPART}@%{EMAILDOMAINPART} EMAILDOMAINPART \b(?:[0-9A-Za-z][0-9A-Za-z-]{0,62})(?:\.(?:[0-9A-Za-z][0-9A-Za-z-]{0,62}))*(\.?|\b) EMAILLOCALPART [a-zA-Z][a-zA-Z0-9_.+-=:]+ AMAVIS_RECIPIENT_LIST (,<%{EMAILADDRESS:amavis_recipient}>)+ AMAVIS_SA_TEST_LIST %{WORD:amavis_sa_test}=%{DATA}(, %{WORD:amavis_sa_test}=%{DATA})* AMAVIS_YES_NO (?:Yes|No) ``` Message example: ``` (13498-03) spam-tag, <ChloéAncaruo@1015thehawk.com> -> <xxxxx@xxxxx.com>, Yes, score=11.078 tagged_above=-1000 required=5 tests=[CRM114_UNSURE=0.1, FROM_EXCESS_BASE64=0.105, FSL_BULK_SIG=0.001, HTML_IMAGE_ONLY_12=1.629, HTML_IMAGE_RATIO_08=0.001, HTML_MESSAGE=0.001, HTML_SHORT_LINK_IMG_2=0.259, MIME_HTML_ONLY=1.105, MR_NOT_ATTRIBUTED_IP=0.2, NO_RDNS=0.5, RAZOR2_CF_RANGE_51_100=0.1, RAZOR2_CF_RANGE_E8_91_100=0.1, RAZOR2_CHECK=0.1, RDNS_NONE=1.274, SERGIO_SUBJECT_PORN014=3.544, T_REMOTE_IMAGE=0.01, UN_PHISHING_PW=0.1, URIBL_ABUSE_SURBL=1.948, URIBL_BLOCKED=0.001] autolearn=disabled ``` ## Expected Behavior graylog should process messages as usual, spiting the messages into named fields. ## Current Behavior The process buffer on all 3 node fills up and the output buffer stays empty. Input buffer then starts to fill up too and nothing gets processed anymore. The problem occurs with graylog 3.0 and graylog 3.1. Extractor's stats: ``` Total time 95th percentile: 9μs 98th percentile: 16μs 99th percentile: 228μs Standard deviation: 121μs Mean: 10μs Minimum: 0μs Maximum: 4,941μs Condition time 95th percentile: 2μs 98th percentile: 2μs 99th percentile: 2μs Standard deviation: 0μs Mean: 1μs Minimum: 0μs Maximum: 17μs Execution time 95th percentile: 420μs 98th percentile: 468μs 99th percentile: 468μs Standard deviation: 77μs Mean: 296μs Minimum: 135μs Maximum: 10,828μs Converter time 95th percentile: 0μs 98th percentile: 0μs 99th percentile: 0μs Standard deviation: 0μs Mean: 0μs Minimum: 0μs Maximum: 68μs ``` **Removing the extractor does not fix the problem**. After removing the extractor, if I restart graylog-server on a node, it will then process normally the disk journal (up to 3000msg/s of ouput) and everything goes back to normal. If i leave the extractor and reboot the server, the output buffer fills up and 65536 messages get processed and the next 64535 messages are stuck in the process buffer. There are no errors in the logs... * Graylog Version: 3.0 and 3.1 * Elasticsearch Version: 6.5.4 * MongoDB Version: 4.0.12 * Operating System: debian stretch 9.9 * Browser version: Not revelant
1.0
Grok patterns break processing - <!--- Provide a general summary of the issue in the Title above --> I am using a 3 node graylog cluster, with a 3 node ES cluster. All 6 nodes are working properly and are healthy indexing between 5 and 10G messages per day. I am trying to process logs from amavisd-new (spamassassin/av interface for mailers). I am adding a new grok patterns extractor on my syslog input (~125 msg/s). The new extractor only matches 1 or 2 messages per second: The extractor: ``` Condition Will only attempt to run if the message includes the string spam-tag Configuration grok_pattern: %{SYSLOGHOST}%{SYSLOGPROG}: %{AMAVIS_SA_RESULT} named_captures_only: ``` The Grok patterns: ``` AMAVIS_SA_RESULT \(%{DATA:amavis_thread}\) spam-tag, <(?:%{DATA:amavis_sender})?> -> <(?:%{EMAILADDRESS:amavis_recipient})?>(%{AMAVIS_RECIPIENT_LIST})?, %{AMAVIS_YES_NO:amavis_sa_spamtag}, score=(%{DATA:amavis_sa_score}|x) tagged_above=%{NOTSPACE} required=%{NOTSPACE} (%{WORD} )?tests=(\[(%{AMAVIS_SA_TEST_LIST})?\]) EMAILADDRESS %{EMAILLOCALPART}@%{EMAILDOMAINPART} EMAILDOMAINPART \b(?:[0-9A-Za-z][0-9A-Za-z-]{0,62})(?:\.(?:[0-9A-Za-z][0-9A-Za-z-]{0,62}))*(\.?|\b) EMAILLOCALPART [a-zA-Z][a-zA-Z0-9_.+-=:]+ AMAVIS_RECIPIENT_LIST (,<%{EMAILADDRESS:amavis_recipient}>)+ AMAVIS_SA_TEST_LIST %{WORD:amavis_sa_test}=%{DATA}(, %{WORD:amavis_sa_test}=%{DATA})* AMAVIS_YES_NO (?:Yes|No) ``` Message example: ``` (13498-03) spam-tag, <ChloéAncaruo@1015thehawk.com> -> <xxxxx@xxxxx.com>, Yes, score=11.078 tagged_above=-1000 required=5 tests=[CRM114_UNSURE=0.1, FROM_EXCESS_BASE64=0.105, FSL_BULK_SIG=0.001, HTML_IMAGE_ONLY_12=1.629, HTML_IMAGE_RATIO_08=0.001, HTML_MESSAGE=0.001, HTML_SHORT_LINK_IMG_2=0.259, MIME_HTML_ONLY=1.105, MR_NOT_ATTRIBUTED_IP=0.2, NO_RDNS=0.5, RAZOR2_CF_RANGE_51_100=0.1, RAZOR2_CF_RANGE_E8_91_100=0.1, RAZOR2_CHECK=0.1, RDNS_NONE=1.274, SERGIO_SUBJECT_PORN014=3.544, T_REMOTE_IMAGE=0.01, UN_PHISHING_PW=0.1, URIBL_ABUSE_SURBL=1.948, URIBL_BLOCKED=0.001] autolearn=disabled ``` ## Expected Behavior graylog should process messages as usual, spiting the messages into named fields. ## Current Behavior The process buffer on all 3 node fills up and the output buffer stays empty. Input buffer then starts to fill up too and nothing gets processed anymore. The problem occurs with graylog 3.0 and graylog 3.1. Extractor's stats: ``` Total time 95th percentile: 9μs 98th percentile: 16μs 99th percentile: 228μs Standard deviation: 121μs Mean: 10μs Minimum: 0μs Maximum: 4,941μs Condition time 95th percentile: 2μs 98th percentile: 2μs 99th percentile: 2μs Standard deviation: 0μs Mean: 1μs Minimum: 0μs Maximum: 17μs Execution time 95th percentile: 420μs 98th percentile: 468μs 99th percentile: 468μs Standard deviation: 77μs Mean: 296μs Minimum: 135μs Maximum: 10,828μs Converter time 95th percentile: 0μs 98th percentile: 0μs 99th percentile: 0μs Standard deviation: 0μs Mean: 0μs Minimum: 0μs Maximum: 68μs ``` **Removing the extractor does not fix the problem**. After removing the extractor, if I restart graylog-server on a node, it will then process normally the disk journal (up to 3000msg/s of ouput) and everything goes back to normal. If i leave the extractor and reboot the server, the output buffer fills up and 65536 messages get processed and the next 64535 messages are stuck in the process buffer. There are no errors in the logs... * Graylog Version: 3.0 and 3.1 * Elasticsearch Version: 6.5.4 * MongoDB Version: 4.0.12 * Operating System: debian stretch 9.9 * Browser version: Not revelant
process
grok patterns break processing i am using a node graylog cluster with a node es cluster all nodes are working properly and are healthy indexing between and messages per day i am trying to process logs from amavisd new spamassassin av interface for mailers i am adding a new grok patterns extractor on my syslog input msg s the new extractor only matches or messages per second the extractor condition will only attempt to run if the message includes the string spam tag configuration grok pattern sysloghost syslogprog amavis sa result named captures only the grok patterns amavis sa result data amavis thread spam tag amavis recipient list amavis yes no amavis sa spamtag score data amavis sa score x tagged above notspace required notspace word tests emailaddress emaillocalpart emaildomainpart emaildomainpart b b emaillocalpart amavis recipient list amavis sa test list word amavis sa test data word amavis sa test data amavis yes no yes no message example spam tag yes score tagged above required tests autolearn disabled expected behavior graylog should process messages as usual spiting the messages into named fields current behavior the process buffer on all node fills up and the output buffer stays empty input buffer then starts to fill up too and nothing gets processed anymore the problem occurs with graylog and graylog extractor s stats total time percentile percentile percentile standard deviation mean minimum maximum condition time percentile percentile percentile standard deviation mean minimum maximum execution time percentile percentile percentile standard deviation mean minimum maximum converter time percentile percentile percentile standard deviation mean minimum maximum removing the extractor does not fix the problem after removing the extractor if i restart graylog server on a node it will then process normally the disk journal up to s of ouput and everything goes back to normal if i leave the extractor and reboot the server the output buffer fills up and messages get processed and the next messages are stuck in the process buffer there are no errors in the logs graylog version and elasticsearch version mongodb version operating system debian stretch browser version not revelant
1
67,452
12,962,600,571
IssuesEvent
2020-07-20 17:24:50
phetsims/energy-skate-park
https://api.github.com/repos/phetsims/energy-skate-park
closed
ReferenceHeightLine dragPositionProperty
dev:code-review
This section was flagged for review. ```js // a Vector2Property for DragListener so we can update the NumberProperty ReferenceHeightProperty with dragging // REVIEW: Shouldn't this be a DynamicProperty? It seems odd to have another independent properly related to the // REVIEW: reference position const dragPositionProperty = new Vector2Property( new Vector2( 0, referenceHeightProperty.value ) ); this.addInputListener( new DragListener( { transform: modelViewTransform, positionProperty: dragPositionProperty, tandem: tandem.createTandem( 'dragListener' ) } ) ); dragPositionProperty.link( dragPosition => { referenceHeightProperty.set( referenceHeightProperty.range.constrainValue( dragPositionProperty.get().y ) ); } ); ``` This was necessary because DragListener requires a Vector2Property, but the model uses a NumberProperty for the reference height. I am not sure yet how DynamicProperty would be better here.
1.0
ReferenceHeightLine dragPositionProperty - This section was flagged for review. ```js // a Vector2Property for DragListener so we can update the NumberProperty ReferenceHeightProperty with dragging // REVIEW: Shouldn't this be a DynamicProperty? It seems odd to have another independent properly related to the // REVIEW: reference position const dragPositionProperty = new Vector2Property( new Vector2( 0, referenceHeightProperty.value ) ); this.addInputListener( new DragListener( { transform: modelViewTransform, positionProperty: dragPositionProperty, tandem: tandem.createTandem( 'dragListener' ) } ) ); dragPositionProperty.link( dragPosition => { referenceHeightProperty.set( referenceHeightProperty.range.constrainValue( dragPositionProperty.get().y ) ); } ); ``` This was necessary because DragListener requires a Vector2Property, but the model uses a NumberProperty for the reference height. I am not sure yet how DynamicProperty would be better here.
non_process
referenceheightline dragpositionproperty this section was flagged for review js a for draglistener so we can update the numberproperty referenceheightproperty with dragging review shouldn t this be a dynamicproperty it seems odd to have another independent properly related to the review reference position const dragpositionproperty new new referenceheightproperty value this addinputlistener new draglistener transform modelviewtransform positionproperty dragpositionproperty tandem tandem createtandem draglistener dragpositionproperty link dragposition referenceheightproperty set referenceheightproperty range constrainvalue dragpositionproperty get y this was necessary because draglistener requires a but the model uses a numberproperty for the reference height i am not sure yet how dynamicproperty would be better here
0
3,160
6,217,762,000
IssuesEvent
2017-07-08 18:02:39
TeamProjectNotch/ProjectNotch
https://api.github.com/repos/TeamProjectNotch/ProjectNotch
closed
Introduce that git branching model
development process
Basically [this](http://nvie.com/posts/a-successful-git-branching-model/). `master` for stable versions/releases. `develop` for active development. Feature branches for, well, features.
1.0
Introduce that git branching model - Basically [this](http://nvie.com/posts/a-successful-git-branching-model/). `master` for stable versions/releases. `develop` for active development. Feature branches for, well, features.
process
introduce that git branching model basically master for stable versions releases develop for active development feature branches for well features
1
8,754
11,873,948,544
IssuesEvent
2020-03-26 18:08:32
googleapis/google-auth-library-nodejs
https://api.github.com/repos/googleapis/google-auth-library-nodejs
opened
the authentication library relies on gaxios setting application/json header
type: process
There is logic in our authentication flow that relies on the upstream gaxios library overwriting the content-type header with `application/json`. We should dig into why this is necessary, and explicitly set this header in the library.
1.0
the authentication library relies on gaxios setting application/json header - There is logic in our authentication flow that relies on the upstream gaxios library overwriting the content-type header with `application/json`. We should dig into why this is necessary, and explicitly set this header in the library.
process
the authentication library relies on gaxios setting application json header there is logic in our authentication flow that relies on the upstream gaxios library overwriting the content type header with application json we should dig into why this is necessary and explicitly set this header in the library
1
21,733
30,246,695,443
IssuesEvent
2023-07-06 17:02:45
w3c/vc-json-schema
https://api.github.com/repos/w3c/vc-json-schema
closed
Being able to normatively reference JSON Schema
planning process
Our largest hurdle for this work item is being able to normatively reference JSON Schema. JSON Schema has broad usage and industry adoption but isn't a standard in the eyes of the W3C, yet. The [latest draft of JSON Schema](https://datatracker.ietf.org/doc/draft-bhutton-json-schema/01/) is being standardized in the IETF, though this make take some time, too much time for us to hit our REC goal. We can non-normatively refer to older versions of the spec. But I believe we need to normatively refer to at least one version of JSON Schema (@msporny @OR13 correct me if I'm wrong). I believe our first step should be reaching out to contacts at the W3C and seeing if we can gain an exception for normatively referencing JSON Schema. I'm not certain how this process works. @iherman can you please advise sensible next steps?
1.0
Being able to normatively reference JSON Schema - Our largest hurdle for this work item is being able to normatively reference JSON Schema. JSON Schema has broad usage and industry adoption but isn't a standard in the eyes of the W3C, yet. The [latest draft of JSON Schema](https://datatracker.ietf.org/doc/draft-bhutton-json-schema/01/) is being standardized in the IETF, though this make take some time, too much time for us to hit our REC goal. We can non-normatively refer to older versions of the spec. But I believe we need to normatively refer to at least one version of JSON Schema (@msporny @OR13 correct me if I'm wrong). I believe our first step should be reaching out to contacts at the W3C and seeing if we can gain an exception for normatively referencing JSON Schema. I'm not certain how this process works. @iherman can you please advise sensible next steps?
process
being able to normatively reference json schema our largest hurdle for this work item is being able to normatively reference json schema json schema has broad usage and industry adoption but isn t a standard in the eyes of the yet the is being standardized in the ietf though this make take some time too much time for us to hit our rec goal we can non normatively refer to older versions of the spec but i believe we need to normatively refer to at least one version of json schema msporny correct me if i m wrong i believe our first step should be reaching out to contacts at the and seeing if we can gain an exception for normatively referencing json schema i m not certain how this process works iherman can you please advise sensible next steps
1
384,697
26,599,236,619
IssuesEvent
2023-01-23 14:41:09
siemens/ix
https://api.github.com/repos/siemens/ix
closed
Overlay example does not work propertly
type: bug component: documentation
### What happened? Close button does not work. Some example content should be included. ### What type of frontend frameware are you seeing the problem on? JavaScript ### Which version of iX do you use? 1.1.1 ### Code to produce this issue. _No response_
1.0
Overlay example does not work propertly - ### What happened? Close button does not work. Some example content should be included. ### What type of frontend frameware are you seeing the problem on? JavaScript ### Which version of iX do you use? 1.1.1 ### Code to produce this issue. _No response_
non_process
overlay example does not work propertly what happened close button does not work some example content should be included what type of frontend frameware are you seeing the problem on javascript which version of ix do you use code to produce this issue no response
0
596,632
18,108,420,658
IssuesEvent
2021-09-22 22:18:45
DSpace/dspace-angular
https://api.github.com/repos/DSpace/dspace-angular
closed
Cannot easily tell similar named EPeople or Groups apart (usability issue)
bug usability high priority e/6
**Describe the bug** In testing other PRs, I've had to recently add some new EPeople & assign them to groups, I've found that the current UI screens rely too much on UUID and Name, making it very difficult to tell similar named EPeople apart. Examples below... **To Reproduce** **Issue 1: Difficult to tell apart two EPeople of same name:** 1. Create two EPerson users with the same name 2. Now, edit a group on the Groups page (http://localhost:4000/admin/access-control/groups) 3. **Usability Issue:** Search for an EPerson with that name...you will not be able to tell the users apart, e.g. in this screenshot there are two "Tim Donohue" Epeople, but no other information provided besides UUID: ![two_epeople](https://user-images.githubusercontent.com/483997/112052383-10626e80-8b21-11eb-9575-42d0b7d2071b.png) 4. **Bug:** Even worse, clicking on one of the names brings me to the list of **Groups** instead of to that EPerson. 5. **EXPECTED BEHAVIOR:** We should list an EPerson's email address next to their name (in same column or separate column) in any browse/search results, as an email address is a simple way of distinguishing two people. **Issue 2: Difficult to tell apart two Groups of a similar name:** 1. Create two Collection Admin groups 2. Now, go to the Groups page (http://localhost:4000/admin/access-control/groups) and try to find the Collection Admin group of a specific Collection. 3. **Usability Issue 1:** You will be unable to tell the Groups apart unless you click on each Collection Admin group one by one. 4. **Usability Issue 2:** Now, click Edit on one of the Collection Admin groups. It's difficult to tell which Collection it refers to, unless you read the detailed note at the top of the page. 4. **EXPECTED BEHAVIOR 1:** Any Role-based Group should list the name of the Community or Collection next to it (either in the Name column or in a new column) in any browse/search results 5. **EXPECTED BEHAVIOR 2:** When editing a Role-based Group, we should show a Read-only field at the top which lists the Collection or Community name which the Group references.
1.0
Cannot easily tell similar named EPeople or Groups apart (usability issue) - **Describe the bug** In testing other PRs, I've had to recently add some new EPeople & assign them to groups, I've found that the current UI screens rely too much on UUID and Name, making it very difficult to tell similar named EPeople apart. Examples below... **To Reproduce** **Issue 1: Difficult to tell apart two EPeople of same name:** 1. Create two EPerson users with the same name 2. Now, edit a group on the Groups page (http://localhost:4000/admin/access-control/groups) 3. **Usability Issue:** Search for an EPerson with that name...you will not be able to tell the users apart, e.g. in this screenshot there are two "Tim Donohue" Epeople, but no other information provided besides UUID: ![two_epeople](https://user-images.githubusercontent.com/483997/112052383-10626e80-8b21-11eb-9575-42d0b7d2071b.png) 4. **Bug:** Even worse, clicking on one of the names brings me to the list of **Groups** instead of to that EPerson. 5. **EXPECTED BEHAVIOR:** We should list an EPerson's email address next to their name (in same column or separate column) in any browse/search results, as an email address is a simple way of distinguishing two people. **Issue 2: Difficult to tell apart two Groups of a similar name:** 1. Create two Collection Admin groups 2. Now, go to the Groups page (http://localhost:4000/admin/access-control/groups) and try to find the Collection Admin group of a specific Collection. 3. **Usability Issue 1:** You will be unable to tell the Groups apart unless you click on each Collection Admin group one by one. 4. **Usability Issue 2:** Now, click Edit on one of the Collection Admin groups. It's difficult to tell which Collection it refers to, unless you read the detailed note at the top of the page. 4. **EXPECTED BEHAVIOR 1:** Any Role-based Group should list the name of the Community or Collection next to it (either in the Name column or in a new column) in any browse/search results 5. **EXPECTED BEHAVIOR 2:** When editing a Role-based Group, we should show a Read-only field at the top which lists the Collection or Community name which the Group references.
non_process
cannot easily tell similar named epeople or groups apart usability issue describe the bug in testing other prs i ve had to recently add some new epeople assign them to groups i ve found that the current ui screens rely too much on uuid and name making it very difficult to tell similar named epeople apart examples below to reproduce issue difficult to tell apart two epeople of same name create two eperson users with the same name now edit a group on the groups page usability issue search for an eperson with that name you will not be able to tell the users apart e g in this screenshot there are two tim donohue epeople but no other information provided besides uuid bug even worse clicking on one of the names brings me to the list of groups instead of to that eperson expected behavior we should list an eperson s email address next to their name in same column or separate column in any browse search results as an email address is a simple way of distinguishing two people issue difficult to tell apart two groups of a similar name create two collection admin groups now go to the groups page and try to find the collection admin group of a specific collection usability issue you will be unable to tell the groups apart unless you click on each collection admin group one by one usability issue now click edit on one of the collection admin groups it s difficult to tell which collection it refers to unless you read the detailed note at the top of the page expected behavior any role based group should list the name of the community or collection next to it either in the name column or in a new column in any browse search results expected behavior when editing a role based group we should show a read only field at the top which lists the collection or community name which the group references
0
8,411
11,577,364,291
IssuesEvent
2020-02-21 13:59:04
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
opened
`AddServiceBusTopicMessagePump` should require subscription name, not prefix
area:message-processing integration:service-bus
`AddServiceBusTopicMessagePump` should require subscription name since you'd always want to specify the subscription to process. If the app runs multiple instances, they have to still process the same subscription regardless of the job. The only exception are things like the Key Vault invalidation job which requires a prefix. For those, I would introduce `AddServiceBusTopicMessagePumpWithPrefix` or so. This is a blocker for v0.1 imo as current approach is not usable.
1.0
`AddServiceBusTopicMessagePump` should require subscription name, not prefix - `AddServiceBusTopicMessagePump` should require subscription name since you'd always want to specify the subscription to process. If the app runs multiple instances, they have to still process the same subscription regardless of the job. The only exception are things like the Key Vault invalidation job which requires a prefix. For those, I would introduce `AddServiceBusTopicMessagePumpWithPrefix` or so. This is a blocker for v0.1 imo as current approach is not usable.
process
addservicebustopicmessagepump should require subscription name not prefix addservicebustopicmessagepump should require subscription name since you d always want to specify the subscription to process if the app runs multiple instances they have to still process the same subscription regardless of the job the only exception are things like the key vault invalidation job which requires a prefix for those i would introduce addservicebustopicmessagepumpwithprefix or so this is a blocker for imo as current approach is not usable
1
4,251
7,188,084,965
IssuesEvent
2018-02-02 08:49:24
nerdalize/nerd
https://api.github.com/repos/nerdalize/nerd
closed
Add flex volume to minikube
Dev Process K8S Integration P0
## Expected Behavior While running `./make.sh test`, it should install flex volumes over ssh to our local cluster. ## Actual Behavior Not added yet.
1.0
Add flex volume to minikube - ## Expected Behavior While running `./make.sh test`, it should install flex volumes over ssh to our local cluster. ## Actual Behavior Not added yet.
process
add flex volume to minikube expected behavior while running make sh test it should install flex volumes over ssh to our local cluster actual behavior not added yet
1
9,664
12,652,470,336
IssuesEvent
2020-06-17 03:41:34
googleapis/java-dataproc
https://api.github.com/repos/googleapis/java-dataproc
closed
Promote to GA
api: dataproc type: process
Package name: **google-cloud-dataproc** Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] At least one integration/smoke test is defined and passing - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [x] Manual code has been reviewed by a DPE responsible for samples - [x] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
Promote to GA - Package name: **google-cloud-dataproc** Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] At least one integration/smoke test is defined and passing - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [x] Manual code has been reviewed by a DPE responsible for samples - [x] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
promote to ga package name google cloud dataproc current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga at least one integration smoke test is defined and passing package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
211,713
7,203,824,871
IssuesEvent
2018-02-06 10:29:14
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
vk.com - site is not usable
browser-firefox-mobile-tablet priority-normal
<!-- @browser: Firefox Mobile (Tablet) 60.0 --> <!-- @ua_header: Mozilla/5.0 (Android 5.1.1; Tablet; rv:60.0) Gecko/60.0 Firefox/60.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://vk.com/ **Browser / Version**: Firefox Mobile (Tablet) 60.0 **Operating System**: Android 5.1.1 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: pasting video file in private message. video doesnt attach, only link **Steps to Reproduce**: _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
vk.com - site is not usable - <!-- @browser: Firefox Mobile (Tablet) 60.0 --> <!-- @ua_header: Mozilla/5.0 (Android 5.1.1; Tablet; rv:60.0) Gecko/60.0 Firefox/60.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://vk.com/ **Browser / Version**: Firefox Mobile (Tablet) 60.0 **Operating System**: Android 5.1.1 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: pasting video file in private message. video doesnt attach, only link **Steps to Reproduce**: _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
vk com site is not usable url browser version firefox mobile tablet operating system android tested another browser no problem type site is not usable description pasting video file in private message video doesnt attach only link steps to reproduce from with ❤️
0
9,132
12,202,787,612
IssuesEvent
2020-04-30 09:30:54
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Update https://github.com/prisma/binary-tester
kind/improvement process/candidate topic: binaries topic: tests
A lot of tests are failing right now, this could be a quick fix said @steebchen 😉 From what I see it needs a small update. See latest run: https://buildkite.com/prisma/binary-testing/builds/38#ee634a35-076e-427d-8668-eedf52848189
1.0
Update https://github.com/prisma/binary-tester - A lot of tests are failing right now, this could be a quick fix said @steebchen 😉 From what I see it needs a small update. See latest run: https://buildkite.com/prisma/binary-testing/builds/38#ee634a35-076e-427d-8668-eedf52848189
process
update a lot of tests are failing right now this could be a quick fix said steebchen 😉 from what i see it needs a small update see latest run
1
9,234
12,262,865,460
IssuesEvent
2020-05-06 23:14:21
googleapis/gaxios
https://api.github.com/repos/googleapis/gaxios
closed
enable publication with wombat-dressing-room
needs more info type: process
currently we don't have publication setup through kokoro, perhaps this could be a candidate for a wombat-dressing-room probot app.
1.0
enable publication with wombat-dressing-room - currently we don't have publication setup through kokoro, perhaps this could be a candidate for a wombat-dressing-room probot app.
process
enable publication with wombat dressing room currently we don t have publication setup through kokoro perhaps this could be a candidate for a wombat dressing room probot app
1
1,019
3,479,666,002
IssuesEvent
2015-12-28 22:06:06
t3kt/vjzual2
https://api.github.com/repos/t3kt/vjzual2
closed
zoom effect module
enhancement video processing
make it a separate module than the blur module (see #249). use the radial blur component from the TD palette.
1.0
zoom effect module - make it a separate module than the blur module (see #249). use the radial blur component from the TD palette.
process
zoom effect module make it a separate module than the blur module see use the radial blur component from the td palette
1
20,247
26,863,878,015
IssuesEvent
2023-02-03 21:11:22
GoogleCloudPlatform/emblem
https://api.github.com/repos/GoogleCloudPlatform/emblem
closed
Decision: document extension model
type: process priority: p3 type: docs
We should document the "extension" model used by `installation-tests` folder. (Thanks @rogerthatdev for highlighting this in #414.)
1.0
Decision: document extension model - We should document the "extension" model used by `installation-tests` folder. (Thanks @rogerthatdev for highlighting this in #414.)
process
decision document extension model we should document the extension model used by installation tests folder thanks rogerthatdev for highlighting this in
1
29,011
11,706,145,964
IssuesEvent
2020-03-07 20:12:57
vlaship/graphql-resolver
https://api.github.com/repos/vlaship/graphql-resolver
opened
CVE-2020-8840 (High) detected in jackson-databind-2.9.9.3.jar
security vulnerability
## CVE-2020-8840 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/graphql-resolver/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar</p> <p> Dependency Hierarchy: - graphql-spring-boot-starter-5.0.2.jar (Root Library) - graphql-spring-boot-autoconfigure-5.0.2.jar - graphql-java-servlet-6.1.2.jar - :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vlaship/graphql-resolver/commit/69b2d878133579d23ef9f8bc407028c32bfc4a47">69b2d878133579d23ef9f8bc407028c32bfc4a47</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.0.0 through 2.9.10.2 lacks certain xbean-reflect/JNDI blocking, as demonstrated by org.apache.xbean.propertyeditor.JndiConverter. <p>Publish Date: 2020-02-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8840>CVE-2020-8840</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2620">https://github.com/FasterXML/jackson-databind/issues/2620</a></p> <p>Release Date: 2020-02-10</p> <p>Fix Resolution: 2.8.11.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8840 (High) detected in jackson-databind-2.9.9.3.jar - ## CVE-2020-8840 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/graphql-resolver/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar</p> <p> Dependency Hierarchy: - graphql-spring-boot-starter-5.0.2.jar (Root Library) - graphql-spring-boot-autoconfigure-5.0.2.jar - graphql-java-servlet-6.1.2.jar - :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vlaship/graphql-resolver/commit/69b2d878133579d23ef9f8bc407028c32bfc4a47">69b2d878133579d23ef9f8bc407028c32bfc4a47</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.0.0 through 2.9.10.2 lacks certain xbean-reflect/JNDI blocking, as demonstrated by org.apache.xbean.propertyeditor.JndiConverter. <p>Publish Date: 2020-02-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8840>CVE-2020-8840</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2620">https://github.com/FasterXML/jackson-databind/issues/2620</a></p> <p>Release Date: 2020-02-10</p> <p>Fix Resolution: 2.8.11.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm graphql resolver build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy graphql spring boot starter jar root library graphql spring boot autoconfigure jar graphql java servlet jar x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind through lacks certain xbean reflect jndi blocking as demonstrated by org apache xbean propertyeditor jndiconverter publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
16,856
22,137,514,258
IssuesEvent
2022-06-03 01:34:22
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Bump release to 0.1.0-rc1
enhancement P1 process
### Problem release branch 0.1 needs to be created ### Solution Bump release to 0.1.0-rc1 ### Alternatives _No response_
1.0
Bump release to 0.1.0-rc1 - ### Problem release branch 0.1 needs to be created ### Solution Bump release to 0.1.0-rc1 ### Alternatives _No response_
process
bump release to problem release branch needs to be created solution bump release to alternatives no response
1
306,518
23,163,709,813
IssuesEvent
2022-07-29 21:01:00
taskless/taskless
https://api.github.com/repos/taskless/taskless
closed
Feature: fixed the position of the sidebar on the docs
documentation
After taking a tour at [taskless/docs](https://taskless.io/docs/packages/express), I noticed that when you scroll, the sidebar scroll as well. I think it should be great if the position is fixed, just to facilitate the navigation and jumps between sections.
1.0
Feature: fixed the position of the sidebar on the docs - After taking a tour at [taskless/docs](https://taskless.io/docs/packages/express), I noticed that when you scroll, the sidebar scroll as well. I think it should be great if the position is fixed, just to facilitate the navigation and jumps between sections.
non_process
feature fixed the position of the sidebar on the docs after taking a tour at i noticed that when you scroll the sidebar scroll as well i think it should be great if the position is fixed just to facilitate the navigation and jumps between sections
0
81,170
7,769,974,829
IssuesEvent
2018-06-04 07:10:52
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
[Failing Test] : gke-ingress-1.11 job failing to checkout Boskos project 1.11-release-blocking
kind/bug milestone/needs-attention priority/critical-urgent priority/failing-test sig/network sig/testing status/approved-for-milestone
## Failing Job * [sig-release-1.11-blocking#gke-ingress-1.11](https://k8s-testgrid.appspot.com/sig-release-1.11-blocking#gke-ingress-1.11) ## Error https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-cos-k8sbeta-ingress/2836 Something went wrong: failed to prepare test environment: --provider=gke boskos failed to acquire project: resource not found /kind bug /priority failing-test /priority critical-urgent /sig testing /milestone v1.11 @kubernetes/sig-testing-bugs cc @jberkus @tpepper /assign @BenTheElder to investigate
2.0
[Failing Test] : gke-ingress-1.11 job failing to checkout Boskos project 1.11-release-blocking - ## Failing Job * [sig-release-1.11-blocking#gke-ingress-1.11](https://k8s-testgrid.appspot.com/sig-release-1.11-blocking#gke-ingress-1.11) ## Error https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-cos-k8sbeta-ingress/2836 Something went wrong: failed to prepare test environment: --provider=gke boskos failed to acquire project: resource not found /kind bug /priority failing-test /priority critical-urgent /sig testing /milestone v1.11 @kubernetes/sig-testing-bugs cc @jberkus @tpepper /assign @BenTheElder to investigate
non_process
gke ingress job failing to checkout boskos project release blocking failing job error something went wrong failed to prepare test environment provider gke boskos failed to acquire project resource not found kind bug priority failing test priority critical urgent sig testing milestone kubernetes sig testing bugs cc jberkus tpepper assign bentheelder to investigate
0
101,764
11,256,063,391
IssuesEvent
2020-01-12 13:53:15
avajs/ava
https://api.github.com/repos/avajs/ava
closed
Recipe for using AVA with Selenium
good for beginner help wanted scope:documentation
It's a common need. We should have a [recipe](https://github.com/avajs/ava/tree/master/docs/recipes) to make it easy for users to get started. I think [Webdriver.IO](http://webdriver.io) would be the best choice.
1.0
Recipe for using AVA with Selenium - It's a common need. We should have a [recipe](https://github.com/avajs/ava/tree/master/docs/recipes) to make it easy for users to get started. I think [Webdriver.IO](http://webdriver.io) would be the best choice.
non_process
recipe for using ava with selenium it s a common need we should have a to make it easy for users to get started i think would be the best choice
0
6,915
10,064,284,222
IssuesEvent
2019-07-23 08:19:15
RohitDhankar/DigitalCognition
https://api.github.com/repos/RohitDhankar/DigitalCognition
opened
Provide an App within DC - for accessing NASDAQ - SEC EDGAR Data , similar to OpenEDGAR with enhanced functionality
Natural Language Processing Python_Utility_Methods
Build upon functionality from OpenEDGAR , focus not only on data ETL and analytics , but also provide Viz using Holoviews and Bokeh . https://arxiv.org/pdf/1806.04973.pdf --
1.0
Provide an App within DC - for accessing NASDAQ - SEC EDGAR Data , similar to OpenEDGAR with enhanced functionality - Build upon functionality from OpenEDGAR , focus not only on data ETL and analytics , but also provide Viz using Holoviews and Bokeh . https://arxiv.org/pdf/1806.04973.pdf --
process
provide an app within dc for accessing nasdaq sec edgar data similar to openedgar with enhanced functionality build upon functionality from openedgar focus not only on data etl and analytics but also provide viz using holoviews and bokeh
1
19,892
26,340,023,854
IssuesEvent
2023-01-10 16:58:04
digitalmethodsinitiative/4cat
https://api.github.com/repos/digitalmethodsinitiative/4cat
closed
Merge datasets
enhancement (mostly) back-end processors
Could be a processor that takes the URL of another dataset as a parameter; and then a new dataset is created containing the data from both the 'source' and 'target' datasets
1.0
Merge datasets - Could be a processor that takes the URL of another dataset as a parameter; and then a new dataset is created containing the data from both the 'source' and 'target' datasets
process
merge datasets could be a processor that takes the url of another dataset as a parameter and then a new dataset is created containing the data from both the source and target datasets
1
11,549
14,433,284,904
IssuesEvent
2020-12-07 04:25:21
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Instructions for stop/start VMs by a list incorrect
automation/svc cxp doc-enhancement process-automation/subsvc triaged
In the section for starting and stopping VMs from using a list https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-config#target-the-start-and-stop-action-by-vm-list it says to use the WHATIF parameter as true - however, this does not make the VMs actually stop or start, for that, the parameter needs to be false. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9aafa67b-4919-6773-d55c-c30e4e3ba703 * Version Independent ID: 51bec2e2-1984-d605-2c6e-6286e40262b8 * Content: [Configure Azure Automation Start/Stop VMs during off-hours](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-config) * Content Source: [articles/automation/automation-solution-vm-management-config.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management-config.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Instructions for stop/start VMs by a list incorrect - In the section for starting and stopping VMs from using a list https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-config#target-the-start-and-stop-action-by-vm-list it says to use the WHATIF parameter as true - however, this does not make the VMs actually stop or start, for that, the parameter needs to be false. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9aafa67b-4919-6773-d55c-c30e4e3ba703 * Version Independent ID: 51bec2e2-1984-d605-2c6e-6286e40262b8 * Content: [Configure Azure Automation Start/Stop VMs during off-hours](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-config) * Content Source: [articles/automation/automation-solution-vm-management-config.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management-config.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
instructions for stop start vms by a list incorrect in the section for starting and stopping vms from using a list it says to use the whatif parameter as true however this does not make the vms actually stop or start for that the parameter needs to be false document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
8,439
11,599,212,124
IssuesEvent
2020-02-25 01:29:17
SE-Garden/tms-webserver
https://api.github.com/repos/SE-Garden/tms-webserver
closed
OpenAPIのプロジェクトをGitSubmoduleで取り込む
kind:アーキ process:検討
## 概要 OpenAPIのプロジェクトをGitSubmoduleで取り込むように修正。 以降のOpenAPIのYAMLの本物をtms-oas3とする。 ## ゴール - git submodule add - rm yaml ## 成果物 - git 設定 ## 関連Issue - none
1.0
OpenAPIのプロジェクトをGitSubmoduleで取り込む - ## 概要 OpenAPIのプロジェクトをGitSubmoduleで取り込むように修正。 以降のOpenAPIのYAMLの本物をtms-oas3とする。 ## ゴール - git submodule add - rm yaml ## 成果物 - git 設定 ## 関連Issue - none
process
openapiのプロジェクトをgitsubmoduleで取り込む 概要 openapiのプロジェクトをgitsubmoduleで取り込むように修正。 以降のopenapiのyamlの本物をtms 。 ゴール git submodule add rm yaml 成果物 git 設定 関連issue none
1
53,083
6,292,464,834
IssuesEvent
2017-07-20 05:46:28
ElucidataInc/ElMaven
https://api.github.com/repos/ElucidataInc/ElMaven
closed
Crash: Bookmarking group from Automatic peak detection table and then going back to isotope of the same group (in automatic peaks table) leads to crash
bug confirmed fixed high tested
Steps to reproduce: 1. Do Automatic peak detection 2. Select any group from the peaks table 3. From the EIC widget, bookmark the group by double clicking. 4. Go to the same group in the peaks table (Obtained from the automatic peak picking). 5. Iterate through the isotopes of that same group. This step will lead to crash. This crash is there in build 776 and elMaven but not in build 682.
1.0
Crash: Bookmarking group from Automatic peak detection table and then going back to isotope of the same group (in automatic peaks table) leads to crash - Steps to reproduce: 1. Do Automatic peak detection 2. Select any group from the peaks table 3. From the EIC widget, bookmark the group by double clicking. 4. Go to the same group in the peaks table (Obtained from the automatic peak picking). 5. Iterate through the isotopes of that same group. This step will lead to crash. This crash is there in build 776 and elMaven but not in build 682.
non_process
crash bookmarking group from automatic peak detection table and then going back to isotope of the same group in automatic peaks table leads to crash steps to reproduce do automatic peak detection select any group from the peaks table from the eic widget bookmark the group by double clicking go to the same group in the peaks table obtained from the automatic peak picking iterate through the isotopes of that same group this step will lead to crash this crash is there in build and elmaven but not in build
0
801,356
28,484,870,952
IssuesEvent
2023-04-18 07:07:14
SirSorensen/Seculizer
https://api.github.com/repos/SirSorensen/Seculizer
closed
Supporting knowledge of other participants
enhancement specification low priority
In some of OFMC's examples, it is specified when a participant has knowledge of another participant. Is this something that should be supported?
1.0
Supporting knowledge of other participants - In some of OFMC's examples, it is specified when a participant has knowledge of another participant. Is this something that should be supported?
non_process
supporting knowledge of other participants in some of ofmc s examples it is specified when a participant has knowledge of another participant is this something that should be supported
0
77,349
15,528,258,122
IssuesEvent
2021-03-13 10:06:17
jonathan-wiens/hwr-kurs19a-g2
https://api.github.com/repos/jonathan-wiens/hwr-kurs19a-g2
opened
CVE-2020-24025 (Medium) detected in node-sass-4.14.1.tgz
security vulnerability
## CVE-2020-24025 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: hwr-kurs19a-g2/package.json</p> <p>Path to vulnerable library: hwr-kurs19a-g2/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jonathan-wiens/hwr-kurs19a-g2/commit/e8dc35c22b4d589d2941dc86cb947a0137795c70">e8dc35c22b4d589d2941dc86cb947a0137795c70</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Certificate validation in node-sass 2.0.0 to 4.14.1 is disabled when requesting binaries even if the user is not specifying an alternative download path. <p>Publish Date: 2021-01-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24025>CVE-2020-24025</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-24025 (Medium) detected in node-sass-4.14.1.tgz - ## CVE-2020-24025 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: hwr-kurs19a-g2/package.json</p> <p>Path to vulnerable library: hwr-kurs19a-g2/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jonathan-wiens/hwr-kurs19a-g2/commit/e8dc35c22b4d589d2941dc86cb947a0137795c70">e8dc35c22b4d589d2941dc86cb947a0137795c70</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Certificate validation in node-sass 2.0.0 to 4.14.1 is disabled when requesting binaries even if the user is not specifying an alternative download path. <p>Publish Date: 2021-01-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24025>CVE-2020-24025</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in node sass tgz cve medium severity vulnerability vulnerable library node sass tgz wrapper around libsass library home page a href path to dependency file hwr package json path to vulnerable library hwr node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details certificate validation in node sass to is disabled when requesting binaries even if the user is not specifying an alternative download path publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href step up your open source security game with whitesource
0
8,550
11,726,304,715
IssuesEvent
2020-03-10 14:19:46
TOMP-WG/TOMP-API
https://api.github.com/repos/TOMP-WG/TOMP-API
opened
Use Git Flow for changes to the API
process
As part of the improved versioning and use of git's features, we should also make better and more consistent use of branches. I suggest we use a variant of the ubiquitous Git Flow style of doing branches. Concretely, this means: - the `master` branch is **only** used for release commits (so v1.2 -> v.1.2.1 -> v1.3 -> v2.0 etc.) - a `develop` branch (or multiple) is used for incremental changes - when a develop version is ready for release, it briefly goes into a `release-v.x.x.x` branch to make the last few changes and bug fixes before being merged into master with the next version number - bigger changes that need a lot of work happen in `feature/<x>` branches that branch off from develop and are merged back into it when completed - if a bug needs a quick fix, a `hotfix` branch can be made branching from a release in master and be merged back into it with a version bump Git Flow is the most common organisation of branches used among projects that have any such organisation. It is completely compatible and even symbiotic with the other development practices we intend to adopt like semantic versioning #102 and releases using tags #107. Since it is about how we use branches I cannot put this change in a PR, so I'll create the branches already to show how it'd work. This is non-destructive so we can revert to the current practice if we do not approve. See [https://nvie.com/posts/a-successful-git-branching-model/](https://nvie.com/posts/a-successful-git-branching-model/) for the Git Flow specification.
1.0
Use Git Flow for changes to the API - As part of the improved versioning and use of git's features, we should also make better and more consistent use of branches. I suggest we use a variant of the ubiquitous Git Flow style of doing branches. Concretely, this means: - the `master` branch is **only** used for release commits (so v1.2 -> v.1.2.1 -> v1.3 -> v2.0 etc.) - a `develop` branch (or multiple) is used for incremental changes - when a develop version is ready for release, it briefly goes into a `release-v.x.x.x` branch to make the last few changes and bug fixes before being merged into master with the next version number - bigger changes that need a lot of work happen in `feature/<x>` branches that branch off from develop and are merged back into it when completed - if a bug needs a quick fix, a `hotfix` branch can be made branching from a release in master and be merged back into it with a version bump Git Flow is the most common organisation of branches used among projects that have any such organisation. It is completely compatible and even symbiotic with the other development practices we intend to adopt like semantic versioning #102 and releases using tags #107. Since it is about how we use branches I cannot put this change in a PR, so I'll create the branches already to show how it'd work. This is non-destructive so we can revert to the current practice if we do not approve. See [https://nvie.com/posts/a-successful-git-branching-model/](https://nvie.com/posts/a-successful-git-branching-model/) for the Git Flow specification.
process
use git flow for changes to the api as part of the improved versioning and use of git s features we should also make better and more consistent use of branches i suggest we use a variant of the ubiquitous git flow style of doing branches concretely this means the master branch is only used for release commits so v etc a develop branch or multiple is used for incremental changes when a develop version is ready for release it briefly goes into a release v x x x branch to make the last few changes and bug fixes before being merged into master with the next version number bigger changes that need a lot of work happen in feature branches that branch off from develop and are merged back into it when completed if a bug needs a quick fix a hotfix branch can be made branching from a release in master and be merged back into it with a version bump git flow is the most common organisation of branches used among projects that have any such organisation it is completely compatible and even symbiotic with the other development practices we intend to adopt like semantic versioning and releases using tags since it is about how we use branches i cannot put this change in a pr so i ll create the branches already to show how it d work this is non destructive so we can revert to the current practice if we do not approve see for the git flow specification
1
30
2,499,412,562
IssuesEvent
2015-01-08 00:12:23
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
closed
Using enums to solve the g.of() anonymous traversal problem. [proposal]
enhancement process
### PROBLEM Now that we can allow nested traversals given that cloning and resetting are native concepts to `Step`, we will start to see a lot more *anonymous traversal* creation via `g.of()` (e.g. `match()`, `choose()`, `local()`, ...and soon, `repeat()` https://github.com/tinkerpop/tinkerpop3/issues/411#issuecomment-68500911). ```java g.of().out().in() ``` In Gremlin2 (and Gremlin1), we supported `_()` for anonymous traversals. There are two problems with this: . `_` is **not** a legal variable or method name in Java8+. . `_` is **not** tied to a `Graph` so you don't know which graph engine you are using and thus, no strategies, no DSLs, etc. ### POTENTIAL SOLUTION I propose that each vendor provides a `G` enum in their `GraphTraversal` implementation. ```java public TinkerGraphTraversal<S,E> implements GraphTraversal<S,E> { public static enum G implements GraphTraversal { _g { public GraphTraversal start() { return new TinkerGraphTraversal().start(); } } } } ``` Now, `_` by itself is no longer allowed in Java8, but `_g` is. So, with `static import`, you can do this: ```java g.V().out().local(_g.outE().order().by('weight').limit(10).inV()).values("name") // vs g.V().out().local(g.of().outE().order().by('weight').limit(10).inV()).values("name") ``` *"So that is all fine and dandy, but what about DSLs (...like `Neo4jGraphTraversal` that supports `cypher()`)."* ```java public Neo4jGraphTraversal<S,E> implements Neo4jTraversal<S,E> { public static enum G implements Neo4jTraversal { _g { public Neo4jTraversal start() { return new Neo4jGraphTraversal().start(); } } } } ``` Thus, when `g instanceof Neo4jGraph`: ``` g.V().out().local(_g.cypher('MATCH (a)->(b)').select('a')) ``` In this way, it is assumed that the user's code base is tied to a particular vendor implementation and thus, they won't `static import` `G` from OrientDB and Neo4j at the same time. ... the only counter argument is the use of `TinkerGraph` and `Neo4j` being used at the same time :( (discussed later). The two problems mentioned in the first section are now solved: . `_g` is a legal, reasonable character sequence to start an anonymous traversal. . `_g` is provided by the vendor and thus, tied to the DSL/strategies of that vendor. ### PROBLEM WITH TWO GRAPHS IN THE SAME CODEBASE Lots of people will use TinkerGraph with their preferred graph engine (e.g. Neo4j, Titan, OrientDB, etc.). This raises the issue of `_g` being statically imported from two places. With: ```java TinkerGraphTraversal { public enum G ... } Neo4jGraphTraversal { public enum G ... } ``` You can `import static` one, and not the other: ```java tinker.V().local(TinkerGraphTraversal.G._g.outE().inV()).values("name") neo4j.V().local(_g.outE().inV()).values("name") ``` Sorta crappy...or we stick with `g.of()`. (which, of course, is the easiest for me :)). ### CONCLUSION Thoughts? @dkuppitz @mbroecheler @spmallette @mpollmeier
1.0
Using enums to solve the g.of() anonymous traversal problem. [proposal] - ### PROBLEM Now that we can allow nested traversals given that cloning and resetting are native concepts to `Step`, we will start to see a lot more *anonymous traversal* creation via `g.of()` (e.g. `match()`, `choose()`, `local()`, ...and soon, `repeat()` https://github.com/tinkerpop/tinkerpop3/issues/411#issuecomment-68500911). ```java g.of().out().in() ``` In Gremlin2 (and Gremlin1), we supported `_()` for anonymous traversals. There are two problems with this: . `_` is **not** a legal variable or method name in Java8+. . `_` is **not** tied to a `Graph` so you don't know which graph engine you are using and thus, no strategies, no DSLs, etc. ### POTENTIAL SOLUTION I propose that each vendor provides a `G` enum in their `GraphTraversal` implementation. ```java public TinkerGraphTraversal<S,E> implements GraphTraversal<S,E> { public static enum G implements GraphTraversal { _g { public GraphTraversal start() { return new TinkerGraphTraversal().start(); } } } } ``` Now, `_` by itself is no longer allowed in Java8, but `_g` is. So, with `static import`, you can do this: ```java g.V().out().local(_g.outE().order().by('weight').limit(10).inV()).values("name") // vs g.V().out().local(g.of().outE().order().by('weight').limit(10).inV()).values("name") ``` *"So that is all fine and dandy, but what about DSLs (...like `Neo4jGraphTraversal` that supports `cypher()`)."* ```java public Neo4jGraphTraversal<S,E> implements Neo4jTraversal<S,E> { public static enum G implements Neo4jTraversal { _g { public Neo4jTraversal start() { return new Neo4jGraphTraversal().start(); } } } } ``` Thus, when `g instanceof Neo4jGraph`: ``` g.V().out().local(_g.cypher('MATCH (a)->(b)').select('a')) ``` In this way, it is assumed that the user's code base is tied to a particular vendor implementation and thus, they won't `static import` `G` from OrientDB and Neo4j at the same time. ... the only counter argument is the use of `TinkerGraph` and `Neo4j` being used at the same time :( (discussed later). The two problems mentioned in the first section are now solved: . `_g` is a legal, reasonable character sequence to start an anonymous traversal. . `_g` is provided by the vendor and thus, tied to the DSL/strategies of that vendor. ### PROBLEM WITH TWO GRAPHS IN THE SAME CODEBASE Lots of people will use TinkerGraph with their preferred graph engine (e.g. Neo4j, Titan, OrientDB, etc.). This raises the issue of `_g` being statically imported from two places. With: ```java TinkerGraphTraversal { public enum G ... } Neo4jGraphTraversal { public enum G ... } ``` You can `import static` one, and not the other: ```java tinker.V().local(TinkerGraphTraversal.G._g.outE().inV()).values("name") neo4j.V().local(_g.outE().inV()).values("name") ``` Sorta crappy...or we stick with `g.of()`. (which, of course, is the easiest for me :)). ### CONCLUSION Thoughts? @dkuppitz @mbroecheler @spmallette @mpollmeier
process
using enums to solve the g of anonymous traversal problem problem now that we can allow nested traversals given that cloning and resetting are native concepts to step we will start to see a lot more anonymous traversal creation via g of e g match choose local and soon repeat java g of out in in and we supported for anonymous traversals there are two problems with this is not a legal variable or method name in is not tied to a graph so you don t know which graph engine you are using and thus no strategies no dsls etc potential solution i propose that each vendor provides a g enum in their graphtraversal implementation java public tinkergraphtraversal implements graphtraversal public static enum g implements graphtraversal g public graphtraversal start return new tinkergraphtraversal start now by itself is no longer allowed in but g is so with static import you can do this java g v out local g oute order by weight limit inv values name vs g v out local g of oute order by weight limit inv values name so that is all fine and dandy but what about dsls like that supports cypher java public implements public static enum g implements g public start return new start thus when g instanceof g v out local g cypher match a b select a in this way it is assumed that the user s code base is tied to a particular vendor implementation and thus they won t static import g from orientdb and at the same time the only counter argument is the use of tinkergraph and being used at the same time discussed later the two problems mentioned in the first section are now solved g is a legal reasonable character sequence to start an anonymous traversal g is provided by the vendor and thus tied to the dsl strategies of that vendor problem with two graphs in the same codebase lots of people will use tinkergraph with their preferred graph engine e g titan orientdb etc this raises the issue of g being statically imported from two places with java tinkergraphtraversal public enum g public enum g you can import static one and not the other java tinker v local tinkergraphtraversal g g oute inv values name v local g oute inv values name sorta crappy or we stick with g of which of course is the easiest for me conclusion thoughts dkuppitz mbroecheler spmallette mpollmeier
1
7,447
10,555,473,334
IssuesEvent
2019-10-03 21:59:10
Stephan-S/FS19_AutoDrive
https://api.github.com/repos/Stephan-S/FS19_AutoDrive
closed
I don't know but submited
Delayed processing
Hi Balu, ready for a headache ? I submit you I think there is mistake somewhere causing size files differ and to a moment AD say me it not access to destination unreachable as solo mode, but after reload the game it can reach (3nd file). The languages translations always not seem taken in multiplayer but work in solo of course because using always the same and latest version ^^. Adding you too some server few lines errors I was copy not sure if this linked too. And also these files have been used in the 2 way servers multiplayer and solo. Ok 1st file I had: [AutoDrive_Village_Yagodnoe_config.xml.txt](https://github.com/Stephan-S/FS19_AutoDrive/files/3186261/AutoDrive_Village_Yagodnoe_config.xml.txt) The 2nd file I have in the server was up without players on it for few days or modifying AD route file sizes was changed as you can see saved after saved: [AutoDrive_Village_Yagodnoe_config.xml.txt](https://github.com/Stephan-S/FS19_AutoDrive/files/3186278/AutoDrive_Village_Yagodnoe_config.xml.txt) The 3rd I just added a joint connection point in solo: [AutoDrive_Village_Yagodnoe_config.xml.txt](https://github.com/Stephan-S/FS19_AutoDrive/files/3186384/AutoDrive_Village_Yagodnoe_config.xml.txt) And the 4th I just added few new points in solo: [AutoDrive_Village_Yagodnoe_config.xml.txt](https://github.com/Stephan-S/FS19_AutoDrive/files/3186316/AutoDrive_Village_Yagodnoe_config.xml.txt) The file size back down again and destinations work perfectly. Server side around the 1st file: ``` d:/servers/service4848/home/games/FarmingSimulator19/mods/FS19_AutoDrive/scripts/AutoDriveInputFunctions.lua:460: attempt to index a nil value Error: Running LUA method 'update'. d:/servers/service4848/home/games/FarmingSimulator19/mods/FS19_AutoDrive/scripts/AutoDriveInputFunctions.lua:442: attempt to index a nil value Error: Running LUA method 'update'. Error: Running LUA method 'update'. d:/servers/service4848/home/games/FarmingSimulator19/mods/FS19_AutoDrive/scripts/AutoDriveInputFunctions.lua:442: attempt to index a nil value AD: creating xml file at d:/servers/service4848/home/games/FarmingSimulator19/FS19_AutoDrive_Export/AutoDrive_Village_Yagodnoe_config.xml AD: Set to recalculating routes Error: Running LUA method 'update'. d:/servers/service4848/home/games/FarmingSimulator19/mods/FS19_AutoDrive/scripts/AutoDriveXML.lua:420: invalid value (nil) at index 2 in table for 'concat' AD: creating xml file at d:/servers/service4848/home/games/FarmingSimulator19/FS19_AutoDrive_Export/AutoDrive_Village_Yagodnoe_config.xml AD: Set to recalculating routes Error: Running LUA method 'update'. ``` I hope this help to improve stability particularly in multiplayer. And if there no issue then forget ! :) My best regard and thaaaaannnnnkkkkksss for AD !
1.0
I don't know but submited - Hi Balu, ready for a headache ? I submit you I think there is mistake somewhere causing size files differ and to a moment AD say me it not access to destination unreachable as solo mode, but after reload the game it can reach (3nd file). The languages translations always not seem taken in multiplayer but work in solo of course because using always the same and latest version ^^. Adding you too some server few lines errors I was copy not sure if this linked too. And also these files have been used in the 2 way servers multiplayer and solo. Ok 1st file I had: [AutoDrive_Village_Yagodnoe_config.xml.txt](https://github.com/Stephan-S/FS19_AutoDrive/files/3186261/AutoDrive_Village_Yagodnoe_config.xml.txt) The 2nd file I have in the server was up without players on it for few days or modifying AD route file sizes was changed as you can see saved after saved: [AutoDrive_Village_Yagodnoe_config.xml.txt](https://github.com/Stephan-S/FS19_AutoDrive/files/3186278/AutoDrive_Village_Yagodnoe_config.xml.txt) The 3rd I just added a joint connection point in solo: [AutoDrive_Village_Yagodnoe_config.xml.txt](https://github.com/Stephan-S/FS19_AutoDrive/files/3186384/AutoDrive_Village_Yagodnoe_config.xml.txt) And the 4th I just added few new points in solo: [AutoDrive_Village_Yagodnoe_config.xml.txt](https://github.com/Stephan-S/FS19_AutoDrive/files/3186316/AutoDrive_Village_Yagodnoe_config.xml.txt) The file size back down again and destinations work perfectly. Server side around the 1st file: ``` d:/servers/service4848/home/games/FarmingSimulator19/mods/FS19_AutoDrive/scripts/AutoDriveInputFunctions.lua:460: attempt to index a nil value Error: Running LUA method 'update'. d:/servers/service4848/home/games/FarmingSimulator19/mods/FS19_AutoDrive/scripts/AutoDriveInputFunctions.lua:442: attempt to index a nil value Error: Running LUA method 'update'. Error: Running LUA method 'update'. d:/servers/service4848/home/games/FarmingSimulator19/mods/FS19_AutoDrive/scripts/AutoDriveInputFunctions.lua:442: attempt to index a nil value AD: creating xml file at d:/servers/service4848/home/games/FarmingSimulator19/FS19_AutoDrive_Export/AutoDrive_Village_Yagodnoe_config.xml AD: Set to recalculating routes Error: Running LUA method 'update'. d:/servers/service4848/home/games/FarmingSimulator19/mods/FS19_AutoDrive/scripts/AutoDriveXML.lua:420: invalid value (nil) at index 2 in table for 'concat' AD: creating xml file at d:/servers/service4848/home/games/FarmingSimulator19/FS19_AutoDrive_Export/AutoDrive_Village_Yagodnoe_config.xml AD: Set to recalculating routes Error: Running LUA method 'update'. ``` I hope this help to improve stability particularly in multiplayer. And if there no issue then forget ! :) My best regard and thaaaaannnnnkkkkksss for AD !
process
i don t know but submited hi balu ready for a headache i submit you i think there is mistake somewhere causing size files differ and to a moment ad say me it not access to destination unreachable as solo mode but after reload the game it can reach file the languages translations always not seem taken in multiplayer but work in solo of course because using always the same and latest version adding you too some server few lines errors i was copy not sure if this linked too and also these files have been used in the way servers multiplayer and solo ok file i had the file i have in the server was up without players on it for few days or modifying ad route file sizes was changed as you can see saved after saved the i just added a joint connection point in solo and the i just added few new points in solo the file size back down again and destinations work perfectly server side around the file d servers home games mods autodrive scripts autodriveinputfunctions lua attempt to index a nil value error running lua method update d servers home games mods autodrive scripts autodriveinputfunctions lua attempt to index a nil value error running lua method update error running lua method update d servers home games mods autodrive scripts autodriveinputfunctions lua attempt to index a nil value ad creating xml file at d servers home games autodrive export autodrive village yagodnoe config xml ad set to recalculating routes error running lua method update d servers home games mods autodrive scripts autodrivexml lua invalid value nil at index in table for concat ad creating xml file at d servers home games autodrive export autodrive village yagodnoe config xml ad set to recalculating routes error running lua method update i hope this help to improve stability particularly in multiplayer and if there no issue then forget my best regard and thaaaaannnnnkkkkksss for ad
1
171,616
6,492,209,766
IssuesEvent
2017-08-21 12:11:32
angular/angular-cli
https://api.github.com/repos/angular/angular-cli
closed
Feature: Exclude files from 'watch' mode
community: help wanted priority: 2 (required) severity2: inconvenient type: bug
I think it would be great if you can exclude files to not watch so that it doesn't recompile every time. As of now, I'm using styles.scss to override the default styles.css. Yes, I do know that component can include the scss but I rather just have 1 style file for entire application (it's very small app). Anyways, everytime I change scss files then it gets compiled twice. Thanks and definitely loving NG2.
1.0
Feature: Exclude files from 'watch' mode - I think it would be great if you can exclude files to not watch so that it doesn't recompile every time. As of now, I'm using styles.scss to override the default styles.css. Yes, I do know that component can include the scss but I rather just have 1 style file for entire application (it's very small app). Anyways, everytime I change scss files then it gets compiled twice. Thanks and definitely loving NG2.
non_process
feature exclude files from watch mode i think it would be great if you can exclude files to not watch so that it doesn t recompile every time as of now i m using styles scss to override the default styles css yes i do know that component can include the scss but i rather just have style file for entire application it s very small app anyways everytime i change scss files then it gets compiled twice thanks and definitely loving
0