Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
970
labels
stringlengths
4
625
body
stringlengths
3
247k
index
stringclasses
9 values
text_combine
stringlengths
96
247k
label
stringclasses
2 values
text
stringlengths
96
218k
binary_label
int64
0
1
27,222
7,924,329,450
IssuesEvent
2018-07-05 16:34:27
WH2-WEC/eom_marienburg
https://api.github.com/repos/WH2-WEC/eom_marienburg
opened
Create mutual protection script
Feature Building
If declared on by Empire factions, gain bonus with bret and vice versa.
1.0
Create mutual protection script - If declared on by Empire factions, gain bonus with bret and vice versa.
non_perf
create mutual protection script if declared on by empire factions gain bonus with bret and vice versa
0
15,724
9,029,940,983
IssuesEvent
2019-02-08 01:09:51
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
AttributeError: module 'tensorflow.compat' has no attribute 'v1'
comp:apis type:bug/performance
I ran cnn_mnist.py on my machine but got "AttributeError: module 'tensorflow.compat' has no attribute 'v1'". can anyone solve this problem?
True
AttributeError: module 'tensorflow.compat' has no attribute 'v1' - I ran cnn_mnist.py on my machine but got "AttributeError: module 'tensorflow.compat' has no attribute 'v1'". can anyone solve this problem?
perf
attributeerror module tensorflow compat has no attribute i ran cnn mnist py on my machine but got attributeerror module tensorflow compat has no attribute can anyone solve this problem
1
256,969
19,480,534,275
IssuesEvent
2021-12-25 06:45:04
TuSimple/naive-ui
https://api.github.com/repos/TuSimple/naive-ui
closed
frFR locale update
documentation
<!-- generated by issue-helper DO NOT REMOVE __BUG__ --> ### TuSimple/naive-ui version (版本) 2.23.1 ### Vue version (Vue 版本) 3.2.26 ### Browser and its version (浏览器及其版本) N/A ### System and its version (系统及其版本) N/A ### Node version (Node 版本) ### Reappearance link (重现链接) https://github.com/TuSimple/naive-ui/blob/main/src/locales/common/frFR.ts ### Reappearance steps (重现步骤) Choose frFr locale ### Expected results (期望的结果) Updated frFR locale translation ### Actual results (实际的结果) Outdated frFR locale translation ### Remarks (补充说明) Here is the updated frFr locale translation: ```js import type { NLocale } from './enUS' const frFR: NLocale = { name: 'fr-FR', global: { undo: 'Défaire', redo: 'Refaire', confirm: 'Confirmer' }, Popconfirm: { positiveText: 'Confirmer', negativeText: 'Annuler' }, Cascader: { placeholder: 'Sélectionner', loading: 'Chargement', loadingRequiredMessage: (label: string): string => `Charger tous les enfants de ${label} avant de le sélectionner` }, Time: { dateFormat: 'dd/MM/yyyy', dateTimeFormat: 'dd/MM/yyyy HH:mm:ss' }, DatePicker: { yearFormat: 'yyyy', monthFormat: 'MMM', dayFormat: 'eeeeee', yearTypeFormat: 'yyyy', monthTypeFormat: 'MM/yyyy', dateFormat: 'dd/MM/yyyy', dateTimeFormat: 'dd/MM/yyyy HH:mm:ss', quarterFormat: 'qqq yyyy', clear: 'Effacer', now: 'Maintenant', confirm: 'Confirmer', selectTime: "Sélectionner l'heure", selectDate: 'Sélectionner la date', datePlaceholder: 'Sélectionner la date', datetimePlaceholder: "Sélectionner la date et l'heure", monthPlaceholder: 'Sélectionner le mois', yearPlaceholder: "Sélectionner l'année", quarterPlaceholder: 'Sélectionner le trimestre', startDatePlaceholder: 'Date de début', endDatePlaceholder: 'Date de fin', startDatetimePlaceholder: 'Date et heure de début', endDatetimePlaceholder: 'Date et heure de fin', monthBeforeYear: true, firstDayOfWeek: 1, today: "Aujourd'hui" }, DataTable: { checkTableAll: 'Sélectionner tout', uncheckTableAll: 'Désélectionner tout', confirm: 'Confirmer', clear: 'Effacer' }, Transfer: { sourceTitle: 'Source', targetTitle: 'Cible' }, Empty: { description: 'Aucune donnée' }, Select: { placeholder: 'Sélectionner' }, TimePicker: { placeholder: "Sélectionner l'heure", positiveText: 'OK', negativeText: 'Annuler', now: 'Maintenant' }, Pagination: { goto: 'Aller à', selectionSuffix: 'page' }, DynamicTags: { add: 'Ajouter' }, Log: { loading: 'Chargement' }, Input: { placeholder: 'Saisir' }, InputNumber: { placeholder: 'Saisir' }, DynamicInput: { create: 'Créer' }, ThemeEditor: { title: 'Editeur de thème', clearAllVars: 'Effacer toutes les variables', clearSearch: 'Effacer la recherche', filterCompName: 'Filtrer par nom de composant', filterVarName: 'Filtrer par nom de variable', import: 'Importer', export: 'Exporter', restore: 'Réinitialiser' } } export default frFR ``` <!-- generated by issue-helper DO NOT REMOVE __BUG__ -->
1.0
frFR locale update - <!-- generated by issue-helper DO NOT REMOVE __BUG__ --> ### TuSimple/naive-ui version (版本) 2.23.1 ### Vue version (Vue 版本) 3.2.26 ### Browser and its version (浏览器及其版本) N/A ### System and its version (系统及其版本) N/A ### Node version (Node 版本) ### Reappearance link (重现链接) https://github.com/TuSimple/naive-ui/blob/main/src/locales/common/frFR.ts ### Reappearance steps (重现步骤) Choose frFr locale ### Expected results (期望的结果) Updated frFR locale translation ### Actual results (实际的结果) Outdated frFR locale translation ### Remarks (补充说明) Here is the updated frFr locale translation: ```js import type { NLocale } from './enUS' const frFR: NLocale = { name: 'fr-FR', global: { undo: 'Défaire', redo: 'Refaire', confirm: 'Confirmer' }, Popconfirm: { positiveText: 'Confirmer', negativeText: 'Annuler' }, Cascader: { placeholder: 'Sélectionner', loading: 'Chargement', loadingRequiredMessage: (label: string): string => `Charger tous les enfants de ${label} avant de le sélectionner` }, Time: { dateFormat: 'dd/MM/yyyy', dateTimeFormat: 'dd/MM/yyyy HH:mm:ss' }, DatePicker: { yearFormat: 'yyyy', monthFormat: 'MMM', dayFormat: 'eeeeee', yearTypeFormat: 'yyyy', monthTypeFormat: 'MM/yyyy', dateFormat: 'dd/MM/yyyy', dateTimeFormat: 'dd/MM/yyyy HH:mm:ss', quarterFormat: 'qqq yyyy', clear: 'Effacer', now: 'Maintenant', confirm: 'Confirmer', selectTime: "Sélectionner l'heure", selectDate: 'Sélectionner la date', datePlaceholder: 'Sélectionner la date', datetimePlaceholder: "Sélectionner la date et l'heure", monthPlaceholder: 'Sélectionner le mois', yearPlaceholder: "Sélectionner l'année", quarterPlaceholder: 'Sélectionner le trimestre', startDatePlaceholder: 'Date de début', endDatePlaceholder: 'Date de fin', startDatetimePlaceholder: 'Date et heure de début', endDatetimePlaceholder: 'Date et heure de fin', monthBeforeYear: true, firstDayOfWeek: 1, today: "Aujourd'hui" }, DataTable: { checkTableAll: 'Sélectionner tout', uncheckTableAll: 'Désélectionner tout', confirm: 'Confirmer', clear: 'Effacer' }, Transfer: { sourceTitle: 'Source', targetTitle: 'Cible' }, Empty: { description: 'Aucune donnée' }, Select: { placeholder: 'Sélectionner' }, TimePicker: { placeholder: "Sélectionner l'heure", positiveText: 'OK', negativeText: 'Annuler', now: 'Maintenant' }, Pagination: { goto: 'Aller à', selectionSuffix: 'page' }, DynamicTags: { add: 'Ajouter' }, Log: { loading: 'Chargement' }, Input: { placeholder: 'Saisir' }, InputNumber: { placeholder: 'Saisir' }, DynamicInput: { create: 'Créer' }, ThemeEditor: { title: 'Editeur de thème', clearAllVars: 'Effacer toutes les variables', clearSearch: 'Effacer la recherche', filterCompName: 'Filtrer par nom de composant', filterVarName: 'Filtrer par nom de variable', import: 'Importer', export: 'Exporter', restore: 'Réinitialiser' } } export default frFR ``` <!-- generated by issue-helper DO NOT REMOVE __BUG__ -->
non_perf
frfr locale update tusimple naive ui version 版本 vue version vue 版本 browser and its version 浏览器及其版本 n a system and its version 系统及其版本 n a node version node 版本 reappearance link 重现链接 reappearance steps 重现步骤 choose frfr locale expected results 期望的结果 updated frfr locale translation actual results 实际的结果 outdated frfr locale translation remarks 补充说明 here is the updated frfr locale translation js import type nlocale from enus const frfr nlocale name fr fr global undo défaire redo refaire confirm confirmer popconfirm positivetext confirmer negativetext annuler cascader placeholder sélectionner loading chargement loadingrequiredmessage label string string charger tous les enfants de label avant de le sélectionner time dateformat dd mm yyyy datetimeformat dd mm yyyy hh mm ss datepicker yearformat yyyy monthformat mmm dayformat eeeeee yeartypeformat yyyy monthtypeformat mm yyyy dateformat dd mm yyyy datetimeformat dd mm yyyy hh mm ss quarterformat qqq yyyy clear effacer now maintenant confirm confirmer selecttime sélectionner l heure selectdate sélectionner la date dateplaceholder sélectionner la date datetimeplaceholder sélectionner la date et l heure monthplaceholder sélectionner le mois yearplaceholder sélectionner l année quarterplaceholder sélectionner le trimestre startdateplaceholder date de début enddateplaceholder date de fin startdatetimeplaceholder date et heure de début enddatetimeplaceholder date et heure de fin monthbeforeyear true firstdayofweek today aujourd hui datatable checktableall sélectionner tout unchecktableall désélectionner tout confirm confirmer clear effacer transfer sourcetitle source targettitle cible empty description aucune donnée select placeholder sélectionner timepicker placeholder sélectionner l heure positivetext ok negativetext annuler now maintenant pagination goto aller à selectionsuffix page dynamictags add ajouter log loading chargement input placeholder saisir inputnumber placeholder saisir dynamicinput create créer themeeditor title editeur de thème clearallvars effacer toutes les variables clearsearch effacer la recherche filtercompname filtrer par nom de composant filtervarname filtrer par nom de variable import importer export exporter restore réinitialiser export default frfr
0
73,443
9,665,463,481
IssuesEvent
2019-05-21 08:34:55
kubernetes/test-infra
https://api.github.com/repos/kubernetes/test-infra
closed
Running kubetest with precompiled kubectl
area/kubetest kind/bug kind/documentation lifecycle/stale sig/contributor-experience
Currently `kubetest` fails with > It looks as if you don't have a compiled kubectl binary even when kubectl is in `client/bin/kubectl` (and in $PATH, but that is secondary issue). This can be fixed user-side by putting the precompiled binary into `_output/bin` or other paths defined by https://github.com/kubernetes/kubernetes/blob/d0439d417b0563d44c65b6e400e2070964dea7d1/cluster/clientbin.sh#L86 Is there a reason kube-test requires a locally compiled binary? With the new containerized kubernetes build, it is quite hard to build a custom container with kubetest as it would require docker in docker during image build (which is not currently possible). This is a followup to #8698 /sig contributor-experience /kind documentation /area kubetest
1.0
Running kubetest with precompiled kubectl - Currently `kubetest` fails with > It looks as if you don't have a compiled kubectl binary even when kubectl is in `client/bin/kubectl` (and in $PATH, but that is secondary issue). This can be fixed user-side by putting the precompiled binary into `_output/bin` or other paths defined by https://github.com/kubernetes/kubernetes/blob/d0439d417b0563d44c65b6e400e2070964dea7d1/cluster/clientbin.sh#L86 Is there a reason kube-test requires a locally compiled binary? With the new containerized kubernetes build, it is quite hard to build a custom container with kubetest as it would require docker in docker during image build (which is not currently possible). This is a followup to #8698 /sig contributor-experience /kind documentation /area kubetest
non_perf
running kubetest with precompiled kubectl currently kubetest fails with it looks as if you don t have a compiled kubectl binary even when kubectl is in client bin kubectl and in path but that is secondary issue this can be fixed user side by putting the precompiled binary into output bin or other paths defined by is there a reason kube test requires a locally compiled binary with the new containerized kubernetes build it is quite hard to build a custom container with kubetest as it would require docker in docker during image build which is not currently possible this is a followup to sig contributor experience kind documentation area kubetest
0
52,155
27,402,939,712
IssuesEvent
2023-03-01 03:03:05
dermotduffy/frigate-hass-card
https://api.github.com/repos/dermotduffy/frigate-hass-card
closed
Desktop so much faster than mobile. Why?
performance
First of all; Thank you for this incredible work! What a card! So my issues; When loading my camera page 6 cams + a snapshot gallery on chrome on my macbook, the page loads fully in about 2 seconds. When viewing the same page on my mobile (companion app on Android) the page takes minimal 5 seconds and in most cases up to 10. Is there anything I can do to optimize? HASS runs on a NUC <10% CPU. Frigate runs on another micro PC (J4125+GPU+Coral) at about 25% CPU. View provider in the card is the first option (homeassistan) but I have not done anything with webrtc yet. Sorry if I posted this in the wrong category
True
Desktop so much faster than mobile. Why? - First of all; Thank you for this incredible work! What a card! So my issues; When loading my camera page 6 cams + a snapshot gallery on chrome on my macbook, the page loads fully in about 2 seconds. When viewing the same page on my mobile (companion app on Android) the page takes minimal 5 seconds and in most cases up to 10. Is there anything I can do to optimize? HASS runs on a NUC <10% CPU. Frigate runs on another micro PC (J4125+GPU+Coral) at about 25% CPU. View provider in the card is the first option (homeassistan) but I have not done anything with webrtc yet. Sorry if I posted this in the wrong category
perf
desktop so much faster than mobile why first of all thank you for this incredible work what a card so my issues when loading my camera page cams a snapshot gallery on chrome on my macbook the page loads fully in about seconds when viewing the same page on my mobile companion app on android the page takes minimal seconds and in most cases up to is there anything i can do to optimize hass runs on a nuc cpu frigate runs on another micro pc gpu coral at about cpu view provider in the card is the first option homeassistan but i have not done anything with webrtc yet sorry if i posted this in the wrong category
1
52,067
27,360,916,180
IssuesEvent
2023-02-27 15:48:55
jupyterlab/jupyterlab
https://api.github.com/repos/jupyterlab/jupyterlab
closed
JupyterLab requests /api/nbconvert twice during load
bug pkg:notebook tag:Performance good first issue
## Description JupyterLab during webpage load performs requests to /api/nbconvert twice. ## Reproduce 1. Open Chrome DevTools, switch to Network tab 2. Reload the page with JupyterLab 3. Find /api/nbconvert requests (you could filter by 'nbconvert') 4. See there are two simultanous requests ## Expected behavior I probably would expect one request not two. The /api/nbconvert response time can be relatively slow (is quite slow in my full, non 'jlab-test' environment for some reason). ## Context - Operating System and version: CentOS - Browser and version: Edge 108.0.1462 - JupyterLab version: 3.5.2 <details><summary>Troubleshoot Output</summary> <pre> $ jupyter troubleshoot $PATH: /home/kungurov/.conda/envs/jlab-test/bin /usr/frog/scicomp/pythonds/conda/condabin /usr/frog/scicomp/pythonds/v1.2/bin /usr/frog/scicomp/pythonds/latex/bin/x86_64-linux /usr/frog/scicomp/pythonds/conda/condabin /usr/local/bin /usr/bin /usr/local/sbin /usr/sbin /home/kungurov/bin sys.path: /home/kungurov/.conda/envs/jlab-test/bin /home/kungurov/.conda/envs/jlab-test/lib/python311.zip /home/kungurov/.conda/envs/jlab-test/lib/python3.11 /home/kungurov/.conda/envs/jlab-test/lib/python3.11/lib-dynload /home/kungurov/.conda/envs/jlab-test/lib/python3.11/site-packages sys.executable: /home/kungurov/.conda/envs/jlab-test/bin/python sys.version: 3.11.0 | packaged by conda-forge | (main, Jan 14 2023, 12:27:40) [GCC 11.3.0] platform.platform(): Linux-3.10.0-1160.76.1.el7.x86_64-x86_64-with-glibc2.17 which -a jupyter: /home/kungurov/.conda/envs/jlab-test/bin/jupyter pip list: Package Version ----------------------------- ----------- anyio 3.6.2 argon2-cffi 21.3.0 argon2-cffi-bindings 21.2.0 asttokens 2.2.1 attrs 22.2.0 Babel 2.11.0 backcall 0.2.0 backports.functools-lru-cache 1.6.4 beautifulsoup4 4.11.1 bleach 5.0.1 brotlipy 0.7.0 certifi 2022.12.7 cffi 1.15.1 charset-normalizer 2.1.1 comm 0.1.2 cryptography 39.0.0 debugpy 1.6.5 decorator 5.1.1 defusedxml 0.7.1 entrypoints 0.4 executing 1.2.0 fastjsonschema 2.16.2 flit_core 3.8.0 idna 3.4 importlib-metadata 6.0.0 importlib-resources 5.10.2 ipykernel 6.20.2 ipython 8.8.0 ipython-genutils 0.2.0 jedi 0.18.2 Jinja2 3.1.2 json5 0.9.5 jsonschema 4.17.3 jupyter_client 7.4.9 jupyter_core 5.1.3 jupyter-events 0.6.3 jupyter_server 2.1.0 jupyter_server_terminals 0.4.4 jupyterlab 3.5.2 jupyterlab-pygments 0.2.2 jupyterlab_server 2.19.0 MarkupSafe 2.1.2 matplotlib-inline 0.1.6 mistune 2.0.4 nbclassic 0.4.8 nbclient 0.7.2 nbconvert 7.2.8 nbformat 5.7.3 nest-asyncio 1.5.6 notebook 6.5.2 notebook_shim 0.2.2 packaging 23.0 pandocfilters 1.5.0 parso 0.8.3 pexpect 4.8.0 pickleshare 0.7.5 pip 22.3.1 pkgutil_resolve_name 1.3.10 platformdirs 2.6.2 prometheus-client 0.15.0 prompt-toolkit 3.0.36 psutil 5.9.4 ptyprocess 0.7.0 pure-eval 0.2.2 pycparser 2.21 Pygments 2.14.0 pyOpenSSL 23.0.0 pyrsistent 0.19.3 PySocks 1.7.1 python-dateutil 2.8.2 python-json-logger 2.0.4 pytz 2022.7.1 PyYAML 6.0 pyzmq 25.0.0 requests 2.28.2 rfc3339-validator 0.1.4 rfc3986-validator 0.1.1 Send2Trash 1.8.0 setuptools 66.1.1 six 1.16.0 sniffio 1.3.0 soupsieve 2.3.2.post1 stack-data 0.6.2 terminado 0.17.1 tinycss2 1.2.1 tomli 2.0.1 tornado 6.2 traitlets 5.8.1 typing_extensions 4.4.0 urllib3 1.26.14 wcwidth 0.2.6 webencodings 0.5.1 websocket-client 1.4.2 wheel 0.38.4 zipp 3.11.0 conda list: # packages in environment at /home/kungurov/.conda/envs/jlab-test: # # Name Version Build Channel _libgcc_mutex 0.1 conda_forge defaults _openmp_mutex 4.5 2_gnu defaults anyio 3.6.2 pyhd8ed1ab_0 defaults argon2-cffi 21.3.0 pyhd8ed1ab_0 defaults argon2-cffi-bindings 21.2.0 py311hd4cff14_3 defaults asttokens 2.2.1 pyhd8ed1ab_0 defaults attrs 22.2.0 pyh71513ae_0 defaults babel 2.11.0 pyhd8ed1ab_0 defaults backcall 0.2.0 pyh9f0ad1d_0 defaults backports 1.0 pyhd8ed1ab_3 defaults backports.functools_lru_cache 1.6.4 pyhd8ed1ab_0 defaults beautifulsoup4 4.11.1 pyha770c72_0 defaults bleach 5.0.1 pyhd8ed1ab_0 defaults brotlipy 0.7.0 py311hd4cff14_1005 defaults bzip2 1.0.8 h7f98852_4 defaults ca-certificates 2022.12.7 ha878542_0 defaults certifi 2022.12.7 pyhd8ed1ab_0 defaults cffi 1.15.1 py311h409f033_3 defaults charset-normalizer 2.1.1 pyhd8ed1ab_0 defaults comm 0.1.2 pyhd8ed1ab_0 defaults cryptography 39.0.0 py311h9b4c7bb_0 defaults debugpy 1.6.5 py311hcafe171_0 defaults decorator 5.1.1 pyhd8ed1ab_0 defaults defusedxml 0.7.1 pyhd8ed1ab_0 defaults entrypoints 0.4 pyhd8ed1ab_0 defaults executing 1.2.0 pyhd8ed1ab_0 defaults flit-core 3.8.0 pyhd8ed1ab_0 defaults idna 3.4 pyhd8ed1ab_0 defaults importlib-metadata 6.0.0 pyha770c72_0 defaults importlib_resources 5.10.2 pyhd8ed1ab_0 defaults ipykernel 6.20.2 pyh210e3f2_0 defaults ipython 8.8.0 pyh41d4057_0 defaults ipython_genutils 0.2.0 py_1 defaults jedi 0.18.2 pyhd8ed1ab_0 defaults jinja2 3.1.2 pyhd8ed1ab_1 defaults json5 0.9.5 pyh9f0ad1d_0 defaults jsonschema 4.17.3 pyhd8ed1ab_0 defaults jupyter_client 7.4.9 pyhd8ed1ab_0 defaults jupyter_core 5.1.3 py311h38be061_0 defaults jupyter_events 0.6.3 pyhd8ed1ab_0 defaults jupyter_server 2.1.0 pyhd8ed1ab_0 defaults jupyter_server_terminals 0.4.4 pyhd8ed1ab_1 defaults jupyterlab 3.5.2 pyhd8ed1ab_0 defaults jupyterlab_pygments 0.2.2 pyhd8ed1ab_0 defaults jupyterlab_server 2.19.0 pyhd8ed1ab_0 defaults ld_impl_linux-64 2.39 hcc3a1bd_1 defaults libffi 3.4.2 h7f98852_5 defaults libgcc-ng 12.2.0 h65d4601_19 defaults libgomp 12.2.0 h65d4601_19 defaults libnsl 2.0.0 h7f98852_0 defaults libsodium 1.0.18 h36c2ea0_1 defaults libsqlite 3.40.0 h753d276_0 defaults libstdcxx-ng 12.2.0 h46fd767_19 defaults libuuid 2.32.1 h7f98852_1000 defaults libzlib 1.2.13 h166bdaf_4 defaults markupsafe 2.1.2 py311h2582759_0 defaults matplotlib-inline 0.1.6 pyhd8ed1ab_0 defaults mistune 2.0.4 pyhd8ed1ab_0 defaults nbclassic 0.4.8 pyhd8ed1ab_0 defaults nbclient 0.7.2 pyhd8ed1ab_0 defaults nbconvert 7.2.8 pyhd8ed1ab_0 defaults nbconvert-core 7.2.8 pyhd8ed1ab_0 defaults nbconvert-pandoc 7.2.8 pyhd8ed1ab_0 defaults nbformat 5.7.3 pyhd8ed1ab_0 defaults ncurses 6.3 h27087fc_1 defaults nest-asyncio 1.5.6 pyhd8ed1ab_0 defaults notebook 6.5.2 pyha770c72_1 defaults notebook-shim 0.2.2 pyhd8ed1ab_0 defaults openssl 3.0.7 h0b41bf4_1 defaults packaging 23.0 pyhd8ed1ab_0 defaults pandoc 2.19.2 h32600fe_1 defaults pandocfilters 1.5.0 pyhd8ed1ab_0 defaults parso 0.8.3 pyhd8ed1ab_0 defaults pexpect 4.8.0 pyh1a96a4e_2 defaults pickleshare 0.7.5 py_1003 defaults pip 22.3.1 pyhd8ed1ab_0 defaults pkgutil-resolve-name 1.3.10 pyhd8ed1ab_0 defaults platformdirs 2.6.2 pyhd8ed1ab_0 defaults prometheus_client 0.15.0 pyhd8ed1ab_0 defaults prompt-toolkit 3.0.36 pyha770c72_0 defaults psutil 5.9.4 py311hd4cff14_0 defaults ptyprocess 0.7.0 pyhd3deb0d_0 defaults pure_eval 0.2.2 pyhd8ed1ab_0 defaults pycparser 2.21 pyhd8ed1ab_0 defaults pygments 2.14.0 pyhd8ed1ab_0 defaults pyopenssl 23.0.0 pyhd8ed1ab_0 defaults pyrsistent 0.19.3 py311h2582759_0 defaults pysocks 1.7.1 pyha2e5f31_6 defaults python 3.11.0 he550d4f_1_cpython defaults python-dateutil 2.8.2 pyhd8ed1ab_0 defaults python-fastjsonschema 2.16.2 pyhd8ed1ab_0 defaults python-json-logger 2.0.4 pyhd8ed1ab_0 defaults python_abi 3.11 3_cp311 defaults pytz 2022.7.1 pyhd8ed1ab_0 defaults pyyaml 6.0 py311hd4cff14_5 defaults pyzmq 25.0.0 py311hd6ccaeb_0 defaults readline 8.1.2 h0f457ee_0 defaults requests 2.28.2 pyhd8ed1ab_0 defaults rfc3339-validator 0.1.4 pyhd8ed1ab_0 defaults rfc3986-validator 0.1.1 pyh9f0ad1d_0 defaults send2trash 1.8.0 pyhd8ed1ab_0 defaults setuptools 66.1.1 pyhd8ed1ab_0 defaults six 1.16.0 pyh6c4a22f_0 defaults sniffio 1.3.0 pyhd8ed1ab_0 defaults soupsieve 2.3.2.post1 pyhd8ed1ab_0 defaults stack_data 0.6.2 pyhd8ed1ab_0 defaults terminado 0.17.1 pyh41d4057_0 defaults tinycss2 1.2.1 pyhd8ed1ab_0 defaults tk 8.6.12 h27826a3_0 defaults tomli 2.0.1 pyhd8ed1ab_0 defaults tornado 6.2 py311hd4cff14_1 defaults traitlets 5.8.1 pyhd8ed1ab_0 defaults typing-extensions 4.4.0 hd8ed1ab_0 defaults typing_extensions 4.4.0 pyha770c72_0 defaults tzdata 2022g h191b570_0 defaults urllib3 1.26.14 pyhd8ed1ab_0 defaults wcwidth 0.2.6 pyhd8ed1ab_0 defaults webencodings 0.5.1 py_1 defaults websocket-client 1.4.2 pyhd8ed1ab_0 defaults wheel 0.38.4 pyhd8ed1ab_0 defaults xz 5.2.6 h166bdaf_0 defaults yaml 0.2.5 h7f98852_2 defaults zeromq 4.3.4 h9c3ff4c_1 defaults zipp 3.11.0 pyhd8ed1ab_0 defaults conda env: name: jlab-test channels: - defaults dependencies: - _libgcc_mutex=0.1=conda_forge - _openmp_mutex=4.5=2_gnu - anyio=3.6.2=pyhd8ed1ab_0 - argon2-cffi=21.3.0=pyhd8ed1ab_0 - argon2-cffi-bindings=21.2.0=py311hd4cff14_3 - asttokens=2.2.1=pyhd8ed1ab_0 - attrs=22.2.0=pyh71513ae_0 - babel=2.11.0=pyhd8ed1ab_0 - backcall=0.2.0=pyh9f0ad1d_0 - backports=1.0=pyhd8ed1ab_3 - backports.functools_lru_cache=1.6.4=pyhd8ed1ab_0 - beautifulsoup4=4.11.1=pyha770c72_0 - bleach=5.0.1=pyhd8ed1ab_0 - brotlipy=0.7.0=py311hd4cff14_1005 - bzip2=1.0.8=h7f98852_4 - ca-certificates=2022.12.7=ha878542_0 - certifi=2022.12.7=pyhd8ed1ab_0 - cffi=1.15.1=py311h409f033_3 - charset-normalizer=2.1.1=pyhd8ed1ab_0 - comm=0.1.2=pyhd8ed1ab_0 - cryptography=39.0.0=py311h9b4c7bb_0 - debugpy=1.6.5=py311hcafe171_0 - decorator=5.1.1=pyhd8ed1ab_0 - defusedxml=0.7.1=pyhd8ed1ab_0 - entrypoints=0.4=pyhd8ed1ab_0 - executing=1.2.0=pyhd8ed1ab_0 - flit-core=3.8.0=pyhd8ed1ab_0 - idna=3.4=pyhd8ed1ab_0 - importlib-metadata=6.0.0=pyha770c72_0 - importlib_resources=5.10.2=pyhd8ed1ab_0 - ipykernel=6.20.2=pyh210e3f2_0 - ipython=8.8.0=pyh41d4057_0 - ipython_genutils=0.2.0=py_1 - jedi=0.18.2=pyhd8ed1ab_0 - jinja2=3.1.2=pyhd8ed1ab_1 - json5=0.9.5=pyh9f0ad1d_0 - jsonschema=4.17.3=pyhd8ed1ab_0 - jupyter_client=7.4.9=pyhd8ed1ab_0 - jupyter_core=5.1.3=py311h38be061_0 - jupyter_events=0.6.3=pyhd8ed1ab_0 - jupyter_server=2.1.0=pyhd8ed1ab_0 - jupyter_server_terminals=0.4.4=pyhd8ed1ab_1 - jupyterlab=3.5.2=pyhd8ed1ab_0 - jupyterlab_pygments=0.2.2=pyhd8ed1ab_0 - jupyterlab_server=2.19.0=pyhd8ed1ab_0 - ld_impl_linux-64=2.39=hcc3a1bd_1 - libffi=3.4.2=h7f98852_5 - libgcc-ng=12.2.0=h65d4601_19 - libgomp=12.2.0=h65d4601_19 - libnsl=2.0.0=h7f98852_0 - libsodium=1.0.18=h36c2ea0_1 - libsqlite=3.40.0=h753d276_0 - libstdcxx-ng=12.2.0=h46fd767_19 - libuuid=2.32.1=h7f98852_1000 - libzlib=1.2.13=h166bdaf_4 - markupsafe=2.1.2=py311h2582759_0 - matplotlib-inline=0.1.6=pyhd8ed1ab_0 - mistune=2.0.4=pyhd8ed1ab_0 - nbclassic=0.4.8=pyhd8ed1ab_0 - nbclient=0.7.2=pyhd8ed1ab_0 - nbconvert=7.2.8=pyhd8ed1ab_0 - nbconvert-core=7.2.8=pyhd8ed1ab_0 - nbconvert-pandoc=7.2.8=pyhd8ed1ab_0 - nbformat=5.7.3=pyhd8ed1ab_0 - ncurses=6.3=h27087fc_1 - nest-asyncio=1.5.6=pyhd8ed1ab_0 - notebook=6.5.2=pyha770c72_1 - notebook-shim=0.2.2=pyhd8ed1ab_0 - openssl=3.0.7=h0b41bf4_1 - packaging=23.0=pyhd8ed1ab_0 - pandoc=2.19.2=h32600fe_1 - pandocfilters=1.5.0=pyhd8ed1ab_0 - parso=0.8.3=pyhd8ed1ab_0 - pexpect=4.8.0=pyh1a96a4e_2 - pickleshare=0.7.5=py_1003 - pip=22.3.1=pyhd8ed1ab_0 - pkgutil-resolve-name=1.3.10=pyhd8ed1ab_0 - platformdirs=2.6.2=pyhd8ed1ab_0 - prometheus_client=0.15.0=pyhd8ed1ab_0 - prompt-toolkit=3.0.36=pyha770c72_0 - psutil=5.9.4=py311hd4cff14_0 - ptyprocess=0.7.0=pyhd3deb0d_0 - pure_eval=0.2.2=pyhd8ed1ab_0 - pycparser=2.21=pyhd8ed1ab_0 - pygments=2.14.0=pyhd8ed1ab_0 - pyopenssl=23.0.0=pyhd8ed1ab_0 - pyrsistent=0.19.3=py311h2582759_0 - pysocks=1.7.1=pyha2e5f31_6 - python=3.11.0=he550d4f_1_cpython - python-dateutil=2.8.2=pyhd8ed1ab_0 - python-fastjsonschema=2.16.2=pyhd8ed1ab_0 - python-json-logger=2.0.4=pyhd8ed1ab_0 - python_abi=3.11=3_cp311 - pytz=2022.7.1=pyhd8ed1ab_0 - pyyaml=6.0=py311hd4cff14_5 - pyzmq=25.0.0=py311hd6ccaeb_0 - readline=8.1.2=h0f457ee_0 - requests=2.28.2=pyhd8ed1ab_0 - rfc3339-validator=0.1.4=pyhd8ed1ab_0 - rfc3986-validator=0.1.1=pyh9f0ad1d_0 - send2trash=1.8.0=pyhd8ed1ab_0 - setuptools=66.1.1=pyhd8ed1ab_0 - six=1.16.0=pyh6c4a22f_0 - sniffio=1.3.0=pyhd8ed1ab_0 - soupsieve=2.3.2.post1=pyhd8ed1ab_0 - stack_data=0.6.2=pyhd8ed1ab_0 - terminado=0.17.1=pyh41d4057_0 - tinycss2=1.2.1=pyhd8ed1ab_0 - tk=8.6.12=h27826a3_0 - tomli=2.0.1=pyhd8ed1ab_0 - tornado=6.2=py311hd4cff14_1 - traitlets=5.8.1=pyhd8ed1ab_0 - typing-extensions=4.4.0=hd8ed1ab_0 - typing_extensions=4.4.0=pyha770c72_0 - tzdata=2022g=h191b570_0 - urllib3=1.26.14=pyhd8ed1ab_0 - wcwidth=0.2.6=pyhd8ed1ab_0 - webencodings=0.5.1=py_1 - websocket-client=1.4.2=pyhd8ed1ab_0 - wheel=0.38.4=pyhd8ed1ab_0 - xz=5.2.6=h166bdaf_0 - yaml=0.2.5=h7f98852_2 - zeromq=4.3.4=h9c3ff4c_1 - zipp=3.11.0=pyhd8ed1ab_0 prefix: /home/kungurov/.conda/envs/jlab-test </pre> </details> I think it's because getExportFormats() called twice in the code (see below). index.js, line 340: ``` void services.nbconvert.getExportFormats().then(response => { if (response) { const formatLabels = Private.getFormatLabels(translato); // Convert export list to palette and menu items. ``` index.js, line 510: ``` void services.nbconvert.getExportFormats().then(response => { if (response) { /** ... ... ... */ const rawFormatExclude = [ ```
True
JupyterLab requests /api/nbconvert twice during load - ## Description JupyterLab during webpage load performs requests to /api/nbconvert twice. ## Reproduce 1. Open Chrome DevTools, switch to Network tab 2. Reload the page with JupyterLab 3. Find /api/nbconvert requests (you could filter by 'nbconvert') 4. See there are two simultanous requests ## Expected behavior I probably would expect one request not two. The /api/nbconvert response time can be relatively slow (is quite slow in my full, non 'jlab-test' environment for some reason). ## Context - Operating System and version: CentOS - Browser and version: Edge 108.0.1462 - JupyterLab version: 3.5.2 <details><summary>Troubleshoot Output</summary> <pre> $ jupyter troubleshoot $PATH: /home/kungurov/.conda/envs/jlab-test/bin /usr/frog/scicomp/pythonds/conda/condabin /usr/frog/scicomp/pythonds/v1.2/bin /usr/frog/scicomp/pythonds/latex/bin/x86_64-linux /usr/frog/scicomp/pythonds/conda/condabin /usr/local/bin /usr/bin /usr/local/sbin /usr/sbin /home/kungurov/bin sys.path: /home/kungurov/.conda/envs/jlab-test/bin /home/kungurov/.conda/envs/jlab-test/lib/python311.zip /home/kungurov/.conda/envs/jlab-test/lib/python3.11 /home/kungurov/.conda/envs/jlab-test/lib/python3.11/lib-dynload /home/kungurov/.conda/envs/jlab-test/lib/python3.11/site-packages sys.executable: /home/kungurov/.conda/envs/jlab-test/bin/python sys.version: 3.11.0 | packaged by conda-forge | (main, Jan 14 2023, 12:27:40) [GCC 11.3.0] platform.platform(): Linux-3.10.0-1160.76.1.el7.x86_64-x86_64-with-glibc2.17 which -a jupyter: /home/kungurov/.conda/envs/jlab-test/bin/jupyter pip list: Package Version ----------------------------- ----------- anyio 3.6.2 argon2-cffi 21.3.0 argon2-cffi-bindings 21.2.0 asttokens 2.2.1 attrs 22.2.0 Babel 2.11.0 backcall 0.2.0 backports.functools-lru-cache 1.6.4 beautifulsoup4 4.11.1 bleach 5.0.1 brotlipy 0.7.0 certifi 2022.12.7 cffi 1.15.1 charset-normalizer 2.1.1 comm 0.1.2 cryptography 39.0.0 debugpy 1.6.5 decorator 5.1.1 defusedxml 0.7.1 entrypoints 0.4 executing 1.2.0 fastjsonschema 2.16.2 flit_core 3.8.0 idna 3.4 importlib-metadata 6.0.0 importlib-resources 5.10.2 ipykernel 6.20.2 ipython 8.8.0 ipython-genutils 0.2.0 jedi 0.18.2 Jinja2 3.1.2 json5 0.9.5 jsonschema 4.17.3 jupyter_client 7.4.9 jupyter_core 5.1.3 jupyter-events 0.6.3 jupyter_server 2.1.0 jupyter_server_terminals 0.4.4 jupyterlab 3.5.2 jupyterlab-pygments 0.2.2 jupyterlab_server 2.19.0 MarkupSafe 2.1.2 matplotlib-inline 0.1.6 mistune 2.0.4 nbclassic 0.4.8 nbclient 0.7.2 nbconvert 7.2.8 nbformat 5.7.3 nest-asyncio 1.5.6 notebook 6.5.2 notebook_shim 0.2.2 packaging 23.0 pandocfilters 1.5.0 parso 0.8.3 pexpect 4.8.0 pickleshare 0.7.5 pip 22.3.1 pkgutil_resolve_name 1.3.10 platformdirs 2.6.2 prometheus-client 0.15.0 prompt-toolkit 3.0.36 psutil 5.9.4 ptyprocess 0.7.0 pure-eval 0.2.2 pycparser 2.21 Pygments 2.14.0 pyOpenSSL 23.0.0 pyrsistent 0.19.3 PySocks 1.7.1 python-dateutil 2.8.2 python-json-logger 2.0.4 pytz 2022.7.1 PyYAML 6.0 pyzmq 25.0.0 requests 2.28.2 rfc3339-validator 0.1.4 rfc3986-validator 0.1.1 Send2Trash 1.8.0 setuptools 66.1.1 six 1.16.0 sniffio 1.3.0 soupsieve 2.3.2.post1 stack-data 0.6.2 terminado 0.17.1 tinycss2 1.2.1 tomli 2.0.1 tornado 6.2 traitlets 5.8.1 typing_extensions 4.4.0 urllib3 1.26.14 wcwidth 0.2.6 webencodings 0.5.1 websocket-client 1.4.2 wheel 0.38.4 zipp 3.11.0 conda list: # packages in environment at /home/kungurov/.conda/envs/jlab-test: # # Name Version Build Channel _libgcc_mutex 0.1 conda_forge defaults _openmp_mutex 4.5 2_gnu defaults anyio 3.6.2 pyhd8ed1ab_0 defaults argon2-cffi 21.3.0 pyhd8ed1ab_0 defaults argon2-cffi-bindings 21.2.0 py311hd4cff14_3 defaults asttokens 2.2.1 pyhd8ed1ab_0 defaults attrs 22.2.0 pyh71513ae_0 defaults babel 2.11.0 pyhd8ed1ab_0 defaults backcall 0.2.0 pyh9f0ad1d_0 defaults backports 1.0 pyhd8ed1ab_3 defaults backports.functools_lru_cache 1.6.4 pyhd8ed1ab_0 defaults beautifulsoup4 4.11.1 pyha770c72_0 defaults bleach 5.0.1 pyhd8ed1ab_0 defaults brotlipy 0.7.0 py311hd4cff14_1005 defaults bzip2 1.0.8 h7f98852_4 defaults ca-certificates 2022.12.7 ha878542_0 defaults certifi 2022.12.7 pyhd8ed1ab_0 defaults cffi 1.15.1 py311h409f033_3 defaults charset-normalizer 2.1.1 pyhd8ed1ab_0 defaults comm 0.1.2 pyhd8ed1ab_0 defaults cryptography 39.0.0 py311h9b4c7bb_0 defaults debugpy 1.6.5 py311hcafe171_0 defaults decorator 5.1.1 pyhd8ed1ab_0 defaults defusedxml 0.7.1 pyhd8ed1ab_0 defaults entrypoints 0.4 pyhd8ed1ab_0 defaults executing 1.2.0 pyhd8ed1ab_0 defaults flit-core 3.8.0 pyhd8ed1ab_0 defaults idna 3.4 pyhd8ed1ab_0 defaults importlib-metadata 6.0.0 pyha770c72_0 defaults importlib_resources 5.10.2 pyhd8ed1ab_0 defaults ipykernel 6.20.2 pyh210e3f2_0 defaults ipython 8.8.0 pyh41d4057_0 defaults ipython_genutils 0.2.0 py_1 defaults jedi 0.18.2 pyhd8ed1ab_0 defaults jinja2 3.1.2 pyhd8ed1ab_1 defaults json5 0.9.5 pyh9f0ad1d_0 defaults jsonschema 4.17.3 pyhd8ed1ab_0 defaults jupyter_client 7.4.9 pyhd8ed1ab_0 defaults jupyter_core 5.1.3 py311h38be061_0 defaults jupyter_events 0.6.3 pyhd8ed1ab_0 defaults jupyter_server 2.1.0 pyhd8ed1ab_0 defaults jupyter_server_terminals 0.4.4 pyhd8ed1ab_1 defaults jupyterlab 3.5.2 pyhd8ed1ab_0 defaults jupyterlab_pygments 0.2.2 pyhd8ed1ab_0 defaults jupyterlab_server 2.19.0 pyhd8ed1ab_0 defaults ld_impl_linux-64 2.39 hcc3a1bd_1 defaults libffi 3.4.2 h7f98852_5 defaults libgcc-ng 12.2.0 h65d4601_19 defaults libgomp 12.2.0 h65d4601_19 defaults libnsl 2.0.0 h7f98852_0 defaults libsodium 1.0.18 h36c2ea0_1 defaults libsqlite 3.40.0 h753d276_0 defaults libstdcxx-ng 12.2.0 h46fd767_19 defaults libuuid 2.32.1 h7f98852_1000 defaults libzlib 1.2.13 h166bdaf_4 defaults markupsafe 2.1.2 py311h2582759_0 defaults matplotlib-inline 0.1.6 pyhd8ed1ab_0 defaults mistune 2.0.4 pyhd8ed1ab_0 defaults nbclassic 0.4.8 pyhd8ed1ab_0 defaults nbclient 0.7.2 pyhd8ed1ab_0 defaults nbconvert 7.2.8 pyhd8ed1ab_0 defaults nbconvert-core 7.2.8 pyhd8ed1ab_0 defaults nbconvert-pandoc 7.2.8 pyhd8ed1ab_0 defaults nbformat 5.7.3 pyhd8ed1ab_0 defaults ncurses 6.3 h27087fc_1 defaults nest-asyncio 1.5.6 pyhd8ed1ab_0 defaults notebook 6.5.2 pyha770c72_1 defaults notebook-shim 0.2.2 pyhd8ed1ab_0 defaults openssl 3.0.7 h0b41bf4_1 defaults packaging 23.0 pyhd8ed1ab_0 defaults pandoc 2.19.2 h32600fe_1 defaults pandocfilters 1.5.0 pyhd8ed1ab_0 defaults parso 0.8.3 pyhd8ed1ab_0 defaults pexpect 4.8.0 pyh1a96a4e_2 defaults pickleshare 0.7.5 py_1003 defaults pip 22.3.1 pyhd8ed1ab_0 defaults pkgutil-resolve-name 1.3.10 pyhd8ed1ab_0 defaults platformdirs 2.6.2 pyhd8ed1ab_0 defaults prometheus_client 0.15.0 pyhd8ed1ab_0 defaults prompt-toolkit 3.0.36 pyha770c72_0 defaults psutil 5.9.4 py311hd4cff14_0 defaults ptyprocess 0.7.0 pyhd3deb0d_0 defaults pure_eval 0.2.2 pyhd8ed1ab_0 defaults pycparser 2.21 pyhd8ed1ab_0 defaults pygments 2.14.0 pyhd8ed1ab_0 defaults pyopenssl 23.0.0 pyhd8ed1ab_0 defaults pyrsistent 0.19.3 py311h2582759_0 defaults pysocks 1.7.1 pyha2e5f31_6 defaults python 3.11.0 he550d4f_1_cpython defaults python-dateutil 2.8.2 pyhd8ed1ab_0 defaults python-fastjsonschema 2.16.2 pyhd8ed1ab_0 defaults python-json-logger 2.0.4 pyhd8ed1ab_0 defaults python_abi 3.11 3_cp311 defaults pytz 2022.7.1 pyhd8ed1ab_0 defaults pyyaml 6.0 py311hd4cff14_5 defaults pyzmq 25.0.0 py311hd6ccaeb_0 defaults readline 8.1.2 h0f457ee_0 defaults requests 2.28.2 pyhd8ed1ab_0 defaults rfc3339-validator 0.1.4 pyhd8ed1ab_0 defaults rfc3986-validator 0.1.1 pyh9f0ad1d_0 defaults send2trash 1.8.0 pyhd8ed1ab_0 defaults setuptools 66.1.1 pyhd8ed1ab_0 defaults six 1.16.0 pyh6c4a22f_0 defaults sniffio 1.3.0 pyhd8ed1ab_0 defaults soupsieve 2.3.2.post1 pyhd8ed1ab_0 defaults stack_data 0.6.2 pyhd8ed1ab_0 defaults terminado 0.17.1 pyh41d4057_0 defaults tinycss2 1.2.1 pyhd8ed1ab_0 defaults tk 8.6.12 h27826a3_0 defaults tomli 2.0.1 pyhd8ed1ab_0 defaults tornado 6.2 py311hd4cff14_1 defaults traitlets 5.8.1 pyhd8ed1ab_0 defaults typing-extensions 4.4.0 hd8ed1ab_0 defaults typing_extensions 4.4.0 pyha770c72_0 defaults tzdata 2022g h191b570_0 defaults urllib3 1.26.14 pyhd8ed1ab_0 defaults wcwidth 0.2.6 pyhd8ed1ab_0 defaults webencodings 0.5.1 py_1 defaults websocket-client 1.4.2 pyhd8ed1ab_0 defaults wheel 0.38.4 pyhd8ed1ab_0 defaults xz 5.2.6 h166bdaf_0 defaults yaml 0.2.5 h7f98852_2 defaults zeromq 4.3.4 h9c3ff4c_1 defaults zipp 3.11.0 pyhd8ed1ab_0 defaults conda env: name: jlab-test channels: - defaults dependencies: - _libgcc_mutex=0.1=conda_forge - _openmp_mutex=4.5=2_gnu - anyio=3.6.2=pyhd8ed1ab_0 - argon2-cffi=21.3.0=pyhd8ed1ab_0 - argon2-cffi-bindings=21.2.0=py311hd4cff14_3 - asttokens=2.2.1=pyhd8ed1ab_0 - attrs=22.2.0=pyh71513ae_0 - babel=2.11.0=pyhd8ed1ab_0 - backcall=0.2.0=pyh9f0ad1d_0 - backports=1.0=pyhd8ed1ab_3 - backports.functools_lru_cache=1.6.4=pyhd8ed1ab_0 - beautifulsoup4=4.11.1=pyha770c72_0 - bleach=5.0.1=pyhd8ed1ab_0 - brotlipy=0.7.0=py311hd4cff14_1005 - bzip2=1.0.8=h7f98852_4 - ca-certificates=2022.12.7=ha878542_0 - certifi=2022.12.7=pyhd8ed1ab_0 - cffi=1.15.1=py311h409f033_3 - charset-normalizer=2.1.1=pyhd8ed1ab_0 - comm=0.1.2=pyhd8ed1ab_0 - cryptography=39.0.0=py311h9b4c7bb_0 - debugpy=1.6.5=py311hcafe171_0 - decorator=5.1.1=pyhd8ed1ab_0 - defusedxml=0.7.1=pyhd8ed1ab_0 - entrypoints=0.4=pyhd8ed1ab_0 - executing=1.2.0=pyhd8ed1ab_0 - flit-core=3.8.0=pyhd8ed1ab_0 - idna=3.4=pyhd8ed1ab_0 - importlib-metadata=6.0.0=pyha770c72_0 - importlib_resources=5.10.2=pyhd8ed1ab_0 - ipykernel=6.20.2=pyh210e3f2_0 - ipython=8.8.0=pyh41d4057_0 - ipython_genutils=0.2.0=py_1 - jedi=0.18.2=pyhd8ed1ab_0 - jinja2=3.1.2=pyhd8ed1ab_1 - json5=0.9.5=pyh9f0ad1d_0 - jsonschema=4.17.3=pyhd8ed1ab_0 - jupyter_client=7.4.9=pyhd8ed1ab_0 - jupyter_core=5.1.3=py311h38be061_0 - jupyter_events=0.6.3=pyhd8ed1ab_0 - jupyter_server=2.1.0=pyhd8ed1ab_0 - jupyter_server_terminals=0.4.4=pyhd8ed1ab_1 - jupyterlab=3.5.2=pyhd8ed1ab_0 - jupyterlab_pygments=0.2.2=pyhd8ed1ab_0 - jupyterlab_server=2.19.0=pyhd8ed1ab_0 - ld_impl_linux-64=2.39=hcc3a1bd_1 - libffi=3.4.2=h7f98852_5 - libgcc-ng=12.2.0=h65d4601_19 - libgomp=12.2.0=h65d4601_19 - libnsl=2.0.0=h7f98852_0 - libsodium=1.0.18=h36c2ea0_1 - libsqlite=3.40.0=h753d276_0 - libstdcxx-ng=12.2.0=h46fd767_19 - libuuid=2.32.1=h7f98852_1000 - libzlib=1.2.13=h166bdaf_4 - markupsafe=2.1.2=py311h2582759_0 - matplotlib-inline=0.1.6=pyhd8ed1ab_0 - mistune=2.0.4=pyhd8ed1ab_0 - nbclassic=0.4.8=pyhd8ed1ab_0 - nbclient=0.7.2=pyhd8ed1ab_0 - nbconvert=7.2.8=pyhd8ed1ab_0 - nbconvert-core=7.2.8=pyhd8ed1ab_0 - nbconvert-pandoc=7.2.8=pyhd8ed1ab_0 - nbformat=5.7.3=pyhd8ed1ab_0 - ncurses=6.3=h27087fc_1 - nest-asyncio=1.5.6=pyhd8ed1ab_0 - notebook=6.5.2=pyha770c72_1 - notebook-shim=0.2.2=pyhd8ed1ab_0 - openssl=3.0.7=h0b41bf4_1 - packaging=23.0=pyhd8ed1ab_0 - pandoc=2.19.2=h32600fe_1 - pandocfilters=1.5.0=pyhd8ed1ab_0 - parso=0.8.3=pyhd8ed1ab_0 - pexpect=4.8.0=pyh1a96a4e_2 - pickleshare=0.7.5=py_1003 - pip=22.3.1=pyhd8ed1ab_0 - pkgutil-resolve-name=1.3.10=pyhd8ed1ab_0 - platformdirs=2.6.2=pyhd8ed1ab_0 - prometheus_client=0.15.0=pyhd8ed1ab_0 - prompt-toolkit=3.0.36=pyha770c72_0 - psutil=5.9.4=py311hd4cff14_0 - ptyprocess=0.7.0=pyhd3deb0d_0 - pure_eval=0.2.2=pyhd8ed1ab_0 - pycparser=2.21=pyhd8ed1ab_0 - pygments=2.14.0=pyhd8ed1ab_0 - pyopenssl=23.0.0=pyhd8ed1ab_0 - pyrsistent=0.19.3=py311h2582759_0 - pysocks=1.7.1=pyha2e5f31_6 - python=3.11.0=he550d4f_1_cpython - python-dateutil=2.8.2=pyhd8ed1ab_0 - python-fastjsonschema=2.16.2=pyhd8ed1ab_0 - python-json-logger=2.0.4=pyhd8ed1ab_0 - python_abi=3.11=3_cp311 - pytz=2022.7.1=pyhd8ed1ab_0 - pyyaml=6.0=py311hd4cff14_5 - pyzmq=25.0.0=py311hd6ccaeb_0 - readline=8.1.2=h0f457ee_0 - requests=2.28.2=pyhd8ed1ab_0 - rfc3339-validator=0.1.4=pyhd8ed1ab_0 - rfc3986-validator=0.1.1=pyh9f0ad1d_0 - send2trash=1.8.0=pyhd8ed1ab_0 - setuptools=66.1.1=pyhd8ed1ab_0 - six=1.16.0=pyh6c4a22f_0 - sniffio=1.3.0=pyhd8ed1ab_0 - soupsieve=2.3.2.post1=pyhd8ed1ab_0 - stack_data=0.6.2=pyhd8ed1ab_0 - terminado=0.17.1=pyh41d4057_0 - tinycss2=1.2.1=pyhd8ed1ab_0 - tk=8.6.12=h27826a3_0 - tomli=2.0.1=pyhd8ed1ab_0 - tornado=6.2=py311hd4cff14_1 - traitlets=5.8.1=pyhd8ed1ab_0 - typing-extensions=4.4.0=hd8ed1ab_0 - typing_extensions=4.4.0=pyha770c72_0 - tzdata=2022g=h191b570_0 - urllib3=1.26.14=pyhd8ed1ab_0 - wcwidth=0.2.6=pyhd8ed1ab_0 - webencodings=0.5.1=py_1 - websocket-client=1.4.2=pyhd8ed1ab_0 - wheel=0.38.4=pyhd8ed1ab_0 - xz=5.2.6=h166bdaf_0 - yaml=0.2.5=h7f98852_2 - zeromq=4.3.4=h9c3ff4c_1 - zipp=3.11.0=pyhd8ed1ab_0 prefix: /home/kungurov/.conda/envs/jlab-test </pre> </details> I think it's because getExportFormats() called twice in the code (see below). index.js, line 340: ``` void services.nbconvert.getExportFormats().then(response => { if (response) { const formatLabels = Private.getFormatLabels(translato); // Convert export list to palette and menu items. ``` index.js, line 510: ``` void services.nbconvert.getExportFormats().then(response => { if (response) { /** ... ... ... */ const rawFormatExclude = [ ```
perf
jupyterlab requests api nbconvert twice during load description jupyterlab during webpage load performs requests to api nbconvert twice reproduce open chrome devtools switch to network tab reload the page with jupyterlab find api nbconvert requests you could filter by nbconvert see there are two simultanous requests expected behavior i probably would expect one request not two the api nbconvert response time can be relatively slow is quite slow in my full non jlab test environment for some reason context operating system and version centos browser and version edge jupyterlab version troubleshoot output jupyter troubleshoot path home kungurov conda envs jlab test bin usr frog scicomp pythonds conda condabin usr frog scicomp pythonds bin usr frog scicomp pythonds latex bin linux usr frog scicomp pythonds conda condabin usr local bin usr bin usr local sbin usr sbin home kungurov bin sys path home kungurov conda envs jlab test bin home kungurov conda envs jlab test lib zip home kungurov conda envs jlab test lib home kungurov conda envs jlab test lib lib dynload home kungurov conda envs jlab test lib site packages sys executable home kungurov conda envs jlab test bin python sys version packaged by conda forge main jan platform platform linux with which a jupyter home kungurov conda envs jlab test bin jupyter pip list package version anyio cffi cffi bindings asttokens attrs babel backcall backports functools lru cache bleach brotlipy certifi cffi charset normalizer comm cryptography debugpy decorator defusedxml entrypoints executing fastjsonschema flit core idna importlib metadata importlib resources ipykernel ipython ipython genutils jedi jsonschema jupyter client jupyter core jupyter events jupyter server jupyter server terminals jupyterlab jupyterlab pygments jupyterlab server markupsafe matplotlib inline mistune nbclassic nbclient nbconvert nbformat nest asyncio notebook notebook shim packaging pandocfilters parso pexpect pickleshare pip pkgutil resolve name platformdirs prometheus client prompt toolkit psutil ptyprocess pure eval pycparser pygments pyopenssl pyrsistent pysocks python dateutil python json logger pytz pyyaml pyzmq requests validator validator setuptools six sniffio soupsieve stack data terminado tomli tornado traitlets typing extensions wcwidth webencodings websocket client wheel zipp conda list packages in environment at home kungurov conda envs jlab test name version build channel libgcc mutex conda forge defaults openmp mutex gnu defaults anyio defaults cffi defaults cffi bindings defaults asttokens defaults attrs defaults babel defaults backcall defaults backports defaults backports functools lru cache defaults defaults bleach defaults brotlipy defaults defaults ca certificates defaults certifi defaults cffi defaults charset normalizer defaults comm defaults cryptography defaults debugpy defaults decorator defaults defusedxml defaults entrypoints defaults executing defaults flit core defaults idna defaults importlib metadata defaults importlib resources defaults ipykernel defaults ipython defaults ipython genutils py defaults jedi defaults defaults defaults jsonschema defaults jupyter client defaults jupyter core defaults jupyter events defaults jupyter server defaults jupyter server terminals defaults jupyterlab defaults jupyterlab pygments defaults jupyterlab server defaults ld impl linux defaults libffi defaults libgcc ng defaults libgomp defaults libnsl defaults libsodium defaults libsqlite defaults libstdcxx ng defaults libuuid defaults libzlib defaults markupsafe defaults matplotlib inline defaults mistune defaults nbclassic defaults nbclient defaults nbconvert defaults nbconvert core defaults nbconvert pandoc defaults nbformat defaults ncurses defaults nest asyncio defaults notebook defaults notebook shim defaults openssl defaults packaging defaults pandoc defaults pandocfilters defaults parso defaults pexpect defaults pickleshare py defaults pip defaults pkgutil resolve name defaults platformdirs defaults prometheus client defaults prompt toolkit defaults psutil defaults ptyprocess defaults pure eval defaults pycparser defaults pygments defaults pyopenssl defaults pyrsistent defaults pysocks defaults python cpython defaults python dateutil defaults python fastjsonschema defaults python json logger defaults python abi defaults pytz defaults pyyaml defaults pyzmq defaults readline defaults requests defaults validator defaults validator defaults defaults setuptools defaults six defaults sniffio defaults soupsieve defaults stack data defaults terminado defaults defaults tk defaults tomli defaults tornado defaults traitlets defaults typing extensions defaults typing extensions defaults tzdata defaults defaults wcwidth defaults webencodings py defaults websocket client defaults wheel defaults xz defaults yaml defaults zeromq defaults zipp defaults conda env name jlab test channels defaults dependencies libgcc mutex conda forge openmp mutex gnu anyio cffi cffi bindings asttokens attrs babel backcall backports backports functools lru cache bleach brotlipy ca certificates certifi cffi charset normalizer comm cryptography debugpy decorator defusedxml entrypoints executing flit core idna importlib metadata importlib resources ipykernel ipython ipython genutils py jedi jsonschema jupyter client jupyter core jupyter events jupyter server jupyter server terminals jupyterlab jupyterlab pygments jupyterlab server ld impl linux libffi libgcc ng libgomp libnsl libsodium libsqlite libstdcxx ng libuuid libzlib markupsafe matplotlib inline mistune nbclassic nbclient nbconvert nbconvert core nbconvert pandoc nbformat ncurses nest asyncio notebook notebook shim openssl packaging pandoc pandocfilters parso pexpect pickleshare py pip pkgutil resolve name platformdirs prometheus client prompt toolkit psutil ptyprocess pure eval pycparser pygments pyopenssl pyrsistent pysocks python cpython python dateutil python fastjsonschema python json logger python abi pytz pyyaml pyzmq readline requests validator validator setuptools six sniffio soupsieve stack data terminado tk tomli tornado traitlets typing extensions typing extensions tzdata wcwidth webencodings py websocket client wheel xz yaml zeromq zipp prefix home kungurov conda envs jlab test i think it s because getexportformats called twice in the code see below index js line void services nbconvert getexportformats then response if response const formatlabels private getformatlabels translato convert export list to palette and menu items index js line void services nbconvert getexportformats then response if response const rawformatexclude
1
121,899
16,050,131,277
IssuesEvent
2021-04-22 18:04:15
t10d/kingdom-python-server
https://api.github.com/repos/t10d/kingdom-python-server
opened
Adapt authorization module to work with current entrypoints
crucial design change refactor
Analogous to #20, but for entrypoint integrations. ## Goal **One liner**: Integrate authorization module services to current implementations of middleware & directives. 1. Adapt authentication middleware and context propagation 2. Adapt authorization middleware and proper scope results handling on query resolvers 3. Proper error handling
1.0
Adapt authorization module to work with current entrypoints - Analogous to #20, but for entrypoint integrations. ## Goal **One liner**: Integrate authorization module services to current implementations of middleware & directives. 1. Adapt authentication middleware and context propagation 2. Adapt authorization middleware and proper scope results handling on query resolvers 3. Proper error handling
non_perf
adapt authorization module to work with current entrypoints analogous to but for entrypoint integrations goal one liner integrate authorization module services to current implementations of middleware directives adapt authentication middleware and context propagation adapt authorization middleware and proper scope results handling on query resolvers proper error handling
0
308,979
9,459,738,633
IssuesEvent
2019-04-17 09:14:30
WallarooLabs/wallaroo
https://api.github.com/repos/WallarooLabs/wallaroo
closed
Add ability to migrate a source to a different worker.
enhancement priority: low
Currently a source cannot be moved from a worker. In order to allow shrink-to-fit on any worker, we need to be able to migrate a source from one worker to the other. We may come up with different solutions for push and pull sources, because whatever connects to a push source will need to be informed about having to reconnect to a different place, or some other trickery will have to be done to make sure we move the connection to the right place. For pull sources, this seems much easier.
1.0
Add ability to migrate a source to a different worker. - Currently a source cannot be moved from a worker. In order to allow shrink-to-fit on any worker, we need to be able to migrate a source from one worker to the other. We may come up with different solutions for push and pull sources, because whatever connects to a push source will need to be informed about having to reconnect to a different place, or some other trickery will have to be done to make sure we move the connection to the right place. For pull sources, this seems much easier.
non_perf
add ability to migrate a source to a different worker currently a source cannot be moved from a worker in order to allow shrink to fit on any worker we need to be able to migrate a source from one worker to the other we may come up with different solutions for push and pull sources because whatever connects to a push source will need to be informed about having to reconnect to a different place or some other trickery will have to be done to make sure we move the connection to the right place for pull sources this seems much easier
0
407,678
27,624,984,463
IssuesEvent
2023-03-10 05:37:41
binance/binance-connector-dotnet
https://api.github.com/repos/binance/binance-connector-dotnet
closed
Documentation: BSwap.Swap not working from USDT to ETH (binance-connector-dotnet)
documentation Reviewing
### Description Hello, i am not sure what i am doing wrong, i am trying to make a swap from USDT to ETH but it gives me an "Exception of type 'Binance.Common.BinanceClientException' error. USDT to BUSD works fine. Any help will be appreciated, thank you. CODE: var bSwap = new BSwap(binanceUrl, binanceApiKey, binanceApiSecret); var result = await bSwap.Swap("USDT", "ETH", 5.0m); print(result);
1.0
Documentation: BSwap.Swap not working from USDT to ETH (binance-connector-dotnet) - ### Description Hello, i am not sure what i am doing wrong, i am trying to make a swap from USDT to ETH but it gives me an "Exception of type 'Binance.Common.BinanceClientException' error. USDT to BUSD works fine. Any help will be appreciated, thank you. CODE: var bSwap = new BSwap(binanceUrl, binanceApiKey, binanceApiSecret); var result = await bSwap.Swap("USDT", "ETH", 5.0m); print(result);
non_perf
documentation bswap swap not working from usdt to eth binance connector dotnet description hello i am not sure what i am doing wrong i am trying to make a swap from usdt to eth but it gives me an exception of type binance common binanceclientexception error usdt to busd works fine any help will be appreciated thank you code var bswap new bswap binanceurl binanceapikey binanceapisecret var result await bswap swap usdt eth print result
0
575,524
17,033,844,413
IssuesEvent
2021-07-05 02:25:20
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
opened
solvers: Ipopt segfaults when a problem has a quadratic objects + a Lorentz cone constraint
component: mathematical program priority: medium team: manipulation type: bug
This runs fine on Mosek, but causes IPOPT to segfault: ``` from pydrake.all import MathematicalProgram, Solve prog = MathematicalProgram() x = prog.NewContinuousVariables(2) prog.AddLorentzConeConstraint(x) prog.AddQuadraticCost(x.dot(x)) result = Solve(prog) print(result.is_success()) print(result.get_solver_id().name()) ```
1.0
solvers: Ipopt segfaults when a problem has a quadratic objects + a Lorentz cone constraint - This runs fine on Mosek, but causes IPOPT to segfault: ``` from pydrake.all import MathematicalProgram, Solve prog = MathematicalProgram() x = prog.NewContinuousVariables(2) prog.AddLorentzConeConstraint(x) prog.AddQuadraticCost(x.dot(x)) result = Solve(prog) print(result.is_success()) print(result.get_solver_id().name()) ```
non_perf
solvers ipopt segfaults when a problem has a quadratic objects a lorentz cone constraint this runs fine on mosek but causes ipopt to segfault from pydrake all import mathematicalprogram solve prog mathematicalprogram x prog newcontinuousvariables prog addlorentzconeconstraint x prog addquadraticcost x dot x result solve prog print result is success print result get solver id name
0
367,050
10,833,483,517
IssuesEvent
2019-11-11 13:02:32
CESARBR/knot-gateway-webui
https://api.github.com/repos/CESARBR/knot-gateway-webui
opened
After removing device, the WebUI shows an error
bug priority: low
**Expected Behavior**: No error should be shown after removing a device. **Current Behavior**: The message `An unexpected error occurred` is showed after removing a device. **Context (Environment):** The WebUI should be configured with a user. **Steps to reproduce:** - Add a device - Remove the device and wait for a few seconds. - The message `An unexpected error occurred` is showed. **Possible solution:** Avoid fetching data from devices removed.
1.0
After removing device, the WebUI shows an error - **Expected Behavior**: No error should be shown after removing a device. **Current Behavior**: The message `An unexpected error occurred` is showed after removing a device. **Context (Environment):** The WebUI should be configured with a user. **Steps to reproduce:** - Add a device - Remove the device and wait for a few seconds. - The message `An unexpected error occurred` is showed. **Possible solution:** Avoid fetching data from devices removed.
non_perf
after removing device the webui shows an error expected behavior no error should be shown after removing a device current behavior the message an unexpected error occurred is showed after removing a device context environment the webui should be configured with a user steps to reproduce add a device remove the device and wait for a few seconds the message an unexpected error occurred is showed possible solution avoid fetching data from devices removed
0
21,115
10,574,280,617
IssuesEvent
2019-10-07 13:44:02
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
pgwire: create an audit log for user logins
A-security A-sql-pgwire C-bug C-enhancement
Requested on forum: https://forum.cockroachlabs.com/t/how-to-audit-login-attempts/3182/2 We need/want a log of success/failed login attempts and also user logouts (or conn dropped). This is currently unavailable via the existing logging and audit facilities.
True
pgwire: create an audit log for user logins - Requested on forum: https://forum.cockroachlabs.com/t/how-to-audit-login-attempts/3182/2 We need/want a log of success/failed login attempts and also user logouts (or conn dropped). This is currently unavailable via the existing logging and audit facilities.
non_perf
pgwire create an audit log for user logins requested on forum we need want a log of success failed login attempts and also user logouts or conn dropped this is currently unavailable via the existing logging and audit facilities
0
19,154
10,326,083,407
IssuesEvent
2019-09-01 23:13:32
modin-project/modin
https://api.github.com/repos/modin-project/modin
closed
UserWarning: `Series.__getstate__` defaulting to pandas implementation.
Performance 🚀 bug :bug:
<!-- General questions should be asked on the mailing list modin-dev@googlegroups.com. Before submitting an issue, please fill out the following form. --> ### System information - **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**:Ubuntu 18.04.1 LTS - **Modin installed from (source or binary)**:binary - **Modin version**:0.5.4 - **Python version**:3.6.8 - **Exact command to reproduce**: The ‘merged’ is a DataFrame that 'modin.pandas.concat(list)' function generates. The column of ‘ticker’ is a series that the type of element is an integer. ``` merged['ticker'] = merged['ticker'].astype(str) ``` <!-- You can obtain the Modin version with python -c "import modin; print(modin.__version__)" --> ### Describe the problem <!-- Describe the problem clearly here. --> Then raise a warning: ``` UserWarning: `Series.__getstate__` defaulting to pandas implementation. ``` The program is stuck and don't go on. ### Source code / logs <!-- Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. Try to provide a reproducible test case that is the bare minimum necessary to generate the problem. -->
True
UserWarning: `Series.__getstate__` defaulting to pandas implementation. - <!-- General questions should be asked on the mailing list modin-dev@googlegroups.com. Before submitting an issue, please fill out the following form. --> ### System information - **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**:Ubuntu 18.04.1 LTS - **Modin installed from (source or binary)**:binary - **Modin version**:0.5.4 - **Python version**:3.6.8 - **Exact command to reproduce**: The ‘merged’ is a DataFrame that 'modin.pandas.concat(list)' function generates. The column of ‘ticker’ is a series that the type of element is an integer. ``` merged['ticker'] = merged['ticker'].astype(str) ``` <!-- You can obtain the Modin version with python -c "import modin; print(modin.__version__)" --> ### Describe the problem <!-- Describe the problem clearly here. --> Then raise a warning: ``` UserWarning: `Series.__getstate__` defaulting to pandas implementation. ``` The program is stuck and don't go on. ### Source code / logs <!-- Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. Try to provide a reproducible test case that is the bare minimum necessary to generate the problem. -->
perf
userwarning series getstate defaulting to pandas implementation general questions should be asked on the mailing list modin dev googlegroups com before submitting an issue please fill out the following form system information os platform and distribution e g linux ubuntu ubuntu lts modin installed from source or binary binary modin version python version exact command to reproduce the ‘merged’ is a dataframe that modin pandas concat list function generates the column of ‘ticker’ is a series that the type of element is an integer merged merged astype str you can obtain the modin version with python c import modin print modin version describe the problem then raise a warning userwarning series getstate defaulting to pandas implementation the program is stuck and don t go on source code logs
1
3,728
2,906,269,741
IssuesEvent
2015-06-19 08:55:08
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
User Registration - New User Account Activation=Admin. Admin does not need to confirm activation
No Code Attached Yet
#### Steps to reproduce the issue Create a new Joomla site, with one Super User (automatic). Version 3.4.1 Allow users to register: Users > User Manager > Options Make the following changes: Allow User Registration = Yes Send Password = No New User Account Activation = Admin On the front end, Create an Account (using Create an Account link) Enter in your name, password and a valid email address. Submit the registration. Verify on the backend (under user manager) that user has been added but that "Enabled" and "Activated" are OFF/Disable. Check your email for the registration email. Click the link in the email to verify the account. You should be directed to the Joomla site with a message that your email address has been verified and once an administrator approves it you can login. Verify in the backend that the new user shows "Activated" but not "Enabled". #### Expected result You should see that the user is "Activated" but not "Enabled". The user should not be able to login. #### Actual result The user is both "Activated and Enabled". No admin intervention was required. #### System information (as much as possible) Joomla Site hosted on Cloudaccess.net. Version 3.4.1 #### Additional comments Recap of existing forum posts for this issue: Perfect description of the issue (but, can happen if you only have one admin) http://forum.joomla.org/viewtopic.php?f=708&t=879213&p=3281053&hilit=user+activation#p3281053 and another one (v 3.4) http://forum.joomla.org/viewtopic.php?f=719&t=880887 http://forum.joomla.org/viewtopic.php?f=708&t=856769&hilit=user+activation not exactly the issue https://github.com/joomla/joomla-cms/issues/4376 And a post I responded to, but unsure that it is being looked into further (and this does not require that two admin accounts be present for it to be an issue) http://forum.joomla.org/viewtopic.php?f=719&t=880887
1.0
User Registration - New User Account Activation=Admin. Admin does not need to confirm activation - #### Steps to reproduce the issue Create a new Joomla site, with one Super User (automatic). Version 3.4.1 Allow users to register: Users > User Manager > Options Make the following changes: Allow User Registration = Yes Send Password = No New User Account Activation = Admin On the front end, Create an Account (using Create an Account link) Enter in your name, password and a valid email address. Submit the registration. Verify on the backend (under user manager) that user has been added but that "Enabled" and "Activated" are OFF/Disable. Check your email for the registration email. Click the link in the email to verify the account. You should be directed to the Joomla site with a message that your email address has been verified and once an administrator approves it you can login. Verify in the backend that the new user shows "Activated" but not "Enabled". #### Expected result You should see that the user is "Activated" but not "Enabled". The user should not be able to login. #### Actual result The user is both "Activated and Enabled". No admin intervention was required. #### System information (as much as possible) Joomla Site hosted on Cloudaccess.net. Version 3.4.1 #### Additional comments Recap of existing forum posts for this issue: Perfect description of the issue (but, can happen if you only have one admin) http://forum.joomla.org/viewtopic.php?f=708&t=879213&p=3281053&hilit=user+activation#p3281053 and another one (v 3.4) http://forum.joomla.org/viewtopic.php?f=719&t=880887 http://forum.joomla.org/viewtopic.php?f=708&t=856769&hilit=user+activation not exactly the issue https://github.com/joomla/joomla-cms/issues/4376 And a post I responded to, but unsure that it is being looked into further (and this does not require that two admin accounts be present for it to be an issue) http://forum.joomla.org/viewtopic.php?f=719&t=880887
non_perf
user registration new user account activation admin admin does not need to confirm activation steps to reproduce the issue create a new joomla site with one super user automatic version allow users to register users user manager options make the following changes allow user registration yes send password no new user account activation admin on the front end create an account using create an account link enter in your name password and a valid email address submit the registration verify on the backend under user manager that user has been added but that enabled and activated are off disable check your email for the registration email click the link in the email to verify the account you should be directed to the joomla site with a message that your email address has been verified and once an administrator approves it you can login verify in the backend that the new user shows activated but not enabled expected result you should see that the user is activated but not enabled the user should not be able to login actual result the user is both activated and enabled no admin intervention was required system information as much as possible joomla site hosted on cloudaccess net version additional comments recap of existing forum posts for this issue perfect description of the issue but can happen if you only have one admin and another one v not exactly the issue and a post i responded to but unsure that it is being looked into further and this does not require that two admin accounts be present for it to be an issue
0
143,025
21,932,070,775
IssuesEvent
2022-05-23 10:37:20
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Explore adding site width and padding to global styles
Needs Design Feedback Global Styles
This iteration is inspired by feedback from those exploring patterns and also the visual work explored in full site editing by @shaunandrews. Global styles have a suggested feature enabling to increase or decrease the typography scale. This could easily lead to lines with a few words or too many for comfortable reading. A great additional option could be to allow a site width setting and padding around that site. I am aware this won't be easy for all themes, so just proposing it here to work around some issues that might come up. I also know that this is being explored in templates, so it might be decided it belongs there instead. ## Option A: textual This version uses text to indicate the width and padding. It's also worth noting all illustrations show 'layout', which I am not set on and could change. Similar, these use the sidebar as that is what we are using for v1 global styles, this could be comfortable in the toolbar as other iterations happen. <img width="1257" alt="global-i1" src="https://user-images.githubusercontent.com/253067/76330175-7056f000-62e5-11ea-9441-1365d790855e.png"> It is worth noting that with the iterations to the interface, we don't have a small input with increasing/decreasing arrows. I explored a little what this could look like with a click in the state showing arrows: <img width="242" alt="Frame 1" src="https://user-images.githubusercontent.com/253067/76330376-ad22e700-62e5-11ea-8292-dddd6464c586.png"> @jasmussen looping you in to get input for interface iterations. Here it is without explicit arrows: <img width="1257" alt="global-i2" src="https://user-images.githubusercontent.com/253067/76330409-b6ac4f00-62e5-11ea-96b8-9f75e247af9a.png"> ## Option B : visual This took a large dose of full site editing inspiration, I brought in some visuals used there to explore what that could look like. I think the iconography might need exploring but for me, the visual has a certain something extra I really like. <img width="1257" alt="global-i3" src="https://user-images.githubusercontent.com/253067/76329911-122a0d00-62e5-11ea-8154-d9402f7026cb.png"> ## Feedback I would love general feedback as this is an early idea and bringing in concepts from elsewhere. That said, it would be good to get specific responses around: - Does this belong in global styles along with templates? - Do you prefer visual or textual?
1.0
Explore adding site width and padding to global styles - This iteration is inspired by feedback from those exploring patterns and also the visual work explored in full site editing by @shaunandrews. Global styles have a suggested feature enabling to increase or decrease the typography scale. This could easily lead to lines with a few words or too many for comfortable reading. A great additional option could be to allow a site width setting and padding around that site. I am aware this won't be easy for all themes, so just proposing it here to work around some issues that might come up. I also know that this is being explored in templates, so it might be decided it belongs there instead. ## Option A: textual This version uses text to indicate the width and padding. It's also worth noting all illustrations show 'layout', which I am not set on and could change. Similar, these use the sidebar as that is what we are using for v1 global styles, this could be comfortable in the toolbar as other iterations happen. <img width="1257" alt="global-i1" src="https://user-images.githubusercontent.com/253067/76330175-7056f000-62e5-11ea-9441-1365d790855e.png"> It is worth noting that with the iterations to the interface, we don't have a small input with increasing/decreasing arrows. I explored a little what this could look like with a click in the state showing arrows: <img width="242" alt="Frame 1" src="https://user-images.githubusercontent.com/253067/76330376-ad22e700-62e5-11ea-8292-dddd6464c586.png"> @jasmussen looping you in to get input for interface iterations. Here it is without explicit arrows: <img width="1257" alt="global-i2" src="https://user-images.githubusercontent.com/253067/76330409-b6ac4f00-62e5-11ea-96b8-9f75e247af9a.png"> ## Option B : visual This took a large dose of full site editing inspiration, I brought in some visuals used there to explore what that could look like. I think the iconography might need exploring but for me, the visual has a certain something extra I really like. <img width="1257" alt="global-i3" src="https://user-images.githubusercontent.com/253067/76329911-122a0d00-62e5-11ea-8154-d9402f7026cb.png"> ## Feedback I would love general feedback as this is an early idea and bringing in concepts from elsewhere. That said, it would be good to get specific responses around: - Does this belong in global styles along with templates? - Do you prefer visual or textual?
non_perf
explore adding site width and padding to global styles this iteration is inspired by feedback from those exploring patterns and also the visual work explored in full site editing by shaunandrews global styles have a suggested feature enabling to increase or decrease the typography scale this could easily lead to lines with a few words or too many for comfortable reading a great additional option could be to allow a site width setting and padding around that site i am aware this won t be easy for all themes so just proposing it here to work around some issues that might come up i also know that this is being explored in templates so it might be decided it belongs there instead option a textual this version uses text to indicate the width and padding it s also worth noting all illustrations show layout which i am not set on and could change similar these use the sidebar as that is what we are using for global styles this could be comfortable in the toolbar as other iterations happen img width alt global src it is worth noting that with the iterations to the interface we don t have a small input with increasing decreasing arrows i explored a little what this could look like with a click in the state showing arrows img width alt frame src jasmussen looping you in to get input for interface iterations here it is without explicit arrows img width alt global src option b visual this took a large dose of full site editing inspiration i brought in some visuals used there to explore what that could look like i think the iconography might need exploring but for me the visual has a certain something extra i really like img width alt global src feedback i would love general feedback as this is an early idea and bringing in concepts from elsewhere that said it would be good to get specific responses around does this belong in global styles along with templates do you prefer visual or textual
0
47,157
24,894,757,522
IssuesEvent
2022-10-28 14:53:03
ualbertalib/jupiter
https://api.github.com/repos/ualbertalib/jupiter
opened
Google detected Largest Contentful Paint
performance
**Describe the bug** Google systems have detected slow LCP (Largest Contentful Paint) on your website. LCP is an important, user-centric metric for measuring perceived page load speed. This might result in lower search rankings. An example page with slow loading speed on your site: https://era.library.ualberta.ca/items/ed56e5c2-4392-4137-9996-e9be26c9872d Visit the Core Web Vitals report linked below to see more affected URLs on your site. You can resolve this issue by improving components on your pages that may cause high LCP. Please review the articles below for more information. * https://search.google.com/search-console/core-web-vitals/drilldown?resource_id=https%3A%2F%2Fera.library.ualberta.ca%2F&item_key=CAMQAhgD&hl=en-CA * https://search.google.com/search-console/core-web-vitals?resource_id=https://era.library.ualberta.ca/&utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA * https://developers.google.com/search/docs/appearance/page-experience?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA * https://web.dev/optimize-lcp/?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA * https://support.google.com/webmasters/community/?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA **To Reproduce** Visit a page like https://era.library.ualberta.ca/items/ed56e5c2-4392-4137-9996-e9be26c9872d **Expected behavior** Pages take less than 4s. **Screenshots** ![image](https://user-images.githubusercontent.com/1220762/198655692-1bc8989d-2742-464f-8519-f1d0c4dcc3df.png) ![image](https://user-images.githubusercontent.com/1220762/198659133-fe3c0055-50df-495b-a744-808e4e7a2c3f.png)
True
Google detected Largest Contentful Paint - **Describe the bug** Google systems have detected slow LCP (Largest Contentful Paint) on your website. LCP is an important, user-centric metric for measuring perceived page load speed. This might result in lower search rankings. An example page with slow loading speed on your site: https://era.library.ualberta.ca/items/ed56e5c2-4392-4137-9996-e9be26c9872d Visit the Core Web Vitals report linked below to see more affected URLs on your site. You can resolve this issue by improving components on your pages that may cause high LCP. Please review the articles below for more information. * https://search.google.com/search-console/core-web-vitals/drilldown?resource_id=https%3A%2F%2Fera.library.ualberta.ca%2F&item_key=CAMQAhgD&hl=en-CA * https://search.google.com/search-console/core-web-vitals?resource_id=https://era.library.ualberta.ca/&utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA * https://developers.google.com/search/docs/appearance/page-experience?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA * https://web.dev/optimize-lcp/?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA * https://support.google.com/webmasters/community/?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA **To Reproduce** Visit a page like https://era.library.ualberta.ca/items/ed56e5c2-4392-4137-9996-e9be26c9872d **Expected behavior** Pages take less than 4s. **Screenshots** ![image](https://user-images.githubusercontent.com/1220762/198655692-1bc8989d-2742-464f-8519-f1d0c4dcc3df.png) ![image](https://user-images.githubusercontent.com/1220762/198659133-fe3c0055-50df-495b-a744-808e4e7a2c3f.png)
perf
google detected largest contentful paint describe the bug google systems have detected slow lcp largest contentful paint on your website lcp is an important user centric metric for measuring perceived page load speed this might result in lower search rankings an example page with slow loading speed on your site visit the core web vitals report linked below to see more affected urls on your site you can resolve this issue by improving components on your pages that may cause high lcp please review the articles below for more information to reproduce visit a page like expected behavior pages take less than screenshots
1
243,203
18,677,925,029
IssuesEvent
2021-10-31 21:43:58
eclipse-ee4j/cargotracker
https://api.github.com/repos/eclipse-ee4j/cargotracker
closed
New migration section in the documentation
documentation Priority: Minor
Create a section in the documentation regarding the changes that has been done for migrating from Java EE 7 to Jakarta EE 8, and also in the future for the next versions. This might help also other folks trying to migrate their projects @alexmario74, do you mind keeping track of the changes / issues you are facing? If @m-reza-rahman thinks could be a good idea adding the new section.
1.0
New migration section in the documentation - Create a section in the documentation regarding the changes that has been done for migrating from Java EE 7 to Jakarta EE 8, and also in the future for the next versions. This might help also other folks trying to migrate their projects @alexmario74, do you mind keeping track of the changes / issues you are facing? If @m-reza-rahman thinks could be a good idea adding the new section.
non_perf
new migration section in the documentation create a section in the documentation regarding the changes that has been done for migrating from java ee to jakarta ee and also in the future for the next versions this might help also other folks trying to migrate their projects do you mind keeping track of the changes issues you are facing if m reza rahman thinks could be a good idea adding the new section
0
46,126
24,374,396,035
IssuesEvent
2022-10-03 22:44:09
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
storage: performance regression in BenchmarkMVCCPutDelete_Pebble
C-performance A-storage T-storage
**Describe the problem** Currently, on the 22.2 release branch there is a ~20% performance regression in `BenchmarkMVCCPutDelete_Pebble`, relative to 22.1: ```bash $ benchstat bench-22.1 bench-22.2 name old time/op new time/op delta MVCCPutDelete_Pebble-16 8.36µs ± 2% 10.09µs ± 1% +20.71% (p=0.000 n=10+10) ``` **To Reproduce** ```bash $ ./dev bench ./pkg/storage --filter MVCCPutDelete_Pebble --count 10 -v --stream-output --ignore-cache ```
True
storage: performance regression in BenchmarkMVCCPutDelete_Pebble - **Describe the problem** Currently, on the 22.2 release branch there is a ~20% performance regression in `BenchmarkMVCCPutDelete_Pebble`, relative to 22.1: ```bash $ benchstat bench-22.1 bench-22.2 name old time/op new time/op delta MVCCPutDelete_Pebble-16 8.36µs ± 2% 10.09µs ± 1% +20.71% (p=0.000 n=10+10) ``` **To Reproduce** ```bash $ ./dev bench ./pkg/storage --filter MVCCPutDelete_Pebble --count 10 -v --stream-output --ignore-cache ```
perf
storage performance regression in benchmarkmvccputdelete pebble describe the problem currently on the release branch there is a performance regression in benchmarkmvccputdelete pebble relative to bash benchstat bench bench name old time op new time op delta mvccputdelete pebble ± ± p n to reproduce bash dev bench pkg storage filter mvccputdelete pebble count v stream output ignore cache
1
34,035
16,385,360,853
IssuesEvent
2021-05-17 09:43:23
space-wizards/RobustToolbox
https://api.github.com/repos/space-wizards/RobustToolbox
closed
Cache GridId and MapId inside TransformComponent
Area: ECS Project: Shared Type: Performance
These now require recursive lookups via the parent to calculate and it's *extremely* expensive.
True
Cache GridId and MapId inside TransformComponent - These now require recursive lookups via the parent to calculate and it's *extremely* expensive.
perf
cache gridid and mapid inside transformcomponent these now require recursive lookups via the parent to calculate and it s extremely expensive
1
35,422
17,079,159,128
IssuesEvent
2021-07-08 00:43:31
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Refactor importer NFT logic to use upsert
P1 enhancement nft parser performance
**Problem** The importer is currently using old logic for handling NFTs and NFT transfers, as opposed to the new upsert logic added in #2006. **Solution** Refactor the importer to use similar upsert logic for NFT and NFT transfer handling. **Alternatives** **Additional Context** There may need to be some rework around how NFT transfers update the accountId of the NFT, primarily during an NFT mint transaction.
True
Refactor importer NFT logic to use upsert - **Problem** The importer is currently using old logic for handling NFTs and NFT transfers, as opposed to the new upsert logic added in #2006. **Solution** Refactor the importer to use similar upsert logic for NFT and NFT transfer handling. **Alternatives** **Additional Context** There may need to be some rework around how NFT transfers update the accountId of the NFT, primarily during an NFT mint transaction.
perf
refactor importer nft logic to use upsert problem the importer is currently using old logic for handling nfts and nft transfers as opposed to the new upsert logic added in solution refactor the importer to use similar upsert logic for nft and nft transfer handling alternatives additional context there may need to be some rework around how nft transfers update the accountid of the nft primarily during an nft mint transaction
1
366,983
10,832,760,749
IssuesEvent
2019-11-11 11:21:03
francismaria/MaTheX2Java
https://api.github.com/repos/francismaria/MaTheX2Java
closed
Add error panel
frontend medium priority
Add an error panel to show messages of errors that occurred during the code generation.
1.0
Add error panel - Add an error panel to show messages of errors that occurred during the code generation.
non_perf
add error panel add an error panel to show messages of errors that occurred during the code generation
0
308,744
23,265,202,973
IssuesEvent
2022-08-04 16:41:12
Qiskit/qiskit_sphinx_theme
https://api.github.com/repos/Qiskit/qiskit_sphinx_theme
opened
Set up COS deployment
documentation docs migration
**As a** Qiskit docs dev, **I need** to be able to deploy html files from this repository to COS when work is merged into main **So that** frontend development work can be more efficient (i.e. decoupled from qiskit releases)
1.0
Set up COS deployment - **As a** Qiskit docs dev, **I need** to be able to deploy html files from this repository to COS when work is merged into main **So that** frontend development work can be more efficient (i.e. decoupled from qiskit releases)
non_perf
set up cos deployment as a qiskit docs dev i need to be able to deploy html files from this repository to cos when work is merged into main so that frontend development work can be more efficient i e decoupled from qiskit releases
0
56,341
31,881,639,719
IssuesEvent
2023-09-16 12:52:44
neovim/neovim
https://api.github.com/repos/neovim/neovim
closed
Lag is noticeable within comment while the treesitter highlight is enabled.
performance treesitter
### Problem This issue is a bit hard to reproduce without following exactly steps. When typing `/**` comment in vue files, the input lag is noticeable. ### Steps to reproduce put `slow-ts.lua` in `/tmp`: ```lua vim.g.mapleader = " " vim.g.maplocalleader = "," vim.opt.number = true for name, url in pairs({ -- ADD PLUGINS _NECESSARY_ TO REPRODUCE THE ISSUE, e.g: -- some_plugin = 'https://github.com/author/plugin.nvim' treesitter = "https://github.com/nvim-treesitter/nvim-treesitter", }) do local install_path = vim.fn.fnamemodify("slow_ts/" .. name, ":p") if vim.fn.isdirectory(install_path) == 0 then if name == "treesitter" then vim.fn.system({ "git", "clone", "--branch", "main", "--depth=1", url, install_path }) else vim.fn.system({ "git", "clone", "--depth=1", url, install_path }) end end vim.opt.runtimepath:append(install_path) end -- =============== require("nvim-treesitter").setup({ ensure_install = { "vue", "javascript", "typescript", "comment", "jsdoc" }, auto_install = false, install_dir = vim.fn.fnamemodify("slow_ts/.site", ":p"), }) local ftau = vim.api.nvim_create_augroup("option_ft", { clear = true }) vim.api.nvim_create_autocmd("FileType", { group = ftau, pattern = { "*" }, callback = function(args) local buf = args.buf -- start highlighter. if not pcall(vim.treesitter.start, buf) then print("ERROR! highlight not started") return end end, }) vim.cmd("colorscheme default") ``` Put a `test.vue` file along side with `slow-ts.lua`: file content: https://gist.github.com/towry/74d6ff3312e18c6fd39367d1b4e477cf#file-test-vue Steps to repro it: 1. `nvim --clean -u slow-ts.lua`, wait for the nvim-treesitter setup is done. 2. make sure `vue`, `comment` parsers is installed. 3. `nvim --clean -u slow-ts.lua test.vue` 4. (important) press `ctrl-d` multiple times to move cursor down and press `ctrl-u` to make sure line 26 is visible and move cursor to line 26 with `j/k`. 5. start insert comment with content `/***` above line 26. Note: it seems if I press `:26` to directly move cursor to line 26, this bug will not be triggered. A lag demo can be seen here: https://github.com/neovim/neovim/issues/25074#issuecomment-1713502035 ### Expected behavior No lag in comment. ### Neovim version (nvim -v) NVIM v0.10.0-dev-1088+g6b5f44817 ### Vim (not Nvim) behaves the same? no ### Operating system/version macos ### Terminal name/version kitty + macos terminal ### $TERM environment variable xterm-256color ### Installation build from repo
True
Lag is noticeable within comment while the treesitter highlight is enabled. - ### Problem This issue is a bit hard to reproduce without following exactly steps. When typing `/**` comment in vue files, the input lag is noticeable. ### Steps to reproduce put `slow-ts.lua` in `/tmp`: ```lua vim.g.mapleader = " " vim.g.maplocalleader = "," vim.opt.number = true for name, url in pairs({ -- ADD PLUGINS _NECESSARY_ TO REPRODUCE THE ISSUE, e.g: -- some_plugin = 'https://github.com/author/plugin.nvim' treesitter = "https://github.com/nvim-treesitter/nvim-treesitter", }) do local install_path = vim.fn.fnamemodify("slow_ts/" .. name, ":p") if vim.fn.isdirectory(install_path) == 0 then if name == "treesitter" then vim.fn.system({ "git", "clone", "--branch", "main", "--depth=1", url, install_path }) else vim.fn.system({ "git", "clone", "--depth=1", url, install_path }) end end vim.opt.runtimepath:append(install_path) end -- =============== require("nvim-treesitter").setup({ ensure_install = { "vue", "javascript", "typescript", "comment", "jsdoc" }, auto_install = false, install_dir = vim.fn.fnamemodify("slow_ts/.site", ":p"), }) local ftau = vim.api.nvim_create_augroup("option_ft", { clear = true }) vim.api.nvim_create_autocmd("FileType", { group = ftau, pattern = { "*" }, callback = function(args) local buf = args.buf -- start highlighter. if not pcall(vim.treesitter.start, buf) then print("ERROR! highlight not started") return end end, }) vim.cmd("colorscheme default") ``` Put a `test.vue` file along side with `slow-ts.lua`: file content: https://gist.github.com/towry/74d6ff3312e18c6fd39367d1b4e477cf#file-test-vue Steps to repro it: 1. `nvim --clean -u slow-ts.lua`, wait for the nvim-treesitter setup is done. 2. make sure `vue`, `comment` parsers is installed. 3. `nvim --clean -u slow-ts.lua test.vue` 4. (important) press `ctrl-d` multiple times to move cursor down and press `ctrl-u` to make sure line 26 is visible and move cursor to line 26 with `j/k`. 5. start insert comment with content `/***` above line 26. Note: it seems if I press `:26` to directly move cursor to line 26, this bug will not be triggered. A lag demo can be seen here: https://github.com/neovim/neovim/issues/25074#issuecomment-1713502035 ### Expected behavior No lag in comment. ### Neovim version (nvim -v) NVIM v0.10.0-dev-1088+g6b5f44817 ### Vim (not Nvim) behaves the same? no ### Operating system/version macos ### Terminal name/version kitty + macos terminal ### $TERM environment variable xterm-256color ### Installation build from repo
perf
lag is noticeable within comment while the treesitter highlight is enabled problem this issue is a bit hard to reproduce without following exactly steps when typing comment in vue files the input lag is noticeable steps to reproduce put slow ts lua in tmp lua vim g mapleader vim g maplocalleader vim opt number true for name url in pairs add plugins necessary to reproduce the issue e g some plugin treesitter do local install path vim fn fnamemodify slow ts name p if vim fn isdirectory install path then if name treesitter then vim fn system git clone branch main depth url install path else vim fn system git clone depth url install path end end vim opt runtimepath append install path end require nvim treesitter setup ensure install vue javascript typescript comment jsdoc auto install false install dir vim fn fnamemodify slow ts site p local ftau vim api nvim create augroup option ft clear true vim api nvim create autocmd filetype group ftau pattern callback function args local buf args buf start highlighter if not pcall vim treesitter start buf then print error highlight not started return end end vim cmd colorscheme default put a test vue file along side with slow ts lua file content steps to repro it nvim clean u slow ts lua wait for the nvim treesitter setup is done make sure vue comment parsers is installed nvim clean u slow ts lua test vue important press ctrl d multiple times to move cursor down and press ctrl u to make sure line is visible and move cursor to line with j k start insert comment with content above line note it seems if i press to directly move cursor to line this bug will not be triggered a lag demo can be seen here expected behavior no lag in comment neovim version nvim v nvim dev vim not nvim behaves the same no operating system version macos terminal name version kitty macos terminal term environment variable xterm installation build from repo
1
17,334
9,703,041,157
IssuesEvent
2019-05-27 10:15:30
Dart-Code/Dart-Code
https://api.github.com/repos/Dart-Code/Dart-Code
opened
Improve performance of completions
important is performance
As well as probably being the cause of #1721, the performance of completions with unimported symbols is poor. Testing on Windows PC with the following test (inside `test\flutter_only\providers\completion_item_provider.test.ts`): ``` describe("with SuggestionSet support", () => { beforeEach("ensure SuggestionSets are supported", function () { if (!extApi.analyzerCapabilities.supportsAvailableSuggestions) this.skip(); }); it.only("includes unimported symbols", async () => { await setTestContent(` main() { ProcessInf } `); const count = 50; const start = Date.now(); for (let i = 0; i < 50; i++) { const completions = await getCompletionsAt("ProcessInf^"); ensureCompletion(completions, vs.CompletionItemKind.Class, "ProcessInfo", "ProcessInfo"); } const end = Date.now(); console.log(`Took ${end - start}ms to do ${count} completion requests`); }); }); ``` Output from a few runs: ``` Took 56426ms to do 50 completion requests Took 57034ms to do 50 completion requests Took 57607ms to do 50 completion requests ```
True
Improve performance of completions - As well as probably being the cause of #1721, the performance of completions with unimported symbols is poor. Testing on Windows PC with the following test (inside `test\flutter_only\providers\completion_item_provider.test.ts`): ``` describe("with SuggestionSet support", () => { beforeEach("ensure SuggestionSets are supported", function () { if (!extApi.analyzerCapabilities.supportsAvailableSuggestions) this.skip(); }); it.only("includes unimported symbols", async () => { await setTestContent(` main() { ProcessInf } `); const count = 50; const start = Date.now(); for (let i = 0; i < 50; i++) { const completions = await getCompletionsAt("ProcessInf^"); ensureCompletion(completions, vs.CompletionItemKind.Class, "ProcessInfo", "ProcessInfo"); } const end = Date.now(); console.log(`Took ${end - start}ms to do ${count} completion requests`); }); }); ``` Output from a few runs: ``` Took 56426ms to do 50 completion requests Took 57034ms to do 50 completion requests Took 57607ms to do 50 completion requests ```
perf
improve performance of completions as well as probably being the cause of the performance of completions with unimported symbols is poor testing on windows pc with the following test inside test flutter only providers completion item provider test ts describe with suggestionset support beforeeach ensure suggestionsets are supported function if extapi analyzercapabilities supportsavailablesuggestions this skip it only includes unimported symbols async await settestcontent main processinf const count const start date now for let i i i const completions await getcompletionsat processinf ensurecompletion completions vs completionitemkind class processinfo processinfo const end date now console log took end start ms to do count completion requests output from a few runs took to do completion requests took to do completion requests took to do completion requests
1
49,950
26,403,482,641
IssuesEvent
2023-01-13 04:56:34
mmtk/mmtk-julia
https://api.github.com/repos/mmtk/mmtk-julia
opened
Improving the performance when querying object size
performance
Julia's stock GC uses page metadata information to ask for the size of smaller objects (pool-allocated). Since MMTk does not have that information available, we perform a [check](https://github.com/mmtk/mmtk-julia/blob/970fc5c49c2f9edfefbffdcff9ec753cf0791fa1/julia/mmtk_julia.c#L318) on the object type to calculate its size, which is likely to cause performance problems.
True
Improving the performance when querying object size - Julia's stock GC uses page metadata information to ask for the size of smaller objects (pool-allocated). Since MMTk does not have that information available, we perform a [check](https://github.com/mmtk/mmtk-julia/blob/970fc5c49c2f9edfefbffdcff9ec753cf0791fa1/julia/mmtk_julia.c#L318) on the object type to calculate its size, which is likely to cause performance problems.
perf
improving the performance when querying object size julia s stock gc uses page metadata information to ask for the size of smaller objects pool allocated since mmtk does not have that information available we perform a on the object type to calculate its size which is likely to cause performance problems
1
34,526
16,594,188,279
IssuesEvent
2021-06-01 11:30:03
ethereumjs/ethereumjs-monorepo
https://api.github.com/repos/ethereumjs/ethereumjs-monorepo
closed
VM: Analyze debug functionality performance behavior
eval: needs discussion eval: needs validation package: vm prio: P3 important type: enhancement type: performance
Debug functionality with the `debug` package introduced in #1080 and follow-up PRs might have been coming along with some performance penalty, as raised by @alcuadrado in the chat. This is still unconfirmed though and we should validate before we act upon this. Performance influencing code parts are mainly (solely?) the preparations for the `debug` strings, posting a random example here from `runBlock.ts`: ```typescript debug(`Invalid gasUsed received=${result.gasUsed} expected=${block.header.gasUsed}`) ``` A structural solution might be to only run/create the `debug` messages if the `VM` has been passed a flag `debug`. This also needs some evaluation/proof if it really solves the problem or if strings are build nonetheless. An alternative/additional first somewhat more targeted step would be to have a look at the different calls and eliminate/refactor all calls with computation. I tried to be really careful with this on implementation, but there are e.g. some `hash()` calls added (or at least one?). Here is an example from `runTx.ts`: ```typescript `debug(`New tx run hash=${opts.tx.hash().toString('hex')} sender=${caller.toString()}`)` ``` (I would personally be a fan of this second approach before here and then give this again some more time or eventually close directly. Respectively we can do this targeted optimization anyhow since this *will* have *some* effect on performance)
True
VM: Analyze debug functionality performance behavior - Debug functionality with the `debug` package introduced in #1080 and follow-up PRs might have been coming along with some performance penalty, as raised by @alcuadrado in the chat. This is still unconfirmed though and we should validate before we act upon this. Performance influencing code parts are mainly (solely?) the preparations for the `debug` strings, posting a random example here from `runBlock.ts`: ```typescript debug(`Invalid gasUsed received=${result.gasUsed} expected=${block.header.gasUsed}`) ``` A structural solution might be to only run/create the `debug` messages if the `VM` has been passed a flag `debug`. This also needs some evaluation/proof if it really solves the problem or if strings are build nonetheless. An alternative/additional first somewhat more targeted step would be to have a look at the different calls and eliminate/refactor all calls with computation. I tried to be really careful with this on implementation, but there are e.g. some `hash()` calls added (or at least one?). Here is an example from `runTx.ts`: ```typescript `debug(`New tx run hash=${opts.tx.hash().toString('hex')} sender=${caller.toString()}`)` ``` (I would personally be a fan of this second approach before here and then give this again some more time or eventually close directly. Respectively we can do this targeted optimization anyhow since this *will* have *some* effect on performance)
perf
vm analyze debug functionality performance behavior debug functionality with the debug package introduced in and follow up prs might have been coming along with some performance penalty as raised by alcuadrado in the chat this is still unconfirmed though and we should validate before we act upon this performance influencing code parts are mainly solely the preparations for the debug strings posting a random example here from runblock ts typescript debug invalid gasused received result gasused expected block header gasused a structural solution might be to only run create the debug messages if the vm has been passed a flag debug this also needs some evaluation proof if it really solves the problem or if strings are build nonetheless an alternative additional first somewhat more targeted step would be to have a look at the different calls and eliminate refactor all calls with computation i tried to be really careful with this on implementation but there are e g some hash calls added or at least one here is an example from runtx ts typescript debug new tx run hash opts tx hash tostring hex sender caller tostring i would personally be a fan of this second approach before here and then give this again some more time or eventually close directly respectively we can do this targeted optimization anyhow since this will have some effect on performance
1
193,625
15,382,659,100
IssuesEvent
2021-03-03 01:05:52
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Documentation for ApplyNegMuCorrection Algorithm
Documentation Muon Stale
Create documentation for ApplyNegMuCorrection Algorithm #13201
1.0
Documentation for ApplyNegMuCorrection Algorithm - Create documentation for ApplyNegMuCorrection Algorithm #13201
non_perf
documentation for applynegmucorrection algorithm create documentation for applynegmucorrection algorithm
0
382,955
26,525,950,286
IssuesEvent
2023-01-19 08:44:50
Tinkoff/taiga-ui
https://api.github.com/repos/Tinkoff/taiga-ui
closed
📚 - Palette in dropdown code example is wrong
documentation P3
### What is the affected URL? https://taiga-ui.dev/next/components/color-picker#palette ### Description The **Palette in dropdown** sample code was published just like its predecessor, **ColorSelector in dropdown** ### Which browsers have you used? - [X] Chrome - [X] Firefox - [ ] Safari - [ ] Edge ### Which operating systems have you used? - [ ] macOS - [X] Windows - [ ] Linux - [ ] iOS - [ ] Android
1.0
📚 - Palette in dropdown code example is wrong - ### What is the affected URL? https://taiga-ui.dev/next/components/color-picker#palette ### Description The **Palette in dropdown** sample code was published just like its predecessor, **ColorSelector in dropdown** ### Which browsers have you used? - [X] Chrome - [X] Firefox - [ ] Safari - [ ] Edge ### Which operating systems have you used? - [ ] macOS - [X] Windows - [ ] Linux - [ ] iOS - [ ] Android
non_perf
📚 palette in dropdown code example is wrong what is the affected url description the palette in dropdown sample code was published just like its predecessor colorselector in dropdown which browsers have you used chrome firefox safari edge which operating systems have you used macos windows linux ios android
0
43,607
23,304,860,903
IssuesEvent
2022-08-07 21:37:35
sourcefrog/cargo-mutants
https://api.github.com/repos/sourcefrog/cargo-mutants
closed
Maybe use `--release` by default
enhancement good first issue performance
`cargo mutants` does many incremental builds and test runs. In most trees, most of the time is spent running the tests. Tests will typically run faster with `--release`; in some crates it will be much faster. So that should at least be an option (#59) but according to the principle of "make it easy" perhaps `--release` should be the default. There are at least two potential downsides: 1. Conceivably in some trees the performance hit from doing the initial and incremental builds in release mode will outweigh the benefit of the tests being faster. However, even if that is true sometimes it's probably a win across most trees. 2. Some projects might have test suites that are less rigorous, or simply don't work properly, in release mode. For example they might rely on `debug_assert` to catch some bugs. I think this would be an unidiomatic use of Rust, but it might well happen. I think those could both reasonably be handled by having an option to go back to debug builds.
True
Maybe use `--release` by default - `cargo mutants` does many incremental builds and test runs. In most trees, most of the time is spent running the tests. Tests will typically run faster with `--release`; in some crates it will be much faster. So that should at least be an option (#59) but according to the principle of "make it easy" perhaps `--release` should be the default. There are at least two potential downsides: 1. Conceivably in some trees the performance hit from doing the initial and incremental builds in release mode will outweigh the benefit of the tests being faster. However, even if that is true sometimes it's probably a win across most trees. 2. Some projects might have test suites that are less rigorous, or simply don't work properly, in release mode. For example they might rely on `debug_assert` to catch some bugs. I think this would be an unidiomatic use of Rust, but it might well happen. I think those could both reasonably be handled by having an option to go back to debug builds.
perf
maybe use release by default cargo mutants does many incremental builds and test runs in most trees most of the time is spent running the tests tests will typically run faster with release in some crates it will be much faster so that should at least be an option but according to the principle of make it easy perhaps release should be the default there are at least two potential downsides conceivably in some trees the performance hit from doing the initial and incremental builds in release mode will outweigh the benefit of the tests being faster however even if that is true sometimes it s probably a win across most trees some projects might have test suites that are less rigorous or simply don t work properly in release mode for example they might rely on debug assert to catch some bugs i think this would be an unidiomatic use of rust but it might well happen i think those could both reasonably be handled by having an option to go back to debug builds
1
321,655
27,546,266,451
IssuesEvent
2023-03-07 12:01:36
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
opened
[Flaky Test] can be created without any text selected
[Type] Flaky Test
<!-- __META_DATA__:{} --> **Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.** ## Test title can be created without any text selected ## Test path `specs/editor/various/links.test.js` ## Errors <!-- __TEST_RESULTS_LIST__ --> <!-- __TEST_RESULT__ --><details> <summary> <time datetime="2023-03-07T12:01:35.760Z"><code>[2023-03-07T12:01:35.760Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4353563197"><code>try/dedicated-block-controls-filter</code></a>. </summary> ``` ● Links › can be created without any text selected expect(jest.fn()).not.toHaveErrored(expected) Expected mock function not to be called but it was called with: ["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."] at Object.assertExpectedCalls (../jest-console/build/@wordpress/jest-console/src/index.js:32:4) at runMicrotasks (<anonymous>) ``` </details><!-- /__TEST_RESULT__ --> <!-- /__TEST_RESULTS_LIST__ -->
1.0
[Flaky Test] can be created without any text selected - <!-- __META_DATA__:{} --> **Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.** ## Test title can be created without any text selected ## Test path `specs/editor/various/links.test.js` ## Errors <!-- __TEST_RESULTS_LIST__ --> <!-- __TEST_RESULT__ --><details> <summary> <time datetime="2023-03-07T12:01:35.760Z"><code>[2023-03-07T12:01:35.760Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4353563197"><code>try/dedicated-block-controls-filter</code></a>. </summary> ``` ● Links › can be created without any text selected expect(jest.fn()).not.toHaveErrored(expected) Expected mock function not to be called but it was called with: ["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."] at Object.assertExpectedCalls (../jest-console/build/@wordpress/jest-console/src/index.js:32:4) at runMicrotasks (<anonymous>) ``` </details><!-- /__TEST_RESULT__ --> <!-- /__TEST_RESULTS_LIST__ -->
non_perf
can be created without any text selected flaky test detected this is an auto generated issue by github actions please do not edit this manually test title can be created without any text selected test path specs editor various links test js errors test passed after failed attempt on a href ● links › can be created without any text selected expect jest fn not tohaveerrored expected expected mock function not to be called but it was called with at object assertexpectedcalls jest console build wordpress jest console src index js at runmicrotasks
0
2,798
3,012,603,091
IssuesEvent
2015-07-29 00:41:56
orbeon/orbeon-forms
https://api.github.com/repos/orbeon/orbeon-forms
closed
Embedding must support Form Builder
Form Builder
Right now, server-side embedding doesn't work well with Form Builder. We need to make this work properly.
1.0
Embedding must support Form Builder - Right now, server-side embedding doesn't work well with Form Builder. We need to make this work properly.
non_perf
embedding must support form builder right now server side embedding doesn t work well with form builder we need to make this work properly
0
10,328
7,154,935,519
IssuesEvent
2018-01-26 10:35:42
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
closed
Run developer-uploaded images through pngcrush/optipng/jpegoptim
component: performance triaged
We load a lot of images we don't control, and we should make sure they are as optimized as they can. This means running them through lossless compression tools like pngcrush, optipng, and jpegoptim. We can save a lot by doing this. We did this in Marketplace a while ago. The main caveat was that cachebusting needed to take this into account, because it had to happen on a separate task, so we'd resize the image, save the file, then optimize them, overwriting the previous copy. If somehow a client accessed the resized image before the optimize step, we needed to load the newer version. It can work with the naive way we do cachebusting right now, but it would be better with https://github.com/mozilla/addons-server/issues/2659 instead.
True
Run developer-uploaded images through pngcrush/optipng/jpegoptim - We load a lot of images we don't control, and we should make sure they are as optimized as they can. This means running them through lossless compression tools like pngcrush, optipng, and jpegoptim. We can save a lot by doing this. We did this in Marketplace a while ago. The main caveat was that cachebusting needed to take this into account, because it had to happen on a separate task, so we'd resize the image, save the file, then optimize them, overwriting the previous copy. If somehow a client accessed the resized image before the optimize step, we needed to load the newer version. It can work with the naive way we do cachebusting right now, but it would be better with https://github.com/mozilla/addons-server/issues/2659 instead.
perf
run developer uploaded images through pngcrush optipng jpegoptim we load a lot of images we don t control and we should make sure they are as optimized as they can this means running them through lossless compression tools like pngcrush optipng and jpegoptim we can save a lot by doing this we did this in marketplace a while ago the main caveat was that cachebusting needed to take this into account because it had to happen on a separate task so we d resize the image save the file then optimize them overwriting the previous copy if somehow a client accessed the resized image before the optimize step we needed to load the newer version it can work with the naive way we do cachebusting right now but it would be better with instead
1
3,192
3,833,125,140
IssuesEvent
2016-04-01 01:00:03
piwik/piwik
https://api.github.com/repos/piwik/piwik
closed
Improve all websites dashboard performance
c: Performance
All websites dashboard loads without any hassle when fetched for single day or week. It is problematic when you fetch 2 days for example, merging the data lasts for very long and usually does not finish. This occurrence was observed in instance with around 2k websites, this definitely is important for testing purposes.
True
Improve all websites dashboard performance - All websites dashboard loads without any hassle when fetched for single day or week. It is problematic when you fetch 2 days for example, merging the data lasts for very long and usually does not finish. This occurrence was observed in instance with around 2k websites, this definitely is important for testing purposes.
perf
improve all websites dashboard performance all websites dashboard loads without any hassle when fetched for single day or week it is problematic when you fetch days for example merging the data lasts for very long and usually does not finish this occurrence was observed in instance with around websites this definitely is important for testing purposes
1
24,084
12,213,898,885
IssuesEvent
2020-05-01 08:26:23
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
closed
FileChannelJournalSegmentReader throws half buffers away
Impact: Performance Scope: broker Type: Bug Type: Maintenance
**Description** During our investigation of the engine latency we found out that the atomix journal is a big hotspot performance wise. We had a deeper look at the code and found the `readNext` method. We have to say before that per Zeebe default the `maxEntrySize` is 4 MB. The reader will allocate a buffer which is `8MB` big. Our assumption is that it wants to do a read a head, but actually it doesn't because it clears the buffer positions etc. ```java /** Reads the next entry in the segment. */ @SuppressWarnings("unchecked") private void readNext() { // Compute the index of the next entry in the segment. final long index = getNextIndex(); try { // Read more bytes from the segment if necessary. if (memory.remaining() < maxEntrySize) { final long position = channel.position() + memory.position(); channel.position(position); memory.clear(); // <=== this resets to position 0, limit to capacity channel.read(memory); // <=== we read now again 8 MB channel.position(position); memory.flip(); } // Mark the buffer so it can be reset if necessary. memory.mark(); try { // Read the length of the entry. final int length = memory.getInt(); // If the buffer length is zero then return. if (length <= 0 || length > maxEntrySize) { memory.reset().limit(memory.position()); nextEntry = null; return; } // Read the checksum of the entry. final long checksum = memory.getInt() & 0xFFFFFFFFL; // Compute the checksum for the entry bytes. final Checksum crc32 = new CRC32(); crc32.update(memory.array(), memory.position(), length); // If the stored checksum equals the computed checksum, return the entry. if (checksum == crc32.getValue()) { final int limit = memory.limit(); memory.limit(memory.position() + length); final E entry = namespace.deserialize(memory); memory.limit(limit); nextEntry = new Indexed<>(index, entry, length); } else { memory.reset().limit(memory.position()); nextEntry = null; } } catch (final BufferUnderflowException e) { memory.reset().limit(memory.position()); nextEntry = null; } } catch (final IOException e) { throw new StorageException(e); } } ``` This means we will read `8 MB` process from these `~4 MB`until our `remaining` is less then `maxEntrySize`, so after we reached half of the buffer. Then we will read again `8MB` where the other remaining is thrown away and will end at the beginning of the buffer, because it is re-read. This problem is more an issue when the `maxEntrySize` and `realEntrySize` is very different, which is our case. We using `4MB` to support big deployments but actually most of our records are less then 1 kb. If the sizes would be nearer then we would probably also read most of the buffer, but this is not our use case so we should fix this!
True
FileChannelJournalSegmentReader throws half buffers away - **Description** During our investigation of the engine latency we found out that the atomix journal is a big hotspot performance wise. We had a deeper look at the code and found the `readNext` method. We have to say before that per Zeebe default the `maxEntrySize` is 4 MB. The reader will allocate a buffer which is `8MB` big. Our assumption is that it wants to do a read a head, but actually it doesn't because it clears the buffer positions etc. ```java /** Reads the next entry in the segment. */ @SuppressWarnings("unchecked") private void readNext() { // Compute the index of the next entry in the segment. final long index = getNextIndex(); try { // Read more bytes from the segment if necessary. if (memory.remaining() < maxEntrySize) { final long position = channel.position() + memory.position(); channel.position(position); memory.clear(); // <=== this resets to position 0, limit to capacity channel.read(memory); // <=== we read now again 8 MB channel.position(position); memory.flip(); } // Mark the buffer so it can be reset if necessary. memory.mark(); try { // Read the length of the entry. final int length = memory.getInt(); // If the buffer length is zero then return. if (length <= 0 || length > maxEntrySize) { memory.reset().limit(memory.position()); nextEntry = null; return; } // Read the checksum of the entry. final long checksum = memory.getInt() & 0xFFFFFFFFL; // Compute the checksum for the entry bytes. final Checksum crc32 = new CRC32(); crc32.update(memory.array(), memory.position(), length); // If the stored checksum equals the computed checksum, return the entry. if (checksum == crc32.getValue()) { final int limit = memory.limit(); memory.limit(memory.position() + length); final E entry = namespace.deserialize(memory); memory.limit(limit); nextEntry = new Indexed<>(index, entry, length); } else { memory.reset().limit(memory.position()); nextEntry = null; } } catch (final BufferUnderflowException e) { memory.reset().limit(memory.position()); nextEntry = null; } } catch (final IOException e) { throw new StorageException(e); } } ``` This means we will read `8 MB` process from these `~4 MB`until our `remaining` is less then `maxEntrySize`, so after we reached half of the buffer. Then we will read again `8MB` where the other remaining is thrown away and will end at the beginning of the buffer, because it is re-read. This problem is more an issue when the `maxEntrySize` and `realEntrySize` is very different, which is our case. We using `4MB` to support big deployments but actually most of our records are less then 1 kb. If the sizes would be nearer then we would probably also read most of the buffer, but this is not our use case so we should fix this!
perf
filechanneljournalsegmentreader throws half buffers away description during our investigation of the engine latency we found out that the atomix journal is a big hotspot performance wise we had a deeper look at the code and found the readnext method we have to say before that per zeebe default the maxentrysize is mb the reader will allocate a buffer which is big our assumption is that it wants to do a read a head but actually it doesn t because it clears the buffer positions etc java reads the next entry in the segment suppresswarnings unchecked private void readnext compute the index of the next entry in the segment final long index getnextindex try read more bytes from the segment if necessary if memory remaining maxentrysize final long position channel position memory position channel position position memory clear this resets to position limit to capacity channel read memory we read now again mb channel position position memory flip mark the buffer so it can be reset if necessary memory mark try read the length of the entry final int length memory getint if the buffer length is zero then return if length maxentrysize memory reset limit memory position nextentry null return read the checksum of the entry final long checksum memory getint compute the checksum for the entry bytes final checksum new update memory array memory position length if the stored checksum equals the computed checksum return the entry if checksum getvalue final int limit memory limit memory limit memory position length final e entry namespace deserialize memory memory limit limit nextentry new indexed index entry length else memory reset limit memory position nextentry null catch final bufferunderflowexception e memory reset limit memory position nextentry null catch final ioexception e throw new storageexception e this means we will read mb process from these mb until our remaining is less then maxentrysize so after we reached half of the buffer then we will read again where the other remaining is thrown away and will end at the beginning of the buffer because it is re read this problem is more an issue when the maxentrysize and realentrysize is very different which is our case we using to support big deployments but actually most of our records are less then kb if the sizes would be nearer then we would probably also read most of the buffer but this is not our use case so we should fix this
1
16,617
9,475,902,622
IssuesEvent
2019-04-19 13:16:05
JuliaOpt/JuMP.jl
https://api.github.com/repos/JuliaOpt/JuMP.jl
closed
Extreme performance degradation from v0.18.5 -> v0.19
bug performance
Heya JuMP team! I've currently noticed a large performance regression (which has been cited in a few cases, notably #1403 and #1905). I was planning on releasing some code for a paper, which was originally written with v0.18.5, but, with the upgrade, decided to do the necessary edits for v0.19. The code is essentially (sadly :( ) unusable in its current form, due to the speed degradation. Two MWEs showing a large difference (I will note that I'm using two slightly different environments, with Julia 1.0 and Julia 1.1) [Julia 1.0, JuMP v0.18.5](https://gist.github.com/angeris/8ac92dee79081c946979bcce8160f656). Time to complete: <2s. Roughly .0001s/constraint. [Julia 1.1, JuMP v0.19](https://gist.github.com/angeris/2f3ae08642c001d8f0ccf7aad03a6a8f). ETA for completing this (as reported by `ProgressMeter`): 18 minutes. Roughly 1.3s/constraint. The problem in question solved was much, much larger and would take several days to formulate at this pace. I'd also be happy to contribute to a fix, but I'd need to learn a bit more about profiling and performance in Julia (and definitely more about the current JuMP implementation) in order to help out! Anyways, thank you so much for this project, it's been super useful :)
True
Extreme performance degradation from v0.18.5 -> v0.19 - Heya JuMP team! I've currently noticed a large performance regression (which has been cited in a few cases, notably #1403 and #1905). I was planning on releasing some code for a paper, which was originally written with v0.18.5, but, with the upgrade, decided to do the necessary edits for v0.19. The code is essentially (sadly :( ) unusable in its current form, due to the speed degradation. Two MWEs showing a large difference (I will note that I'm using two slightly different environments, with Julia 1.0 and Julia 1.1) [Julia 1.0, JuMP v0.18.5](https://gist.github.com/angeris/8ac92dee79081c946979bcce8160f656). Time to complete: <2s. Roughly .0001s/constraint. [Julia 1.1, JuMP v0.19](https://gist.github.com/angeris/2f3ae08642c001d8f0ccf7aad03a6a8f). ETA for completing this (as reported by `ProgressMeter`): 18 minutes. Roughly 1.3s/constraint. The problem in question solved was much, much larger and would take several days to formulate at this pace. I'd also be happy to contribute to a fix, but I'd need to learn a bit more about profiling and performance in Julia (and definitely more about the current JuMP implementation) in order to help out! Anyways, thank you so much for this project, it's been super useful :)
perf
extreme performance degradation from heya jump team i ve currently noticed a large performance regression which has been cited in a few cases notably and i was planning on releasing some code for a paper which was originally written with but with the upgrade decided to do the necessary edits for the code is essentially sadly unusable in its current form due to the speed degradation two mwes showing a large difference i will note that i m using two slightly different environments with julia and julia time to complete roughly constraint eta for completing this as reported by progressmeter minutes roughly constraint the problem in question solved was much much larger and would take several days to formulate at this pace i d also be happy to contribute to a fix but i d need to learn a bit more about profiling and performance in julia and definitely more about the current jump implementation in order to help out anyways thank you so much for this project it s been super useful
1
186,056
6,733,153,079
IssuesEvent
2017-10-18 14:00:28
dwyl/library
https://api.github.com/repos/dwyl/library
opened
Create button component to do different things depending on state
enhancement priority-2
We have a button on the book component, this button should do different things depending on the state of the book/application: + If the book is in the library and is not on loan then it should say "Borrow book" + If it is in the library but is on loan it should say "Join queue" + If it is in the library and is on loan to the user viewing it it should say "Return" + If it is not in the library and the user is not an admin it should say "Request book" + If it is not in the library and the user is an admin it should say "Add book" The buttons should of course have relevant actions which fulfil the actions represented on them in text.
1.0
Create button component to do different things depending on state - We have a button on the book component, this button should do different things depending on the state of the book/application: + If the book is in the library and is not on loan then it should say "Borrow book" + If it is in the library but is on loan it should say "Join queue" + If it is in the library and is on loan to the user viewing it it should say "Return" + If it is not in the library and the user is not an admin it should say "Request book" + If it is not in the library and the user is an admin it should say "Add book" The buttons should of course have relevant actions which fulfil the actions represented on them in text.
non_perf
create button component to do different things depending on state we have a button on the book component this button should do different things depending on the state of the book application if the book is in the library and is not on loan then it should say borrow book if it is in the library but is on loan it should say join queue if it is in the library and is on loan to the user viewing it it should say return if it is not in the library and the user is not an admin it should say request book if it is not in the library and the user is an admin it should say add book the buttons should of course have relevant actions which fulfil the actions represented on them in text
0
6,377
3,796,715,523
IssuesEvent
2016-03-23 02:20:58
mapbox/mapbox-gl-native
https://api.github.com/repos/mapbox/mapbox-gl-native
closed
remove build-time dependency on llvm.org
build
This host has poor reliability, and when it goes down our builds fail.
1.0
remove build-time dependency on llvm.org - This host has poor reliability, and when it goes down our builds fail.
non_perf
remove build time dependency on llvm org this host has poor reliability and when it goes down our builds fail
0
300,756
25,993,454,863
IssuesEvent
2022-12-20 09:38:06
finos/waltz
https://api.github.com/repos/finos/waltz
closed
Prep 1.46
fixed (test & close) task
### Description - [x] mvn versions - [x] pacakge.json version - [x] prep liquibase version ### Resourcing We intend to implement this task
1.0
Prep 1.46 - ### Description - [x] mvn versions - [x] pacakge.json version - [x] prep liquibase version ### Resourcing We intend to implement this task
non_perf
prep description mvn versions pacakge json version prep liquibase version resourcing we intend to implement this task
0
382,010
26,481,759,612
IssuesEvent
2023-01-17 15:09:37
StaticJsCMS/static-cms
https://api.github.com/repos/StaticJsCMS/static-cms
closed
Create NetlifyCMS to StaticCMS Migration Guide
documentation
Document all steps necessary to migrate from NetlifyCMS to StaticCMS, covering as many scenarios as possible. - [ ] All deprecated features were removed - [ ] Some widget properties/options have changed - [ ] New markdown editor comes with new plugin interface, old editor components no longer work - [ ] Widget control properties have changed - [ ] List widget - Dropped support for the `field` property. A single field in the `fields` property does the same behavior. - [ ] Dropping Moment for date-fns - [ ] Passing a config to `.init()` will now completely override `config.yml` (they are not exclusive), instead of being merged with `config.yml` - [ ] CMS must be explicitly initiated now. - [ ] Dropped the following beta features from NetlifyCMS. - GraphQL support for GitHub and GitLab - Remark plugins (new markdown editor has its own plugin system) - [ ] Dropped support for AssetStore integration - [ ] Dropped support for Azure backend - [ ] `registerWidget` props have changed - [ ] `isValid` in widget control component replaced by `validator` function in `registerWidget` props - [ ] `createClass` is deprecated, functional components should be used instead. Basic react hooks are provided as globals now. - [ ] React 18.2.0
1.0
Create NetlifyCMS to StaticCMS Migration Guide - Document all steps necessary to migrate from NetlifyCMS to StaticCMS, covering as many scenarios as possible. - [ ] All deprecated features were removed - [ ] Some widget properties/options have changed - [ ] New markdown editor comes with new plugin interface, old editor components no longer work - [ ] Widget control properties have changed - [ ] List widget - Dropped support for the `field` property. A single field in the `fields` property does the same behavior. - [ ] Dropping Moment for date-fns - [ ] Passing a config to `.init()` will now completely override `config.yml` (they are not exclusive), instead of being merged with `config.yml` - [ ] CMS must be explicitly initiated now. - [ ] Dropped the following beta features from NetlifyCMS. - GraphQL support for GitHub and GitLab - Remark plugins (new markdown editor has its own plugin system) - [ ] Dropped support for AssetStore integration - [ ] Dropped support for Azure backend - [ ] `registerWidget` props have changed - [ ] `isValid` in widget control component replaced by `validator` function in `registerWidget` props - [ ] `createClass` is deprecated, functional components should be used instead. Basic react hooks are provided as globals now. - [ ] React 18.2.0
non_perf
create netlifycms to staticcms migration guide document all steps necessary to migrate from netlifycms to staticcms covering as many scenarios as possible all deprecated features were removed some widget properties options have changed new markdown editor comes with new plugin interface old editor components no longer work widget control properties have changed list widget dropped support for the field property a single field in the fields property does the same behavior dropping moment for date fns passing a config to init will now completely override config yml they are not exclusive instead of being merged with config yml cms must be explicitly initiated now dropped the following beta features from netlifycms graphql support for github and gitlab remark plugins new markdown editor has its own plugin system dropped support for assetstore integration dropped support for azure backend registerwidget props have changed isvalid in widget control component replaced by validator function in registerwidget props createclass is deprecated functional components should be used instead basic react hooks are provided as globals now react
0
142,042
19,024,488,951
IssuesEvent
2021-11-24 00:35:28
udsm-dhis2-project/dim-api
https://api.github.com/repos/udsm-dhis2-project/dim-api
closed
CVE-2021-3795 (Medium) detected in semver-regex-2.0.0.tgz
security vulnerability no-issue-activity
## CVE-2021-3795 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>semver-regex-2.0.0.tgz</b></p></summary> <p>Regular expression for matching semver versions</p> <p>Library home page: <a href="https://registry.npmjs.org/semver-regex/-/semver-regex-2.0.0.tgz">https://registry.npmjs.org/semver-regex/-/semver-regex-2.0.0.tgz</a></p> <p>Path to dependency file: dim-api/package.json</p> <p>Path to vulnerable library: dim-api/node_modules/semver-regex/package.json</p> <p> Dependency Hierarchy: - husky-4.2.5.tgz (Root Library) - find-versions-3.2.0.tgz - :x: **semver-regex-2.0.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> semver-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3795>CVE-2021-3795</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sindresorhus/semver-regex/releases/tag/v4.0.1">https://github.com/sindresorhus/semver-regex/releases/tag/v4.0.1</a></p> <p>Release Date: 2021-09-15</p> <p>Fix Resolution: semver-regex - 3.1.3,4.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3795 (Medium) detected in semver-regex-2.0.0.tgz - ## CVE-2021-3795 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>semver-regex-2.0.0.tgz</b></p></summary> <p>Regular expression for matching semver versions</p> <p>Library home page: <a href="https://registry.npmjs.org/semver-regex/-/semver-regex-2.0.0.tgz">https://registry.npmjs.org/semver-regex/-/semver-regex-2.0.0.tgz</a></p> <p>Path to dependency file: dim-api/package.json</p> <p>Path to vulnerable library: dim-api/node_modules/semver-regex/package.json</p> <p> Dependency Hierarchy: - husky-4.2.5.tgz (Root Library) - find-versions-3.2.0.tgz - :x: **semver-regex-2.0.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> semver-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3795>CVE-2021-3795</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sindresorhus/semver-regex/releases/tag/v4.0.1">https://github.com/sindresorhus/semver-regex/releases/tag/v4.0.1</a></p> <p>Release Date: 2021-09-15</p> <p>Fix Resolution: semver-regex - 3.1.3,4.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_perf
cve medium detected in semver regex tgz cve medium severity vulnerability vulnerable library semver regex tgz regular expression for matching semver versions library home page a href path to dependency file dim api package json path to vulnerable library dim api node modules semver regex package json dependency hierarchy husky tgz root library find versions tgz x semver regex tgz vulnerable library found in base branch master vulnerability details semver regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution semver regex step up your open source security game with whitesource
0
28,902
13,884,128,745
IssuesEvent
2020-10-18 14:56:33
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
BUG: GroupBy().fillna() performance regression
Groupby Missing-data Performance
- [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the latest version of pandas. - [ ] (optional) I have confirmed this bug exists on the master branch of pandas. --- ```python import pandas as pd import numpy as np N = 2000 df = pd.DataFrame({"A": [1] * N, "B": [np.nan, 1.0] * (N // 2)}) df = df.sort_values("A").set_index("A") df["B"] = df.groupby("A")["B"].fillna(method="ffill") ``` #### Problem description The groupby + fillna gets extremely slow increasing the N. This is a regression from 1.0.5->1.1.0. Note: if I remove the `.set_index("A")` it's fast again. #### Expected Output Same output, just faster. #### Output of ``pd.show_versions()`` <details> INSTALLED VERSIONS ------------------ commit : d9fff2792bf16178d4e450fe7384244e50635733 python : 3.7.8.final.0 python-bits : 64 OS : Linux OS-release : 4.4.110-1.el7.elrepo.x86_64 Version : #1 SMP Fri Jan 5 11:35:48 EST 2018 machine : x86_64 processor : x86_64 byteorder : little LC_ALL : None LANG : en_US.UTF-8 LOCALE : en_US.UTF-8 pandas : 1.1.0 numpy : 1.19.1 pytz : 2020.1 dateutil : 2.8.1 pip : 20.2.3 setuptools : 49.6.0.post20200917 Cython : None pytest : None hypothesis : None sphinx : None blosc : None feather : None xlsxwriter : None lxml.etree : None html5lib : None pymysql : None psycopg2 : None jinja2 : None IPython : None pandas_datareader: None bs4 : None bottleneck : None fsspec : None fastparquet : None gcsfs : None matplotlib : None numexpr : None odfpy : None openpyxl : None pandas_gbq : None pyarrow : None pytables : None pyxlsb : None s3fs : None scipy : None sqlalchemy : None tables : None tabulate : None xarray : None xlrd : None xlwt : None numba : None </details>
True
BUG: GroupBy().fillna() performance regression - - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the latest version of pandas. - [ ] (optional) I have confirmed this bug exists on the master branch of pandas. --- ```python import pandas as pd import numpy as np N = 2000 df = pd.DataFrame({"A": [1] * N, "B": [np.nan, 1.0] * (N // 2)}) df = df.sort_values("A").set_index("A") df["B"] = df.groupby("A")["B"].fillna(method="ffill") ``` #### Problem description The groupby + fillna gets extremely slow increasing the N. This is a regression from 1.0.5->1.1.0. Note: if I remove the `.set_index("A")` it's fast again. #### Expected Output Same output, just faster. #### Output of ``pd.show_versions()`` <details> INSTALLED VERSIONS ------------------ commit : d9fff2792bf16178d4e450fe7384244e50635733 python : 3.7.8.final.0 python-bits : 64 OS : Linux OS-release : 4.4.110-1.el7.elrepo.x86_64 Version : #1 SMP Fri Jan 5 11:35:48 EST 2018 machine : x86_64 processor : x86_64 byteorder : little LC_ALL : None LANG : en_US.UTF-8 LOCALE : en_US.UTF-8 pandas : 1.1.0 numpy : 1.19.1 pytz : 2020.1 dateutil : 2.8.1 pip : 20.2.3 setuptools : 49.6.0.post20200917 Cython : None pytest : None hypothesis : None sphinx : None blosc : None feather : None xlsxwriter : None lxml.etree : None html5lib : None pymysql : None psycopg2 : None jinja2 : None IPython : None pandas_datareader: None bs4 : None bottleneck : None fsspec : None fastparquet : None gcsfs : None matplotlib : None numexpr : None odfpy : None openpyxl : None pandas_gbq : None pyarrow : None pytables : None pyxlsb : None s3fs : None scipy : None sqlalchemy : None tables : None tabulate : None xarray : None xlrd : None xlwt : None numba : None </details>
perf
bug groupby fillna performance regression i have checked that this issue has not already been reported i have confirmed this bug exists on the latest version of pandas optional i have confirmed this bug exists on the master branch of pandas python import pandas as pd import numpy as np n df pd dataframe a n b n df df sort values a set index a df df groupby a fillna method ffill problem description the groupby fillna gets extremely slow increasing the n this is a regression from note if i remove the set index a it s fast again expected output same output just faster output of pd show versions installed versions commit python final python bits os linux os release elrepo version smp fri jan est machine processor byteorder little lc all none lang en us utf locale en us utf pandas numpy pytz dateutil pip setuptools cython none pytest none hypothesis none sphinx none blosc none feather none xlsxwriter none lxml etree none none pymysql none none none ipython none pandas datareader none none bottleneck none fsspec none fastparquet none gcsfs none matplotlib none numexpr none odfpy none openpyxl none pandas gbq none pyarrow none pytables none pyxlsb none none scipy none sqlalchemy none tables none tabulate none xarray none xlrd none xlwt none numba none
1
19,040
10,300,096,313
IssuesEvent
2019-08-28 12:09:07
ArkEcosystem/desktop-wallet
https://api.github.com/repos/ArkEcosystem/desktop-wallet
closed
Change Peer function slow compared to Mobile Wallet
Status: Stale Type: Performance
When using the desktop wallet, I have noticed that change peer (AKA "connect to best") is slow or doesn't do anything when clicked. My peer latencies on desktop wallet are also often in the red. The mobile wallet always does changes peers for me lightning fast and without an issue (and 99% of the time, the mobile wallet stays connected as intended without the need to manually change). Also, specifying a custom peer, like when I actually type in the peer I see in the mobile wallet into the desktop wallet, returns an error of not being able to connect on the desktop wallet. So I don't understand why my desktop wallet is having so much trouble with peers compared to the mobile wallet
True
Change Peer function slow compared to Mobile Wallet - When using the desktop wallet, I have noticed that change peer (AKA "connect to best") is slow or doesn't do anything when clicked. My peer latencies on desktop wallet are also often in the red. The mobile wallet always does changes peers for me lightning fast and without an issue (and 99% of the time, the mobile wallet stays connected as intended without the need to manually change). Also, specifying a custom peer, like when I actually type in the peer I see in the mobile wallet into the desktop wallet, returns an error of not being able to connect on the desktop wallet. So I don't understand why my desktop wallet is having so much trouble with peers compared to the mobile wallet
perf
change peer function slow compared to mobile wallet when using the desktop wallet i have noticed that change peer aka connect to best is slow or doesn t do anything when clicked my peer latencies on desktop wallet are also often in the red the mobile wallet always does changes peers for me lightning fast and without an issue and of the time the mobile wallet stays connected as intended without the need to manually change also specifying a custom peer like when i actually type in the peer i see in the mobile wallet into the desktop wallet returns an error of not being able to connect on the desktop wallet so i don t understand why my desktop wallet is having so much trouble with peers compared to the mobile wallet
1
102,849
16,590,754,951
IssuesEvent
2021-06-01 07:25:01
Yoavmartin/vulnerable-node
https://api.github.com/repos/Yoavmartin/vulnerable-node
closed
CVE-2020-8116 (High) detected in opennmsopennms-source-24.1.2-1 - autoclosed
security vulnerability
## CVE-2020-8116 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-24.1.2-1</b></p></summary> <p> <p>A Java based fault and performance management system</p> <p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p> <p>Found in HEAD commit: <a href="https://github.com/Yoavmartin/vulnerable-node/commit/c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c">c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in dot-prop npm package versions before 4.2.1 and versions 5.x before 5.1.1 allows an attacker to add arbitrary properties to JavaScript language constructs such as objects. <p>Publish Date: 2020-02-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p> <p>Release Date: 2020-02-04</p> <p>Fix Resolution: dot-prop - 5.1.1</p> </p> </details> <p></p>
True
CVE-2020-8116 (High) detected in opennmsopennms-source-24.1.2-1 - autoclosed - ## CVE-2020-8116 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-24.1.2-1</b></p></summary> <p> <p>A Java based fault and performance management system</p> <p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p> <p>Found in HEAD commit: <a href="https://github.com/Yoavmartin/vulnerable-node/commit/c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c">c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in dot-prop npm package versions before 4.2.1 and versions 5.x before 5.1.1 allows an attacker to add arbitrary properties to JavaScript language constructs such as objects. <p>Publish Date: 2020-02-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p> <p>Release Date: 2020-02-04</p> <p>Fix Resolution: dot-prop - 5.1.1</p> </p> </details> <p></p>
non_perf
cve high detected in opennmsopennms source autoclosed cve high severity vulnerability vulnerable library opennmsopennms source a java based fault and performance management system library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details prototype pollution vulnerability in dot prop npm package versions before and versions x before allows an attacker to add arbitrary properties to javascript language constructs such as objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution dot prop
0
308,637
9,441,338,708
IssuesEvent
2019-04-15 00:55:45
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Solution for building Docker images from Jupyter
area/jupyter help wanted priority/p2
We need a solution for building docker images from Jupyter. I think a common use case will be developing/prototyping code in JupyterHub and then wanting to package that code up in a Docker image and either run a job (e.g. a TF training job) or deploy a model (e.g. via Seldon). There are lots of ways to build Docker images - bazel - docker build - GCB - [Repo2Docker](https://github.com/jupyter/repo2docker) part of JupyterHub's binder project - Argo What's the difference between Repo2Docker and using Argo to pull the code from Git and then run a build?
1.0
Solution for building Docker images from Jupyter - We need a solution for building docker images from Jupyter. I think a common use case will be developing/prototyping code in JupyterHub and then wanting to package that code up in a Docker image and either run a job (e.g. a TF training job) or deploy a model (e.g. via Seldon). There are lots of ways to build Docker images - bazel - docker build - GCB - [Repo2Docker](https://github.com/jupyter/repo2docker) part of JupyterHub's binder project - Argo What's the difference between Repo2Docker and using Argo to pull the code from Git and then run a build?
non_perf
solution for building docker images from jupyter we need a solution for building docker images from jupyter i think a common use case will be developing prototyping code in jupyterhub and then wanting to package that code up in a docker image and either run a job e g a tf training job or deploy a model e g via seldon there are lots of ways to build docker images bazel docker build gcb part of jupyterhub s binder project argo what s the difference between and using argo to pull the code from git and then run a build
0
1,737
3,089,681,432
IssuesEvent
2015-08-25 22:58:38
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Why does the speed of kubelet running pause container have slowed down?
area/performance kind/support team/node
Today I raise a 2 minions cluster and create a 50 replica redis-controller-master rc, the rc take about 10 minutes to launch, that's much longer than one week ago ( about 50 seconds). I see the kubelet's log and found it take 10 seconds to pulled a pause images and another 10 seconds to start it: ![default](https://cloud.githubusercontent.com/assets/13072423/9002760/76950efa-379a-11e5-8c0a-3672758b02dd.PNG) And after all the pause containers has started, it only takes several seconds to started all the redis containers: ![default](https://cloud.githubusercontent.com/assets/13072423/9002791/bae11d74-379a-11e5-8857-4d25bc1810de.PNG) I thought the behavior of starting redis containers is reasonable and that about pause container is unreasonable, why it?
True
Why does the speed of kubelet running pause container have slowed down? - Today I raise a 2 minions cluster and create a 50 replica redis-controller-master rc, the rc take about 10 minutes to launch, that's much longer than one week ago ( about 50 seconds). I see the kubelet's log and found it take 10 seconds to pulled a pause images and another 10 seconds to start it: ![default](https://cloud.githubusercontent.com/assets/13072423/9002760/76950efa-379a-11e5-8c0a-3672758b02dd.PNG) And after all the pause containers has started, it only takes several seconds to started all the redis containers: ![default](https://cloud.githubusercontent.com/assets/13072423/9002791/bae11d74-379a-11e5-8857-4d25bc1810de.PNG) I thought the behavior of starting redis containers is reasonable and that about pause container is unreasonable, why it?
perf
why does the speed of kubelet running pause container have slowed down today i raise a minions cluster and create a replica redis controller master rc the rc take about minutes to launch that s much longer than one week ago about seconds i see the kubelet s log and found it take seconds to pulled a pause images and another seconds to start it and after all the pause containers has started it only takes several seconds to started all the redis containers i thought the behavior of starting redis containers is reasonable and that about pause container is unreasonable why it
1
7,835
6,260,378,918
IssuesEvent
2017-07-14 20:30:19
KazDragon/terminalpp
https://api.github.com/repos/KazDragon/terminalpp
closed
Add forwarding references to lambda visitor
Improvement in progress Performance
Althought it's supposed to take lambdas, it might also take moveable function objects, so adding forwarding references makes sense.
True
Add forwarding references to lambda visitor - Althought it's supposed to take lambdas, it might also take moveable function objects, so adding forwarding references makes sense.
perf
add forwarding references to lambda visitor althought it s supposed to take lambdas it might also take moveable function objects so adding forwarding references makes sense
1
279,727
30,733,971,743
IssuesEvent
2023-07-28 05:49:46
DevSeconsTL/WebGoat-WhiteSource-Bolt
https://api.github.com/repos/DevSeconsTL/WebGoat-WhiteSource-Bolt
opened
commons-io-1.3.2.jar: 1 vulnerabilities (highest severity is: 4.8)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-io-1.3.2.jar</b></p></summary> <p>Commons-IO contains utility classes, stream implementations, file filters, and endian classes.</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-io/commons-io/1.3.2/commons-io-1.3.2.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/DevSeconsTL/WebGoat-WhiteSource-Bolt/commit/a4b4e21bf0133636e16df71363955c6ce17b0791">a4b4e21bf0133636e16df71363955c6ce17b0791</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (commons-io version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-29425](https://www.mend.io/vulnerability-database/CVE-2021-29425) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 4.8 | commons-io-1.3.2.jar | Direct | 2.7 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-29425</summary> ### Vulnerable Library - <b>commons-io-1.3.2.jar</b></p> <p>Commons-IO contains utility classes, stream implementations, file filters, and endian classes.</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-io/commons-io/1.3.2/commons-io-1.3.2.jar</p> <p> Dependency Hierarchy: - :x: **commons-io-1.3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DevSeconsTL/WebGoat-WhiteSource-Bolt/commit/a4b4e21bf0133636e16df71363955c6ce17b0791">a4b4e21bf0133636e16df71363955c6ce17b0791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Commons IO before 2.7, When invoking the method FileNameUtils.normalize with an improper input string, like "//../foo", or "\\..\foo", the result would be the same value, thus possibly providing access to files in the parent directory, but not further above (thus "limited" path traversal), if the calling code would use the result to construct a path value. <p>Publish Date: 2021-04-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29425>CVE-2021-29425</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>4.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-29425">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-29425</a></p> <p>Release Date: 2021-04-13</p> <p>Fix Resolution: 2.7</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
commons-io-1.3.2.jar: 1 vulnerabilities (highest severity is: 4.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-io-1.3.2.jar</b></p></summary> <p>Commons-IO contains utility classes, stream implementations, file filters, and endian classes.</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-io/commons-io/1.3.2/commons-io-1.3.2.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/DevSeconsTL/WebGoat-WhiteSource-Bolt/commit/a4b4e21bf0133636e16df71363955c6ce17b0791">a4b4e21bf0133636e16df71363955c6ce17b0791</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (commons-io version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-29425](https://www.mend.io/vulnerability-database/CVE-2021-29425) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 4.8 | commons-io-1.3.2.jar | Direct | 2.7 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-29425</summary> ### Vulnerable Library - <b>commons-io-1.3.2.jar</b></p> <p>Commons-IO contains utility classes, stream implementations, file filters, and endian classes.</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-io/commons-io/1.3.2/commons-io-1.3.2.jar</p> <p> Dependency Hierarchy: - :x: **commons-io-1.3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DevSeconsTL/WebGoat-WhiteSource-Bolt/commit/a4b4e21bf0133636e16df71363955c6ce17b0791">a4b4e21bf0133636e16df71363955c6ce17b0791</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Commons IO before 2.7, When invoking the method FileNameUtils.normalize with an improper input string, like "//../foo", or "\\..\foo", the result would be the same value, thus possibly providing access to files in the parent directory, but not further above (thus "limited" path traversal), if the calling code would use the result to construct a path value. <p>Publish Date: 2021-04-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29425>CVE-2021-29425</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>4.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-29425">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-29425</a></p> <p>Release Date: 2021-04-13</p> <p>Fix Resolution: 2.7</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_perf
commons io jar vulnerabilities highest severity is vulnerable library commons io jar commons io contains utility classes stream implementations file filters and endian classes path to dependency file pom xml path to vulnerable library home wss scanner repository commons io commons io commons io jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in commons io version remediation available medium commons io jar direct details cve vulnerable library commons io jar commons io contains utility classes stream implementations file filters and endian classes path to dependency file pom xml path to vulnerable library home wss scanner repository commons io commons io commons io jar dependency hierarchy x commons io jar vulnerable library found in head commit a href found in base branch master vulnerability details in apache commons io before when invoking the method filenameutils normalize with an improper input string like foo or foo the result would be the same value thus possibly providing access to files in the parent directory but not further above thus limited path traversal if the calling code would use the result to construct a path value publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
25,951
7,763,198,148
IssuesEvent
2018-06-01 15:46:20
apache/incubator-openwhisk-release
https://api.github.com/repos/apache/incubator-openwhisk-release
opened
Remove wskdeploy specification PDFs (archives) and diagram from source release
Release build priority-low
The following folders and their contents for the wskdeploy repo. should be excluded from the TAR file produced for the source code release: - specification/archive - specification/diagrams This would mean (binary) files of the following format/extensions would be excluded: - .pdf - .pptx - .jpg
1.0
Remove wskdeploy specification PDFs (archives) and diagram from source release - The following folders and their contents for the wskdeploy repo. should be excluded from the TAR file produced for the source code release: - specification/archive - specification/diagrams This would mean (binary) files of the following format/extensions would be excluded: - .pdf - .pptx - .jpg
non_perf
remove wskdeploy specification pdfs archives and diagram from source release the following folders and their contents for the wskdeploy repo should be excluded from the tar file produced for the source code release specification archive specification diagrams this would mean binary files of the following format extensions would be excluded pdf pptx jpg
0
11,185
7,460,943,233
IssuesEvent
2018-03-30 22:14:10
zfsonlinux/zfs
https://api.github.com/repos/zfsonlinux/zfs
closed
Large synchronous writes are slow when a slog is present
Performance
Note that this issue seems to impact all ZFS implementations, not just ZFS On Linux. ZFS uses a complicated process when it comes to deciding whether a write should be logged in indirect mode (written once by the DMU, the log records store a pointer) or in immediate mode (written in the log record, rewritten later by the DMU). Basically, it goes like this: - Write in indirect mode to the data vdevs if: - `logbias=throughput`, or - There is no slog and the write is larger than `zfs_immediate_write_sz`. - Write in immediate mode to the data vdevs if `logbias=latency` and: - There is no slog and the write is smaller than `zfs_immediate_write_sz`, or - There is a slog and the total commit size if larger than `zil_slog_limit`. - Write in immediate mode to the slog vdevs if `logbias=latency`, there is a slog, and the total commit size is smaller than `zil_slog_limit`. The decision to use indirect or immediate mode is implemented in `zfs_log_write()` and `zvol_log_write()`. The decision to use the slog or the normal vdevs is implemented in the `USE_SLOG()` macro used by `zil_lwb_write_start`. The issue is, this decision process makes sense except for one particularly painful edge case, when these conditions are all true: - `logbias=latency`, and - There is a slog, and - There are large writes in the ZIL to be commited (e.g. > 100 MB). In this situation, the optimal choice would be to write to the normal pool in indirect mode, which should give us the minimum latency considering this is a large sequential write. Indeed, for very large writes, you don't want to use immediate mode because it means writing the data twice. Even if you write the log records to the slog, this will be slower with most pool configurations with e.g. lots of spindles and one SSD slog because the aggregate sequential write throughput of all the spindles is usually greater than the SSD's. Instead, the algorithm makes the worst decision possible: it writes the data in immediate mode to the main data disks. This means that all the (large) data will be commited as ZIL log records on the data disks first, then immediately after, it will get written _again_ by the DMU. This means the overall throughput is halved, and if this is a sustained load, the ZIL commit latency will be doubled compared to indirect mode. It is shockingly easy to reproduce this issue. In pseudo-code: ``` open(file) write(file, lots of data) // e.g. 2 GB fsync(file) ``` Watch the `zil_stats` kstat page when that runs. If you don't have a slog in your pool, then the `fsync()` call will complete in roughly the time it takes to write 2 GB sequentially to your main disks. This is optimal. If you have a slog in your pool, then the `fsync()` call will generate twice as much write activity, and will write up to 4 GB to your main disks. Ironically, the slog won't be used at all when that happens. The solution would be to modify the algorithm `zfs_log_write()` and `zvol_log_write()` so that, in the conditions mentioned above, it switches to indirect writes when the commit size reaches a certain threshold (e.g. 32 MB). I would gladly write a patch, but I won't have the time to do it, so I'm just leaving the result of my research here in case anyone's interested. If anyone wants to write the patch, it should be very simple to implement it.
True
Large synchronous writes are slow when a slog is present - Note that this issue seems to impact all ZFS implementations, not just ZFS On Linux. ZFS uses a complicated process when it comes to deciding whether a write should be logged in indirect mode (written once by the DMU, the log records store a pointer) or in immediate mode (written in the log record, rewritten later by the DMU). Basically, it goes like this: - Write in indirect mode to the data vdevs if: - `logbias=throughput`, or - There is no slog and the write is larger than `zfs_immediate_write_sz`. - Write in immediate mode to the data vdevs if `logbias=latency` and: - There is no slog and the write is smaller than `zfs_immediate_write_sz`, or - There is a slog and the total commit size if larger than `zil_slog_limit`. - Write in immediate mode to the slog vdevs if `logbias=latency`, there is a slog, and the total commit size is smaller than `zil_slog_limit`. The decision to use indirect or immediate mode is implemented in `zfs_log_write()` and `zvol_log_write()`. The decision to use the slog or the normal vdevs is implemented in the `USE_SLOG()` macro used by `zil_lwb_write_start`. The issue is, this decision process makes sense except for one particularly painful edge case, when these conditions are all true: - `logbias=latency`, and - There is a slog, and - There are large writes in the ZIL to be commited (e.g. > 100 MB). In this situation, the optimal choice would be to write to the normal pool in indirect mode, which should give us the minimum latency considering this is a large sequential write. Indeed, for very large writes, you don't want to use immediate mode because it means writing the data twice. Even if you write the log records to the slog, this will be slower with most pool configurations with e.g. lots of spindles and one SSD slog because the aggregate sequential write throughput of all the spindles is usually greater than the SSD's. Instead, the algorithm makes the worst decision possible: it writes the data in immediate mode to the main data disks. This means that all the (large) data will be commited as ZIL log records on the data disks first, then immediately after, it will get written _again_ by the DMU. This means the overall throughput is halved, and if this is a sustained load, the ZIL commit latency will be doubled compared to indirect mode. It is shockingly easy to reproduce this issue. In pseudo-code: ``` open(file) write(file, lots of data) // e.g. 2 GB fsync(file) ``` Watch the `zil_stats` kstat page when that runs. If you don't have a slog in your pool, then the `fsync()` call will complete in roughly the time it takes to write 2 GB sequentially to your main disks. This is optimal. If you have a slog in your pool, then the `fsync()` call will generate twice as much write activity, and will write up to 4 GB to your main disks. Ironically, the slog won't be used at all when that happens. The solution would be to modify the algorithm `zfs_log_write()` and `zvol_log_write()` so that, in the conditions mentioned above, it switches to indirect writes when the commit size reaches a certain threshold (e.g. 32 MB). I would gladly write a patch, but I won't have the time to do it, so I'm just leaving the result of my research here in case anyone's interested. If anyone wants to write the patch, it should be very simple to implement it.
perf
large synchronous writes are slow when a slog is present note that this issue seems to impact all zfs implementations not just zfs on linux zfs uses a complicated process when it comes to deciding whether a write should be logged in indirect mode written once by the dmu the log records store a pointer or in immediate mode written in the log record rewritten later by the dmu basically it goes like this write in indirect mode to the data vdevs if logbias throughput or there is no slog and the write is larger than zfs immediate write sz write in immediate mode to the data vdevs if logbias latency and there is no slog and the write is smaller than zfs immediate write sz or there is a slog and the total commit size if larger than zil slog limit write in immediate mode to the slog vdevs if logbias latency there is a slog and the total commit size is smaller than zil slog limit the decision to use indirect or immediate mode is implemented in zfs log write and zvol log write the decision to use the slog or the normal vdevs is implemented in the use slog macro used by zil lwb write start the issue is this decision process makes sense except for one particularly painful edge case when these conditions are all true logbias latency and there is a slog and there are large writes in the zil to be commited e g mb in this situation the optimal choice would be to write to the normal pool in indirect mode which should give us the minimum latency considering this is a large sequential write indeed for very large writes you don t want to use immediate mode because it means writing the data twice even if you write the log records to the slog this will be slower with most pool configurations with e g lots of spindles and one ssd slog because the aggregate sequential write throughput of all the spindles is usually greater than the ssd s instead the algorithm makes the worst decision possible it writes the data in immediate mode to the main data disks this means that all the large data will be commited as zil log records on the data disks first then immediately after it will get written again by the dmu this means the overall throughput is halved and if this is a sustained load the zil commit latency will be doubled compared to indirect mode it is shockingly easy to reproduce this issue in pseudo code open file write file lots of data e g gb fsync file watch the zil stats kstat page when that runs if you don t have a slog in your pool then the fsync call will complete in roughly the time it takes to write gb sequentially to your main disks this is optimal if you have a slog in your pool then the fsync call will generate twice as much write activity and will write up to gb to your main disks ironically the slog won t be used at all when that happens the solution would be to modify the algorithm zfs log write and zvol log write so that in the conditions mentioned above it switches to indirect writes when the commit size reaches a certain threshold e g mb i would gladly write a patch but i won t have the time to do it so i m just leaving the result of my research here in case anyone s interested if anyone wants to write the patch it should be very simple to implement it
1
222,818
24,711,335,768
IssuesEvent
2022-10-20 01:14:34
Lob2018/YannLobjois_6_25032021
https://api.github.com/repos/Lob2018/YannLobjois_6_25032021
opened
CVE-2022-37603 (High) detected in loader-utils-2.0.0.tgz
security vulnerability
## CVE-2022-37603 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-2.0.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/css-loader/node_modules/loader-utils/package.json,/node_modules/style-loader/node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - style-loader-2.0.0.tgz (Root Library) - :x: **loader-utils-2.0.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js. <p>Publish Date: 2022-10-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37603>CVE-2022-37603</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-14</p> <p>Fix Resolution (loader-utils): 2.0.1</p> <p>Direct dependency fix Resolution (style-loader): 3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-37603 (High) detected in loader-utils-2.0.0.tgz - ## CVE-2022-37603 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-2.0.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/css-loader/node_modules/loader-utils/package.json,/node_modules/style-loader/node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - style-loader-2.0.0.tgz (Root Library) - :x: **loader-utils-2.0.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js. <p>Publish Date: 2022-10-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37603>CVE-2022-37603</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-14</p> <p>Fix Resolution (loader-utils): 2.0.1</p> <p>Direct dependency fix Resolution (style-loader): 3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_perf
cve high detected in loader utils tgz cve high severity vulnerability vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules css loader node modules loader utils package json node modules style loader node modules loader utils package json dependency hierarchy style loader tgz root library x loader utils tgz vulnerable library found in base branch main vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the url variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution loader utils direct dependency fix resolution style loader step up your open source security game with mend
0
1,297
2,931,188,117
IssuesEvent
2015-06-29 10:42:00
mapbox/mapbox-gl-native
https://api.github.com/repos/mapbox/mapbox-gl-native
closed
All polylines flicker when adding adding/removing a polyline (iOS)
bug iOS performance
I see an old issue (#1712) referencing flickering of annotations that's marked as closed for 0.4, but the issue still exists. Steps to reproduce: - Add a number of polylines to a map when it's first made - Add a new polyline or remove an existing one triggered by a user action http://f.cl.ly/items/0w1e2K2w3N2q3Y2K0V1x/Flicker.m4v
True
All polylines flicker when adding adding/removing a polyline (iOS) - I see an old issue (#1712) referencing flickering of annotations that's marked as closed for 0.4, but the issue still exists. Steps to reproduce: - Add a number of polylines to a map when it's first made - Add a new polyline or remove an existing one triggered by a user action http://f.cl.ly/items/0w1e2K2w3N2q3Y2K0V1x/Flicker.m4v
perf
all polylines flicker when adding adding removing a polyline ios i see an old issue referencing flickering of annotations that s marked as closed for but the issue still exists steps to reproduce add a number of polylines to a map when it s first made add a new polyline or remove an existing one triggered by a user action
1
386,387
11,437,879,599
IssuesEvent
2020-02-05 01:25:37
zulip/zulip
https://api.github.com/repos/zulip/zulip
opened
Follow-ups to addition of new stream_post_policy feature
area: stream settings priority: blocker
Copied from https://github.com/zulip/zulip/pull/13768#issuecomment-582196283: There's some follow-ups we should definitely do: * [ ] Once this is deployed to chat.zulip.org (very soon), test to make sure the mobile app can still create streams and still provides a good error message posting to streams with posting restrictions. * [ ] I think it'd be helpful for `stream_types.hbs` to have a link to the /help/ documentation for what a "full member" is, since it's not clear from context. This can use the same styling we use in the "settings" pages that define these users. * [ ] The real-time synchronization for changes to this field does not propagate to other browsers' display. @pragatiagrawal31 can you take this part since you're actively working on that component? * [ ] The `test_messages` testing approach cleanups I suggested to use `do_change_` functions rather than raw `stream.save()` etc. (did you see them? I'm concerned GitHub may have lost my comment). * [ ] Investigate further improving our migrations as described here: https://github.com/zulip/zulip/pull/13768/files/3bec321bf9e32751ddd62c18763c4b6cb00f107e#diff-52fee2c5eed6104b21eefefd0d2ef76e * [ ] Reworking the `create_stream` and `ajaxSubscribeForCreation` interface as we discussed above to not involve N positional arguments?
1.0
Follow-ups to addition of new stream_post_policy feature - Copied from https://github.com/zulip/zulip/pull/13768#issuecomment-582196283: There's some follow-ups we should definitely do: * [ ] Once this is deployed to chat.zulip.org (very soon), test to make sure the mobile app can still create streams and still provides a good error message posting to streams with posting restrictions. * [ ] I think it'd be helpful for `stream_types.hbs` to have a link to the /help/ documentation for what a "full member" is, since it's not clear from context. This can use the same styling we use in the "settings" pages that define these users. * [ ] The real-time synchronization for changes to this field does not propagate to other browsers' display. @pragatiagrawal31 can you take this part since you're actively working on that component? * [ ] The `test_messages` testing approach cleanups I suggested to use `do_change_` functions rather than raw `stream.save()` etc. (did you see them? I'm concerned GitHub may have lost my comment). * [ ] Investigate further improving our migrations as described here: https://github.com/zulip/zulip/pull/13768/files/3bec321bf9e32751ddd62c18763c4b6cb00f107e#diff-52fee2c5eed6104b21eefefd0d2ef76e * [ ] Reworking the `create_stream` and `ajaxSubscribeForCreation` interface as we discussed above to not involve N positional arguments?
non_perf
follow ups to addition of new stream post policy feature copied from there s some follow ups we should definitely do once this is deployed to chat zulip org very soon test to make sure the mobile app can still create streams and still provides a good error message posting to streams with posting restrictions i think it d be helpful for stream types hbs to have a link to the help documentation for what a full member is since it s not clear from context this can use the same styling we use in the settings pages that define these users the real time synchronization for changes to this field does not propagate to other browsers display can you take this part since you re actively working on that component the test messages testing approach cleanups i suggested to use do change functions rather than raw stream save etc did you see them i m concerned github may have lost my comment investigate further improving our migrations as described here reworking the create stream and ajaxsubscribeforcreation interface as we discussed above to not involve n positional arguments
0
45,507
24,084,716,302
IssuesEvent
2022-09-19 09:52:14
iterative/dvc
https://api.github.com/repos/iterative/dvc
closed
regression: dvc init is slow
p2-medium performance regression
The 3 major issue seems to be: - [x] `tpi` (see https://github.com/iterative/tpi/issues/14) (~300ms) - [ ] `fsspec` callback (~200ms) - [x] prompts import on experiments from `rich` (for `dvc exp init`) (~50ms) #6738 - [ ] [`dvc.fs` importing fsspec by default](https://github.com/iterative/dvc/blob/612696cf4ae01885a89269b5d26b4631380614b8/dvc/fs/azure.py#L5) (~100ms) (There may not be much point in making fsspec lazy, as we'll need local filesystem anyway). The following is a startup log for `dvc init`. <img width="1107" alt="Screen Shot 2021-09-29 at 17 52 16" src="https://user-images.githubusercontent.com/18718008/135265276-7f9279b5-8610-497a-8f59-654a7a859870.png"> Following is the benchmark for `dvc init`. ```console $ hyperfine "dvc init" --warmup 3 -p "rm -rf .dvc" Benchmark #1: dvc init Time (mean ± σ): 1.311 s ± 0.065 s [User: 1.066 s, System: 0.203 s] Range (min … max): 1.230 s … 1.435 s 10 runs ``` `dvc init` crossing 1 sec is really bad, this was the result a month ago (before tpi and prompts, and with fsspec.Callback made lazy): ```console $ hyperfine "dvc init" --warmup 3 -p "rm -rf .dvc" Benchmark #1: dvc init Time (mean ± σ): 506.8 ms ± 31.4 ms [User: 369.2 ms, System: 103.7 ms] Range (min … max): 477.4 ms … 566.6 ms 10 runs ```
True
regression: dvc init is slow - The 3 major issue seems to be: - [x] `tpi` (see https://github.com/iterative/tpi/issues/14) (~300ms) - [ ] `fsspec` callback (~200ms) - [x] prompts import on experiments from `rich` (for `dvc exp init`) (~50ms) #6738 - [ ] [`dvc.fs` importing fsspec by default](https://github.com/iterative/dvc/blob/612696cf4ae01885a89269b5d26b4631380614b8/dvc/fs/azure.py#L5) (~100ms) (There may not be much point in making fsspec lazy, as we'll need local filesystem anyway). The following is a startup log for `dvc init`. <img width="1107" alt="Screen Shot 2021-09-29 at 17 52 16" src="https://user-images.githubusercontent.com/18718008/135265276-7f9279b5-8610-497a-8f59-654a7a859870.png"> Following is the benchmark for `dvc init`. ```console $ hyperfine "dvc init" --warmup 3 -p "rm -rf .dvc" Benchmark #1: dvc init Time (mean ± σ): 1.311 s ± 0.065 s [User: 1.066 s, System: 0.203 s] Range (min … max): 1.230 s … 1.435 s 10 runs ``` `dvc init` crossing 1 sec is really bad, this was the result a month ago (before tpi and prompts, and with fsspec.Callback made lazy): ```console $ hyperfine "dvc init" --warmup 3 -p "rm -rf .dvc" Benchmark #1: dvc init Time (mean ± σ): 506.8 ms ± 31.4 ms [User: 369.2 ms, System: 103.7 ms] Range (min … max): 477.4 ms … 566.6 ms 10 runs ```
perf
regression dvc init is slow the major issue seems to be tpi see fsspec callback prompts import on experiments from rich for dvc exp init there may not be much point in making fsspec lazy as we ll need local filesystem anyway the following is a startup log for dvc init img width alt screen shot at src following is the benchmark for dvc init console hyperfine dvc init warmup p rm rf dvc benchmark dvc init time mean ± σ s ± s range min … max s … s runs dvc init crossing sec is really bad this was the result a month ago before tpi and prompts and with fsspec callback made lazy console hyperfine dvc init warmup p rm rf dvc benchmark dvc init time mean ± σ ms ± ms range min … max ms … ms runs
1
254,912
19,277,021,646
IssuesEvent
2021-12-10 13:05:51
Djordje261/github-slideshow
https://api.github.com/repos/Djordje261/github-slideshow
closed
Getting Started with GitHub
documentation
# :wave: Bienvenido a "Introducción a GitHub" de GitHub Learning Lab Para empezar, te guiaré a través de unos primeros pasos que son importantes para programar y colaborar en GitHub. :point_down: _Esta flecha indica que puedes expandir la ventana! Haz clic en ellas durante el curso para más información._ <details><summary>¿Qué es GitHub?</summary> <hr> ## ¿Qué es GitHub? ¡Me alegra que lo hayas preguntado! Muchas personas usan GitHub porque quieren contribuir a proyectos de código abierto <sup>[:book:](https://help.github.com/articles/github-glossary/#open-source)</sup>, o porque son invitados por compañeros de trabajo o de clase que usan GitHub en sus proyectos. ¿Por qué la gente usa GitHub en sus proyectos? **En esencia, GitHub es una plataforma de colaboración.** Desde software hasta documentos legales, puedes contar con GitHub para ayudarte a hacer tu mejor trabajo con las herramientas de colaboración y seguridad que tu equipo necesita. Con GitHub, puedes mantener proyectos completamente privados, invitar al mundo a colaborar y optimizar cada paso de tu proyecto. **GitHub también es una poderosa herramienta de versión de control.** GitHub usa Git <sup>[:book:](https://help.github.com/articles/github-glossary/#git)</sup>, el software de control de versión más popular para rastrear cada contribución y colaborador <sup>[:book:](https://help.github.com/articles/github-glossary/#contributor)</sup> de tu proyecto--de esta manera sabrás exactamente de dónde vino cada línea de código de tu proyecto. **GitHub ayuda a las personas a hacer mucho más.** La gente usa GitHub para construir varias de las tecnologías más avanzadas del mundo. Ya sea que estes visualizando datos o construyendo un nuevo juego, hay una comunidad entera y un set de herramientas en GitHub que te puedan llevar al siguiente nivel. Este curso empieza con lo básico pero ¡profundizaremos el resto después! :tv: [Video: ¿Qué es GitHub?](https://www.youtube.com/watch?v=w3jLJU7DT5E) <hr> </details><br> <details><summary>Explorando un repositorio de GitHub</summary> <hr> ## Explorando un repositorio de GitHub :tv: [Video: Explorando un repositorio](https://www.youtube.com/watch?v=R8OAwrcMlRw) ### Más características El video cubre algunas de las características usadas con mayor frecuencia. Aquí hay otras cosas que puedes encontrar en los repositorios de GitHub: - Tableros para el proyecto: Crear un tablero estilo Kanban para seguir tus tareas dentro de GitHub - Wiki: Crea y guarda documentación relevante acerca de tu proyecto. - Estadísticas: Ve un menú desplegable que contiene enlaces a herramientas de análisis para tu repositorio, que incluyen: - Pulso: Encuentra información sobre el trabajo que ha sido completado y el trabajo que está en progreso en este panel del proyecto. - Gráficas: Las gráficas proporcionan una vista más granular de la actividad del repositorio, incluyendo quién contribuyó al repositorio, quién lo bifurcó y cuándo completaron el trabajo. ### Archivos especiales En el video aprendiste sobre un archivo especial llamado README.md. Aquí hay una lista de algunos otros archivos especiales que puedes añadir a tus repositorios: - CONTRIBUTING.md: El archivo `CONTRIBUTING.md` es usado para describir el proceso para contribuir al repositorio. Un link al archivo `CONTRIBUTING.md` es mostrado cada vez que alguien crea una nueva propuesta o una solicitud de extracción. - ISSUE_TEMPLATE.md: El archivo `ISSUE_TEMPLATE.md` lo puedes usar para rellenar el cuerpo de una propuesta. Por ejemplo, si siempre necesitas el mismo tipo de información para reportar errores puedes incluirlo en la plantilla de tu propuesta y cada nueva propuesta será abierta con el texto que has recomendado para empezar. <hr> </details> ### Usando las propuestas (o _issues_) Esto es una propuesta (o _issue_) <sup>[:book:](https://help.github.com/articles/github-glossary/#issue)</sup>: un lugar en el que puedes tener conversaciones sobre los errores en tu código, revisiones de código y casi cualquier otra cosas que te puedas imaginar. Los títulos de tus propuestas son como la línea de asunto de un email. Avisan de un vistazo a tus colaboradores de qué trata tu propuesta. Por ejemplo, el título de esta propuesta es Getting Started with GitHub. <details><summary>Usando GitHub Issues (propuestas)</summary> ## Usando propuestas de GitHub Las propuestas se usan para discutir ideas, mejoras, tareas y errores. Facilitan la colaboración al: - Proveer a todos (incluso miembros futuros) la historia completa en un solo lugar - Permitirte entrelazar tu propuesta con otras propuestas o una solicitud de extracción (o _pull request_) <sup>[:book:](https://help.github.com/articles/github-glossary/#pull-request)</sup> - Crear un registro único y completo de cómo y por qué tomaste ciertas decisiones - Permitirte atraer fácilmente a las personas y los equipos adecuados a una conversación con @mencion :tv: [Video: Usando propuestas](https://www.youtube.com/watch?v=Zhj46r5D0nQ) <hr> </details> <details><summary>Administrando notificaciones</summary> <hr> ## Administrando notificaciones :tv: [Video: Observando, notificaciones, estrellas y explorar](https://www.youtube.com/watch?v=ocQldxF7fMY) Una vez que hayas comentado en una propuesta o solicitud de extracción, empezarás a recibir notificaciones por email cada vez que haya actividad en tu hilo. ## Cómo silenciar o dejar de silenciar conversaciones en específico. 1. Ve a la propuesta o solicitud de extracción 2. Bajo _"Notifications"_, haz clic en el botón de **Unsubscribe** a la derecha para silenciar notificaciones o **Subscribe** para dejar de silenciarlas. Verás una descripción corta que explica tu estado actual de notificaciones. ### Cómo personalizar notificaciones en Settings 1. Haz clic en el icono de tu perfil 2. Haz clic en **Settings** 3. Haz clic en **Notifications** desde el menú a tu izquierda y [ajusta tus preferencias de notificaciones](https://help.github.com/articles/managing-notification-delivery-methods/) ### Opciones de notificación de un repositorio * **Watch**: Recibirás una notificación cuando una nueva propuesta, solicitud de extracción o comentario ha sido publicado y también cuando una propuesta ha sido cerrado o una solicitud de extracción ha sido fusionada. * **Not watching**: No vas a recibir notificaciones al menos que te hayan mencionado usando @ * **Ignore**: No vas a recibir ningún tipo de notificación sobre este respoitorio ## Cómo revisar las notificaciones de los repositorios que estás observando 1. Haz clic en tu icono de perfil 2. Haz clic en **Settings** 3. Haz clic en **Notification** desde el menú a tu izquierda 4. Haz clic en el link de [repositorios que estás observando](https://github.com/watching) 5. Selecciona la pestaña **Watching** 6. Haz clic en el botón de **Unwatch** para inhabilitar las notificaciones o en **Watch** para habilitarlas. <hr> </details> <hr> <h3 align="center">Sigue leyendo a continuación para encontrar tu primera tarea</h3>
1.0
Getting Started with GitHub - # :wave: Bienvenido a "Introducción a GitHub" de GitHub Learning Lab Para empezar, te guiaré a través de unos primeros pasos que son importantes para programar y colaborar en GitHub. :point_down: _Esta flecha indica que puedes expandir la ventana! Haz clic en ellas durante el curso para más información._ <details><summary>¿Qué es GitHub?</summary> <hr> ## ¿Qué es GitHub? ¡Me alegra que lo hayas preguntado! Muchas personas usan GitHub porque quieren contribuir a proyectos de código abierto <sup>[:book:](https://help.github.com/articles/github-glossary/#open-source)</sup>, o porque son invitados por compañeros de trabajo o de clase que usan GitHub en sus proyectos. ¿Por qué la gente usa GitHub en sus proyectos? **En esencia, GitHub es una plataforma de colaboración.** Desde software hasta documentos legales, puedes contar con GitHub para ayudarte a hacer tu mejor trabajo con las herramientas de colaboración y seguridad que tu equipo necesita. Con GitHub, puedes mantener proyectos completamente privados, invitar al mundo a colaborar y optimizar cada paso de tu proyecto. **GitHub también es una poderosa herramienta de versión de control.** GitHub usa Git <sup>[:book:](https://help.github.com/articles/github-glossary/#git)</sup>, el software de control de versión más popular para rastrear cada contribución y colaborador <sup>[:book:](https://help.github.com/articles/github-glossary/#contributor)</sup> de tu proyecto--de esta manera sabrás exactamente de dónde vino cada línea de código de tu proyecto. **GitHub ayuda a las personas a hacer mucho más.** La gente usa GitHub para construir varias de las tecnologías más avanzadas del mundo. Ya sea que estes visualizando datos o construyendo un nuevo juego, hay una comunidad entera y un set de herramientas en GitHub que te puedan llevar al siguiente nivel. Este curso empieza con lo básico pero ¡profundizaremos el resto después! :tv: [Video: ¿Qué es GitHub?](https://www.youtube.com/watch?v=w3jLJU7DT5E) <hr> </details><br> <details><summary>Explorando un repositorio de GitHub</summary> <hr> ## Explorando un repositorio de GitHub :tv: [Video: Explorando un repositorio](https://www.youtube.com/watch?v=R8OAwrcMlRw) ### Más características El video cubre algunas de las características usadas con mayor frecuencia. Aquí hay otras cosas que puedes encontrar en los repositorios de GitHub: - Tableros para el proyecto: Crear un tablero estilo Kanban para seguir tus tareas dentro de GitHub - Wiki: Crea y guarda documentación relevante acerca de tu proyecto. - Estadísticas: Ve un menú desplegable que contiene enlaces a herramientas de análisis para tu repositorio, que incluyen: - Pulso: Encuentra información sobre el trabajo que ha sido completado y el trabajo que está en progreso en este panel del proyecto. - Gráficas: Las gráficas proporcionan una vista más granular de la actividad del repositorio, incluyendo quién contribuyó al repositorio, quién lo bifurcó y cuándo completaron el trabajo. ### Archivos especiales En el video aprendiste sobre un archivo especial llamado README.md. Aquí hay una lista de algunos otros archivos especiales que puedes añadir a tus repositorios: - CONTRIBUTING.md: El archivo `CONTRIBUTING.md` es usado para describir el proceso para contribuir al repositorio. Un link al archivo `CONTRIBUTING.md` es mostrado cada vez que alguien crea una nueva propuesta o una solicitud de extracción. - ISSUE_TEMPLATE.md: El archivo `ISSUE_TEMPLATE.md` lo puedes usar para rellenar el cuerpo de una propuesta. Por ejemplo, si siempre necesitas el mismo tipo de información para reportar errores puedes incluirlo en la plantilla de tu propuesta y cada nueva propuesta será abierta con el texto que has recomendado para empezar. <hr> </details> ### Usando las propuestas (o _issues_) Esto es una propuesta (o _issue_) <sup>[:book:](https://help.github.com/articles/github-glossary/#issue)</sup>: un lugar en el que puedes tener conversaciones sobre los errores en tu código, revisiones de código y casi cualquier otra cosas que te puedas imaginar. Los títulos de tus propuestas son como la línea de asunto de un email. Avisan de un vistazo a tus colaboradores de qué trata tu propuesta. Por ejemplo, el título de esta propuesta es Getting Started with GitHub. <details><summary>Usando GitHub Issues (propuestas)</summary> ## Usando propuestas de GitHub Las propuestas se usan para discutir ideas, mejoras, tareas y errores. Facilitan la colaboración al: - Proveer a todos (incluso miembros futuros) la historia completa en un solo lugar - Permitirte entrelazar tu propuesta con otras propuestas o una solicitud de extracción (o _pull request_) <sup>[:book:](https://help.github.com/articles/github-glossary/#pull-request)</sup> - Crear un registro único y completo de cómo y por qué tomaste ciertas decisiones - Permitirte atraer fácilmente a las personas y los equipos adecuados a una conversación con @mencion :tv: [Video: Usando propuestas](https://www.youtube.com/watch?v=Zhj46r5D0nQ) <hr> </details> <details><summary>Administrando notificaciones</summary> <hr> ## Administrando notificaciones :tv: [Video: Observando, notificaciones, estrellas y explorar](https://www.youtube.com/watch?v=ocQldxF7fMY) Una vez que hayas comentado en una propuesta o solicitud de extracción, empezarás a recibir notificaciones por email cada vez que haya actividad en tu hilo. ## Cómo silenciar o dejar de silenciar conversaciones en específico. 1. Ve a la propuesta o solicitud de extracción 2. Bajo _"Notifications"_, haz clic en el botón de **Unsubscribe** a la derecha para silenciar notificaciones o **Subscribe** para dejar de silenciarlas. Verás una descripción corta que explica tu estado actual de notificaciones. ### Cómo personalizar notificaciones en Settings 1. Haz clic en el icono de tu perfil 2. Haz clic en **Settings** 3. Haz clic en **Notifications** desde el menú a tu izquierda y [ajusta tus preferencias de notificaciones](https://help.github.com/articles/managing-notification-delivery-methods/) ### Opciones de notificación de un repositorio * **Watch**: Recibirás una notificación cuando una nueva propuesta, solicitud de extracción o comentario ha sido publicado y también cuando una propuesta ha sido cerrado o una solicitud de extracción ha sido fusionada. * **Not watching**: No vas a recibir notificaciones al menos que te hayan mencionado usando @ * **Ignore**: No vas a recibir ningún tipo de notificación sobre este respoitorio ## Cómo revisar las notificaciones de los repositorios que estás observando 1. Haz clic en tu icono de perfil 2. Haz clic en **Settings** 3. Haz clic en **Notification** desde el menú a tu izquierda 4. Haz clic en el link de [repositorios que estás observando](https://github.com/watching) 5. Selecciona la pestaña **Watching** 6. Haz clic en el botón de **Unwatch** para inhabilitar las notificaciones o en **Watch** para habilitarlas. <hr> </details> <hr> <h3 align="center">Sigue leyendo a continuación para encontrar tu primera tarea</h3>
non_perf
getting started with github wave bienvenido a introducción a github de github learning lab para empezar te guiaré a través de unos primeros pasos que son importantes para programar y colaborar en github point down esta flecha indica que puedes expandir la ventana haz clic en ellas durante el curso para más información ¿qué es github ¿qué es github ¡me alegra que lo hayas preguntado muchas personas usan github porque quieren contribuir a proyectos de código abierto o porque son invitados por compañeros de trabajo o de clase que usan github en sus proyectos ¿por qué la gente usa github en sus proyectos en esencia github es una plataforma de colaboración desde software hasta documentos legales puedes contar con github para ayudarte a hacer tu mejor trabajo con las herramientas de colaboración y seguridad que tu equipo necesita con github puedes mantener proyectos completamente privados invitar al mundo a colaborar y optimizar cada paso de tu proyecto github también es una poderosa herramienta de versión de control github usa git el software de control de versión más popular para rastrear cada contribución y colaborador de tu proyecto de esta manera sabrás exactamente de dónde vino cada línea de código de tu proyecto github ayuda a las personas a hacer mucho más la gente usa github para construir varias de las tecnologías más avanzadas del mundo ya sea que estes visualizando datos o construyendo un nuevo juego hay una comunidad entera y un set de herramientas en github que te puedan llevar al siguiente nivel este curso empieza con lo básico pero ¡profundizaremos el resto después tv explorando un repositorio de github explorando un repositorio de github tv más características el video cubre algunas de las características usadas con mayor frecuencia aquí hay otras cosas que puedes encontrar en los repositorios de github tableros para el proyecto crear un tablero estilo kanban para seguir tus tareas dentro de github wiki crea y guarda documentación relevante acerca de tu proyecto estadísticas ve un menú desplegable que contiene enlaces a herramientas de análisis para tu repositorio que incluyen pulso encuentra información sobre el trabajo que ha sido completado y el trabajo que está en progreso en este panel del proyecto gráficas las gráficas proporcionan una vista más granular de la actividad del repositorio incluyendo quién contribuyó al repositorio quién lo bifurcó y cuándo completaron el trabajo archivos especiales en el video aprendiste sobre un archivo especial llamado readme md aquí hay una lista de algunos otros archivos especiales que puedes añadir a tus repositorios contributing md el archivo contributing md es usado para describir el proceso para contribuir al repositorio un link al archivo contributing md es mostrado cada vez que alguien crea una nueva propuesta o una solicitud de extracción issue template md el archivo issue template md lo puedes usar para rellenar el cuerpo de una propuesta por ejemplo si siempre necesitas el mismo tipo de información para reportar errores puedes incluirlo en la plantilla de tu propuesta y cada nueva propuesta será abierta con el texto que has recomendado para empezar usando las propuestas o issues esto es una propuesta o issue un lugar en el que puedes tener conversaciones sobre los errores en tu código revisiones de código y casi cualquier otra cosas que te puedas imaginar los títulos de tus propuestas son como la línea de asunto de un email avisan de un vistazo a tus colaboradores de qué trata tu propuesta por ejemplo el título de esta propuesta es getting started with github usando github issues propuestas usando propuestas de github las propuestas se usan para discutir ideas mejoras tareas y errores facilitan la colaboración al proveer a todos incluso miembros futuros la historia completa en un solo lugar permitirte entrelazar tu propuesta con otras propuestas o una solicitud de extracción o pull request crear un registro único y completo de cómo y por qué tomaste ciertas decisiones permitirte atraer fácilmente a las personas y los equipos adecuados a una conversación con mencion tv administrando notificaciones administrando notificaciones tv una vez que hayas comentado en una propuesta o solicitud de extracción empezarás a recibir notificaciones por email cada vez que haya actividad en tu hilo cómo silenciar o dejar de silenciar conversaciones en específico ve a la propuesta o solicitud de extracción bajo notifications haz clic en el botón de unsubscribe a la derecha para silenciar notificaciones o subscribe para dejar de silenciarlas verás una descripción corta que explica tu estado actual de notificaciones cómo personalizar notificaciones en settings haz clic en el icono de tu perfil haz clic en settings haz clic en notifications desde el menú a tu izquierda y opciones de notificación de un repositorio watch recibirás una notificación cuando una nueva propuesta solicitud de extracción o comentario ha sido publicado y también cuando una propuesta ha sido cerrado o una solicitud de extracción ha sido fusionada not watching no vas a recibir notificaciones al menos que te hayan mencionado usando ignore no vas a recibir ningún tipo de notificación sobre este respoitorio cómo revisar las notificaciones de los repositorios que estás observando haz clic en tu icono de perfil haz clic en settings haz clic en notification desde el menú a tu izquierda haz clic en el link de selecciona la pestaña watching haz clic en el botón de unwatch para inhabilitar las notificaciones o en watch para habilitarlas sigue leyendo a continuación para encontrar tu primera tarea
0
22,179
11,502,662,263
IssuesEvent
2020-02-12 19:32:13
influxdata/influxdb
https://api.github.com/repos/influxdata/influxdb
closed
investigate why cli is so slow
area/cli performance
with the influxd locally. compare the cli time and api time. have 2 orgs curl http://localhost:9999/api/v2/orgs took 0.02 secs. run ``` influx org find|gnomon ``` got ``` 2.5534s ID Name 0.0006s 60b45ab5d3826209 org2 0.0011s b3ef3f1a0e3cebc0 org1 0.0002s Total 2.5562s ``` almost all command took at least 2.5 sec. exp: `influx -h|gnomon` ``` 2.7557s Influx Client 0.0014s 0.0001s Usage: 0.0001s influx [flags] 0.0001s influx [command] 0.0001s 0.0001s Available Commands: 0.0001s auth Authorization management commands 0.0001s bucket Bucket management commands 0.0001s help Help about any command 0.0001s org Organization management commands 0.0001s ping Check the InfluxDB /health endpoint 0.0001s query Execute a Flux query 0.0001s repl Interactive REPL (read-eval-print-loop) 0.0001s setup Setup instance with initial user, org, bucket 0.0001s task Task management commands 0.0001s user User management commands 0.0001s write Write points to InfluxDB 0.0000s 0.0000s Flags: 0.0000s -h, --help Help for the influx command 0.0000s --host string HTTP address of Influx (default "http://loca lhost:9999") 0.0000s -t, --token string API token to be used throughout client calls 0.0010s Use "influx [command] --help" for more information about a command. 0.0001s Use "influx [command] --help" for more information about a command. Total 2.7628s ```
True
investigate why cli is so slow - with the influxd locally. compare the cli time and api time. have 2 orgs curl http://localhost:9999/api/v2/orgs took 0.02 secs. run ``` influx org find|gnomon ``` got ``` 2.5534s ID Name 0.0006s 60b45ab5d3826209 org2 0.0011s b3ef3f1a0e3cebc0 org1 0.0002s Total 2.5562s ``` almost all command took at least 2.5 sec. exp: `influx -h|gnomon` ``` 2.7557s Influx Client 0.0014s 0.0001s Usage: 0.0001s influx [flags] 0.0001s influx [command] 0.0001s 0.0001s Available Commands: 0.0001s auth Authorization management commands 0.0001s bucket Bucket management commands 0.0001s help Help about any command 0.0001s org Organization management commands 0.0001s ping Check the InfluxDB /health endpoint 0.0001s query Execute a Flux query 0.0001s repl Interactive REPL (read-eval-print-loop) 0.0001s setup Setup instance with initial user, org, bucket 0.0001s task Task management commands 0.0001s user User management commands 0.0001s write Write points to InfluxDB 0.0000s 0.0000s Flags: 0.0000s -h, --help Help for the influx command 0.0000s --host string HTTP address of Influx (default "http://loca lhost:9999") 0.0000s -t, --token string API token to be used throughout client calls 0.0010s Use "influx [command] --help" for more information about a command. 0.0001s Use "influx [command] --help" for more information about a command. Total 2.7628s ```
perf
investigate why cli is so slow with the influxd locally compare the cli time and api time have orgs curl took secs run influx org find gnomon got id name total almost all command took at least sec exp influx h gnomon influx client usage influx influx available commands auth authorization management commands bucket bucket management commands help help about any command org organization management commands ping check the influxdb health endpoint query execute a flux query repl interactive repl read eval print loop setup setup instance with initial user org bucket task task management commands user user management commands write write points to influxdb flags h help help for the influx command host string http address of influx default lhost t token string api token to be used throughout client calls use influx help for more information about a command use influx help for more information about a command total
1
448,375
31,791,092,011
IssuesEvent
2023-09-13 03:25:40
networkx/networkx
https://api.github.com/repos/networkx/networkx
opened
Some examples in tournament.py don't use tournament graphs
type: Documentation
Most `tournament` functions assume the input graph is a tournament. And the docs say the user must check whether it is a tournament. But some of the example don't use tournament graphs. The docs say that the function's output is undefined when the graph is not a tournament graph. So, while the examples do produce the output they claim to produce, it is a meaningless result. For example: `tournament.is_strongly_connected(G)` returns False even though the graph in the example is strongly connected. ```python >>> G = nx.DiGraph([(0, 1), (0, 2), (0, 3), (1, 2), (1, 3), (2, 3), (3, 0)]) >>> nx.tournament.is_strongly_connected(G) True >>> G.remove_edge(1, 3) >>> nx.tournament.is_strongly_connected(G) # this graph is strongly connected. False ``` The function `is_strongly_connected` returns `False` even though the graph G is strongly connected. This is because both examples used are not tournament graphs (exactly one directed edge between each pair of nodes). I think every example in this module should add a line checking if the graph is a tournament before running the function. There may be others that are not valid example uses of the function. And those examples will reinforce the good practice of checking each time a graph is changed.
1.0
Some examples in tournament.py don't use tournament graphs - Most `tournament` functions assume the input graph is a tournament. And the docs say the user must check whether it is a tournament. But some of the example don't use tournament graphs. The docs say that the function's output is undefined when the graph is not a tournament graph. So, while the examples do produce the output they claim to produce, it is a meaningless result. For example: `tournament.is_strongly_connected(G)` returns False even though the graph in the example is strongly connected. ```python >>> G = nx.DiGraph([(0, 1), (0, 2), (0, 3), (1, 2), (1, 3), (2, 3), (3, 0)]) >>> nx.tournament.is_strongly_connected(G) True >>> G.remove_edge(1, 3) >>> nx.tournament.is_strongly_connected(G) # this graph is strongly connected. False ``` The function `is_strongly_connected` returns `False` even though the graph G is strongly connected. This is because both examples used are not tournament graphs (exactly one directed edge between each pair of nodes). I think every example in this module should add a line checking if the graph is a tournament before running the function. There may be others that are not valid example uses of the function. And those examples will reinforce the good practice of checking each time a graph is changed.
non_perf
some examples in tournament py don t use tournament graphs most tournament functions assume the input graph is a tournament and the docs say the user must check whether it is a tournament but some of the example don t use tournament graphs the docs say that the function s output is undefined when the graph is not a tournament graph so while the examples do produce the output they claim to produce it is a meaningless result for example tournament is strongly connected g returns false even though the graph in the example is strongly connected python g nx digraph nx tournament is strongly connected g true g remove edge nx tournament is strongly connected g this graph is strongly connected false the function is strongly connected returns false even though the graph g is strongly connected this is because both examples used are not tournament graphs exactly one directed edge between each pair of nodes i think every example in this module should add a line checking if the graph is a tournament before running the function there may be others that are not valid example uses of the function and those examples will reinforce the good practice of checking each time a graph is changed
0
7,919
6,301,749,950
IssuesEvent
2017-07-21 09:00:50
akvo/akvo-flow
https://api.github.com/repos/akvo/akvo-flow
reopened
Improve loading time of device list
Improving performance Ready for release
I noticed on a partner dashboard that the Device list is loading all the devices ever connected to the dashboard (with some being last connected in 2013) <img width="1184" alt="screen shot 2017-04-11 at 16 20 56" src="https://cloud.githubusercontent.com/assets/12456965/24913898/e50c0ab6-1ed2-11e7-878d-5e172c466dca.png"> Considering the user can organise the Device list based on all the header items, I think we need to reconsider how many entries we show in the list. Secondly, are we showing only the last time device connected to the instance or are we showing every time a device connected to the instance?
True
Improve loading time of device list - I noticed on a partner dashboard that the Device list is loading all the devices ever connected to the dashboard (with some being last connected in 2013) <img width="1184" alt="screen shot 2017-04-11 at 16 20 56" src="https://cloud.githubusercontent.com/assets/12456965/24913898/e50c0ab6-1ed2-11e7-878d-5e172c466dca.png"> Considering the user can organise the Device list based on all the header items, I think we need to reconsider how many entries we show in the list. Secondly, are we showing only the last time device connected to the instance or are we showing every time a device connected to the instance?
perf
improve loading time of device list i noticed on a partner dashboard that the device list is loading all the devices ever connected to the dashboard with some being last connected in img width alt screen shot at src considering the user can organise the device list based on all the header items i think we need to reconsider how many entries we show in the list secondly are we showing only the last time device connected to the instance or are we showing every time a device connected to the instance
1
46,931
7,295,757,005
IssuesEvent
2018-02-26 08:23:43
laravel-enso/VueDatatable
https://api.github.com/repos/laravel-enso/VueDatatable
closed
Missing dependencies
documentation
Hi there and thank you for the component. Just a heads-up that using this component independant of laravel/enso project needs a lot more setup than described in the readme. Don't know if it's in scope, but I'll try to resume here what needs to be done: 1. In the documentation, on step 4 there is a Toastr import that is not in the assets for this package. I assume that this package should have laravel-enso/vuecomponents as a dependency, since there is where the Toastr component lives. 2. There are a few npm dependencies missing for the missing icons if using the example provided: - @ fortawesome/fontawesome - @ fortawesome/fontawesome-free-solid - @ fortawesome/fontawesome-free-brands This icons need to be imported at some point during the app like this: import { faUpload } from '@fortawesome/fontawesome-free-solid/shakable.es'; fontawesome.library.add([faUpload]); This has to be done for any icons that will be used in the datatables (or referenced by table json that gets loaded into the controller). 3. @ fortawesome/vue-fontawesome needs to be installed with npm/yarn, imported in the bootstrap file and the component registered globally: ``` import FontAwesomeIcon from '@fortawesome/vue-fontawesome'; Vue.component('fa', FontAwesomeIcon); ``` Hope it helps someone.
1.0
Missing dependencies - Hi there and thank you for the component. Just a heads-up that using this component independant of laravel/enso project needs a lot more setup than described in the readme. Don't know if it's in scope, but I'll try to resume here what needs to be done: 1. In the documentation, on step 4 there is a Toastr import that is not in the assets for this package. I assume that this package should have laravel-enso/vuecomponents as a dependency, since there is where the Toastr component lives. 2. There are a few npm dependencies missing for the missing icons if using the example provided: - @ fortawesome/fontawesome - @ fortawesome/fontawesome-free-solid - @ fortawesome/fontawesome-free-brands This icons need to be imported at some point during the app like this: import { faUpload } from '@fortawesome/fontawesome-free-solid/shakable.es'; fontawesome.library.add([faUpload]); This has to be done for any icons that will be used in the datatables (or referenced by table json that gets loaded into the controller). 3. @ fortawesome/vue-fontawesome needs to be installed with npm/yarn, imported in the bootstrap file and the component registered globally: ``` import FontAwesomeIcon from '@fortawesome/vue-fontawesome'; Vue.component('fa', FontAwesomeIcon); ``` Hope it helps someone.
non_perf
missing dependencies hi there and thank you for the component just a heads up that using this component independant of laravel enso project needs a lot more setup than described in the readme don t know if it s in scope but i ll try to resume here what needs to be done in the documentation on step there is a toastr import that is not in the assets for this package i assume that this package should have laravel enso vuecomponents as a dependency since there is where the toastr component lives there are a few npm dependencies missing for the missing icons if using the example provided fortawesome fontawesome fortawesome fontawesome free solid fortawesome fontawesome free brands this icons need to be imported at some point during the app like this import faupload from fortawesome fontawesome free solid shakable es fontawesome library add this has to be done for any icons that will be used in the datatables or referenced by table json that gets loaded into the controller fortawesome vue fontawesome needs to be installed with npm yarn imported in the bootstrap file and the component registered globally import fontawesomeicon from fortawesome vue fontawesome vue component fa fontawesomeicon hope it helps someone
0
109,086
16,828,286,102
IssuesEvent
2021-06-17 22:03:37
jtimberlake/rl-bakery
https://api.github.com/repos/jtimberlake/rl-bakery
opened
CVE-2021-33503 (High) detected in urllib3-1.25.11-py2.py3-none-any.whl
security vulnerability
## CVE-2021-33503 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>urllib3-1.25.11-py2.py3-none-any.whl</b></p></summary> <p>HTTP library with thread-safe connection pooling, file post, and more.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl</a></p> <p>Path to dependency file: rl-bakery/requirements.txt</p> <p>Path to vulnerable library: rl-bakery/requirements.txt,rl-bakery</p> <p> Dependency Hierarchy: - moto-1.3.14-py2.py3-none-any.whl (Root Library) - boto3-1.12.18-py2.py3-none-any.whl - botocore-1.15.49-py2.py3-none-any.whl - :x: **urllib3-1.25.11-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/rl-bakery/commit/d91469bc8923d4e9b3605580b1f374632d029ad0">d91469bc8923d4e9b3605580b1f374632d029ad0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A security issue has been found in python-urllib3 before version 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL was passed as a parameter or redirected to via an HTTP redirect. <p>Publish Date: 2021-05-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503>CVE-2021-33503</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg">https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg</a></p> <p>Release Date: 2021-05-22</p> <p>Fix Resolution: urllib3 - 1.26.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"urllib3","packageVersion":"1.25.11","packageFilePaths":["/requirements.txt","rl-bakery"],"isTransitiveDependency":true,"dependencyTree":"moto:1.3.14;boto3:1.12.18;botocore:1.15.49;urllib3:1.25.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"urllib3 - 1.26.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33503","vulnerabilityDetails":"A security issue has been found in python-urllib3 before version 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL was passed as a parameter or redirected to via an HTTP redirect.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-33503 (High) detected in urllib3-1.25.11-py2.py3-none-any.whl - ## CVE-2021-33503 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>urllib3-1.25.11-py2.py3-none-any.whl</b></p></summary> <p>HTTP library with thread-safe connection pooling, file post, and more.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl</a></p> <p>Path to dependency file: rl-bakery/requirements.txt</p> <p>Path to vulnerable library: rl-bakery/requirements.txt,rl-bakery</p> <p> Dependency Hierarchy: - moto-1.3.14-py2.py3-none-any.whl (Root Library) - boto3-1.12.18-py2.py3-none-any.whl - botocore-1.15.49-py2.py3-none-any.whl - :x: **urllib3-1.25.11-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/rl-bakery/commit/d91469bc8923d4e9b3605580b1f374632d029ad0">d91469bc8923d4e9b3605580b1f374632d029ad0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A security issue has been found in python-urllib3 before version 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL was passed as a parameter or redirected to via an HTTP redirect. <p>Publish Date: 2021-05-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503>CVE-2021-33503</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg">https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg</a></p> <p>Release Date: 2021-05-22</p> <p>Fix Resolution: urllib3 - 1.26.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"urllib3","packageVersion":"1.25.11","packageFilePaths":["/requirements.txt","rl-bakery"],"isTransitiveDependency":true,"dependencyTree":"moto:1.3.14;boto3:1.12.18;botocore:1.15.49;urllib3:1.25.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"urllib3 - 1.26.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33503","vulnerabilityDetails":"A security issue has been found in python-urllib3 before version 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL was passed as a parameter or redirected to via an HTTP redirect.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_perf
cve high detected in none any whl cve high severity vulnerability vulnerable library none any whl http library with thread safe connection pooling file post and more library home page a href path to dependency file rl bakery requirements txt path to vulnerable library rl bakery requirements txt rl bakery dependency hierarchy moto none any whl root library none any whl botocore none any whl x none any whl vulnerable library found in head commit a href found in base branch master vulnerability details a security issue has been found in python before version when provided with a url containing many characters in the authority component the authority regular expression exhibits catastrophic backtracking causing a denial of service if a url was passed as a parameter or redirected to via an http redirect publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree moto botocore isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails a security issue has been found in python before version when provided with a url containing many characters in the authority component the authority regular expression exhibits catastrophic backtracking causing a denial of service if a url was passed as a parameter or redirected to via an http redirect vulnerabilityurl
0
51,535
27,125,630,965
IssuesEvent
2023-02-16 04:54:42
keras-team/keras
https://api.github.com/repos/keras-team/keras
closed
tf.keras.layers.Normalization(invert=True) does not restore from save
type:bug/performance stat:awaiting response from contributor
**System information**. - Have I written custom code: yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 22.04 - TensorFlow installed from (source or binary): tf-nightly from pip - TensorFlow version (use command below): 2.13.0-dev20230212 - Python version: 3.10.6 - Bazel version (if compiling from source): - GPU model and memory: CPU/N/A - Exact command to reproduce: run the code below: **Describe the problem**. `keras.models.load_model()` does not load `tf.keras.layers.Normalization` layer with `invert=True`. **Describe the current behavior**. It seems to omit `invert=True` flag when loading (or saving) the model. **Describe the expected behavior**. original trained and saved/loaded model output the same when using `Normalization` with `invert=True`. **Standalone code to reproduce the issue**. ```python import numpy as np import pandas as pd import sys import tensorflow as tf from tensorflow import keras from tensorflow.keras import layers print(tf.__version__) dataset = pd.read_csv('https://raw.githubusercontent.com/mzakharo/micropython/tubby/ports/esp32/orp2.csv') train_dataset = dataset.sample(frac=0.9, random_state=0) train_features = train_dataset.copy() train_labels = train_features.pop('orp') inorm = tf.keras.layers.Normalization(axis=-1, input_shape=[2, ]) inorm.adapt(np.array(train_features)) onorm = tf.keras.layers.Normalization(axis=-1, invert=True) onorm.adapt(train_labels) def build_and_compile_model(inorm, onorm): model = keras.Sequential([ inorm, layers.Dense(10, activation='relu'), layers.Dense(1), onorm, ]) model.compile(loss='mean_squared_error', optimizer='adam') return model model = build_and_compile_model(inorm, onorm) history = model.fit( train_features, train_labels, validation_split=0.1, verbose=0, epochs=10) fname = 'model.h5' model.save(fname) p0 = 7.6 p1 = 0.85 v = model.predict([(p0, p1)], verbose=0)[0] print('original output: ', v) lmodel = keras.models.load_model(fname) v = lmodel.predict([(p0, p1)], verbose=0)[0] print('restored output: ', v) ``` **logs**. ``` original output: [667.36383] restored output: [-8.889455] ```
True
tf.keras.layers.Normalization(invert=True) does not restore from save - **System information**. - Have I written custom code: yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 22.04 - TensorFlow installed from (source or binary): tf-nightly from pip - TensorFlow version (use command below): 2.13.0-dev20230212 - Python version: 3.10.6 - Bazel version (if compiling from source): - GPU model and memory: CPU/N/A - Exact command to reproduce: run the code below: **Describe the problem**. `keras.models.load_model()` does not load `tf.keras.layers.Normalization` layer with `invert=True`. **Describe the current behavior**. It seems to omit `invert=True` flag when loading (or saving) the model. **Describe the expected behavior**. original trained and saved/loaded model output the same when using `Normalization` with `invert=True`. **Standalone code to reproduce the issue**. ```python import numpy as np import pandas as pd import sys import tensorflow as tf from tensorflow import keras from tensorflow.keras import layers print(tf.__version__) dataset = pd.read_csv('https://raw.githubusercontent.com/mzakharo/micropython/tubby/ports/esp32/orp2.csv') train_dataset = dataset.sample(frac=0.9, random_state=0) train_features = train_dataset.copy() train_labels = train_features.pop('orp') inorm = tf.keras.layers.Normalization(axis=-1, input_shape=[2, ]) inorm.adapt(np.array(train_features)) onorm = tf.keras.layers.Normalization(axis=-1, invert=True) onorm.adapt(train_labels) def build_and_compile_model(inorm, onorm): model = keras.Sequential([ inorm, layers.Dense(10, activation='relu'), layers.Dense(1), onorm, ]) model.compile(loss='mean_squared_error', optimizer='adam') return model model = build_and_compile_model(inorm, onorm) history = model.fit( train_features, train_labels, validation_split=0.1, verbose=0, epochs=10) fname = 'model.h5' model.save(fname) p0 = 7.6 p1 = 0.85 v = model.predict([(p0, p1)], verbose=0)[0] print('original output: ', v) lmodel = keras.models.load_model(fname) v = lmodel.predict([(p0, p1)], verbose=0)[0] print('restored output: ', v) ``` **logs**. ``` original output: [667.36383] restored output: [-8.889455] ```
perf
tf keras layers normalization invert true does not restore from save system information have i written custom code yes os platform and distribution e g linux ubuntu ubuntu tensorflow installed from source or binary tf nightly from pip tensorflow version use command below python version bazel version if compiling from source gpu model and memory cpu n a exact command to reproduce run the code below describe the problem keras models load model does not load tf keras layers normalization layer with invert true describe the current behavior it seems to omit invert true flag when loading or saving the model describe the expected behavior original trained and saved loaded model output the same when using normalization with invert true standalone code to reproduce the issue python import numpy as np import pandas as pd import sys import tensorflow as tf from tensorflow import keras from tensorflow keras import layers print tf version dataset pd read csv train dataset dataset sample frac random state train features train dataset copy train labels train features pop orp inorm tf keras layers normalization axis input shape inorm adapt np array train features onorm tf keras layers normalization axis invert true onorm adapt train labels def build and compile model inorm onorm model keras sequential inorm layers dense activation relu layers dense onorm model compile loss mean squared error optimizer adam return model model build and compile model inorm onorm history model fit train features train labels validation split verbose epochs fname model model save fname v model predict verbose print original output v lmodel keras models load model fname v lmodel predict verbose print restored output v logs original output restored output
1
6,758
5,626,997,676
IssuesEvent
2017-04-04 23:41:47
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
Engine roll caused regression to Android stock_layout_iteration times
performance regression ▣ platform-android
This PR: https://github.com/flutter/flutter/pull/9137 ...which rolled the engine from: https://github.com/flutter/engine/commit/0005285781cd8eab7a43571bccdaad8354c6faa4 ...to: https://github.com/flutter/engine/commit/4c05830aafb75c2893ea98bfd932ac95e883bce2 ...may have caused a minor (1.05x) regression to the stock_layout_iteration benchmark on Android.
True
Engine roll caused regression to Android stock_layout_iteration times - This PR: https://github.com/flutter/flutter/pull/9137 ...which rolled the engine from: https://github.com/flutter/engine/commit/0005285781cd8eab7a43571bccdaad8354c6faa4 ...to: https://github.com/flutter/engine/commit/4c05830aafb75c2893ea98bfd932ac95e883bce2 ...may have caused a minor (1.05x) regression to the stock_layout_iteration benchmark on Android.
perf
engine roll caused regression to android stock layout iteration times this pr which rolled the engine from to may have caused a minor regression to the stock layout iteration benchmark on android
1
37,575
18,536,617,108
IssuesEvent
2021-10-21 12:14:30
getsentry/sentry-javascript
https://api.github.com/repos/getsentry/sentry-javascript
closed
Question: Can we customize the end-time of Page Load transactions? Plus, capture paint time
Type: Discussion Feature: Performance Status: Stale
Hey folks! I'm coming up to speed on Sentry and trying to integrate it with my organization's frontend. One thing we'd like to do is understand better how Sentry starts / ends Page Load (and React Router) transactions. We've got some pages that we _believe_ are taking longer to "load" than Sentry is reporting, and we'd like a way to customize when those transactions are completed. Basically, I think our TTI is later than Sentry seems to believe it is. For reference, here's a sample transaction and a few spans from it: ![image](https://user-images.githubusercontent.com/4821111/125306002-d352ff80-e2fc-11eb-841b-f47f29d05db4.png) We're also using the React Router v5 integration to track load times for internal page switches, and we'd like to include some React metrics in there as well. It looks like, currently, only Navigation time and HTTP requests are being tracked. This would also be handleable if we knew how these transactions are being completed, and if we could attach more data to them. ![image](https://user-images.githubusercontent.com/4821111/125306443-3a70b400-e2fd-11eb-8e23-c0741d00264a.png) Finally, we've got a component that's taking a LONG time to paint that we'd like to capture some metrics about. We can create a transaction inside React to handle the `render` bits, but it's a Canvas component and I believe the paint is taking upwards of 1s. We'd really like to capture that time (along with some metadata like number of painted elements) so we can measure improvement. Does Sentry have a way to capture this paint time? Thanks again!
True
Question: Can we customize the end-time of Page Load transactions? Plus, capture paint time - Hey folks! I'm coming up to speed on Sentry and trying to integrate it with my organization's frontend. One thing we'd like to do is understand better how Sentry starts / ends Page Load (and React Router) transactions. We've got some pages that we _believe_ are taking longer to "load" than Sentry is reporting, and we'd like a way to customize when those transactions are completed. Basically, I think our TTI is later than Sentry seems to believe it is. For reference, here's a sample transaction and a few spans from it: ![image](https://user-images.githubusercontent.com/4821111/125306002-d352ff80-e2fc-11eb-841b-f47f29d05db4.png) We're also using the React Router v5 integration to track load times for internal page switches, and we'd like to include some React metrics in there as well. It looks like, currently, only Navigation time and HTTP requests are being tracked. This would also be handleable if we knew how these transactions are being completed, and if we could attach more data to them. ![image](https://user-images.githubusercontent.com/4821111/125306443-3a70b400-e2fd-11eb-8e23-c0741d00264a.png) Finally, we've got a component that's taking a LONG time to paint that we'd like to capture some metrics about. We can create a transaction inside React to handle the `render` bits, but it's a Canvas component and I believe the paint is taking upwards of 1s. We'd really like to capture that time (along with some metadata like number of painted elements) so we can measure improvement. Does Sentry have a way to capture this paint time? Thanks again!
perf
question can we customize the end time of page load transactions plus capture paint time hey folks i m coming up to speed on sentry and trying to integrate it with my organization s frontend one thing we d like to do is understand better how sentry starts ends page load and react router transactions we ve got some pages that we believe are taking longer to load than sentry is reporting and we d like a way to customize when those transactions are completed basically i think our tti is later than sentry seems to believe it is for reference here s a sample transaction and a few spans from it we re also using the react router integration to track load times for internal page switches and we d like to include some react metrics in there as well it looks like currently only navigation time and http requests are being tracked this would also be handleable if we knew how these transactions are being completed and if we could attach more data to them finally we ve got a component that s taking a long time to paint that we d like to capture some metrics about we can create a transaction inside react to handle the render bits but it s a canvas component and i believe the paint is taking upwards of we d really like to capture that time along with some metadata like number of painted elements so we can measure improvement does sentry have a way to capture this paint time thanks again
1
32,738
15,605,943,770
IssuesEvent
2021-03-19 07:12:05
ropensci/targets
https://api.github.com/repos/ropensci/targets
closed
tar_make slow when callr_function not NULL
depends: reprex topic: performance
## Prework * [x] Read and agree to the [code of conduct](https://ropensci.org/code-of-conduct/) and [contributing guidelines](https://github.com/ropensci/targets/blob/main/CONTRIBUTING.md). * [x] If there is [already a relevant issue](https://github.com/ropensci/targets/issues), whether open or closed, comment on the existing thread instead of posting a new issue. * [ ] Post a [minimal reproducible example](https://www.tidyverse.org/help/) like [this one](https://github.com/ropensci/targets/issues/256#issuecomment-754229683) so the maintainer can troubleshoot the problems you identify. A reproducible example is: * [ ] **Runnable**: post enough R code and data so any onlooker can create the error on their own computer. * [ ] **Minimal**: reduce runtime wherever possible and remove complicated details that are irrelevant to the issue at hand. * [ ] **Readable**: format your code according to the [tidyverse style guide](https://style.tidyverse.org/). ## Description `tar_make()` takes a lot of time to start up (several minutes). In comparison, `tar_make(callr_function = NULL)` starts almost without delay. See the benchmarks. ## Reproducible example Since I'm working with confidential data, creating a reprex is quite some work. Therefore, I hope the benchmarks already help. If they don't I will take the time to create a reprex. ## Benchmarks I tried to use `proffer`, but I got some errors. Also, as far as I understand, `proffer` should only work with `callr_function = NULL`. <details> ```r > proffer::pprof(tar_make()) * run target mifid_files v skip target ids v skip branch mifid_e5ccf3fc v skip branch mifid_276b0dbc v skip branch mifid_85ba955c v skip branch mifid_32c58c89 v skip branch ids_list_d984c7ca v skip branch ids_list_afadd32b v skip branch ids_list_1622817d v skip branch ids_list_5ff49e59 v skip branch id_counts_07b60a7a v skip branch id_counts_4ffd5edf v skip branch id_counts_6d3fac2a v skip branch id_counts_f10827a2 v skip branch id_counts_259d208a v skip branch id_counts_5bced0f6 v skip branch sdc_07b60a7a v skip branch sdc_4ffd5edf v skip branch sdc_6d3fac2a v skip branch sdc_f10827a2 v skip branch sdc_259d208a v skip branch sdc_5bced0f6 * end pipeline Error in rethrow_call(c_processx_exec, command, c(command, args), stdin, : Command '' not found @win/processx.c:994 (processx_exec) Type .Last.error.trace to see where the error occured > proffer::pprof(tar_make(callr_function = NULL)) * run target mifid_files v skip target ids v skip branch mifid_e5ccf3fc v skip branch mifid_276b0dbc v skip branch mifid_85ba955c v skip branch mifid_32c58c89 v skip branch ids_list_d984c7ca v skip branch ids_list_afadd32b v skip branch ids_list_1622817d v skip branch ids_list_5ff49e59 v skip branch id_counts_07b60a7a v skip branch id_counts_4ffd5edf v skip branch id_counts_6d3fac2a v skip branch id_counts_f10827a2 v skip branch id_counts_259d208a v skip branch id_counts_5bced0f6 v skip branch sdc_07b60a7a v skip branch sdc_4ffd5edf v skip branch sdc_6d3fac2a v skip branch sdc_f10827a2 v skip branch sdc_259d208a v skip branch sdc_5bced0f6 * end pipeline Error in rethrow_call(c_processx_exec, command, c(command, args), stdin, : Command '' not found @win/processx.c:994 (processx_exec) Type .Last.error.trace to see where the error occured > .Last.error.trace Stack trace: 1. proffer::pprof(tar_make(callr_function = NULL)) 2. proffer:::serve_pprof(pprof = pprof, host = host, port = port, ... 3. proffer:::serve_pprof_impl(args) 4. proffer:::with_safe_path(Sys.getenv("PROFFER_GRAPHVIZ_BIN"), ... 5. processx::process$new(command = pprof_path(), args = args, stdout = "|", ... 6. .subset2(public_bind_env, "initialize")(...) 7. processx:::process_initialize(self, private, command, args, stdin, ... 8. rethrow_call(c_processx_exec, command, c(command, args), stdin, ... x Command '' not found @win/processx.c:994 (processx_exec) ``` </details> Thus, I measured the time for `tar_make()` with `system.time()`: ```r > system.time(tar_make(callr_function = NULL)) * run target mifid_files v skip target ids v skip branch mifid_e5ccf3fc v skip branch mifid_276b0dbc v skip branch mifid_85ba955c v skip branch mifid_32c58c89 v skip branch ids_list_d984c7ca v skip branch ids_list_afadd32b v skip branch ids_list_1622817d v skip branch ids_list_5ff49e59 v skip branch id_counts_07b60a7a v skip branch id_counts_4ffd5edf v skip branch id_counts_6d3fac2a v skip branch id_counts_f10827a2 v skip branch id_counts_259d208a v skip branch id_counts_5bced0f6 v skip branch sdc_07b60a7a v skip branch sdc_4ffd5edf v skip branch sdc_6d3fac2a v skip branch sdc_f10827a2 v skip branch sdc_259d208a v skip branch sdc_5bced0f6 * end pipeline user system elapsed 0.20 0.39 10.23 > system.time(tar_make()) * run target mifid_files v skip target ids v skip branch mifid_e5ccf3fc v skip branch mifid_276b0dbc v skip branch mifid_85ba955c v skip branch mifid_32c58c89 v skip branch ids_list_d984c7ca v skip branch ids_list_afadd32b v skip branch ids_list_1622817d v skip branch ids_list_5ff49e59 v skip branch id_counts_07b60a7a v skip branch id_counts_4ffd5edf v skip branch id_counts_6d3fac2a v skip branch id_counts_f10827a2 v skip branch id_counts_259d208a v skip branch id_counts_5bced0f6 v skip branch sdc_07b60a7a v skip branch sdc_4ffd5edf v skip branch sdc_6d3fac2a v skip branch sdc_f10827a2 v skip branch sdc_259d208a v skip branch sdc_5bced0f6 * end pipeline user system elapsed 2.39 0.59 372.45 ``` I also profiled using `profvis`. Not sure if this helps, but see for yourself. For `profvis::profvis(tar_make(callr_function = NULL))`: ![image](https://user-images.githubusercontent.com/6517388/111629891-0fb09c00-87f2-11eb-98e7-608499177491.png) And for `profvis::profvis(tar_make())`: ![image](https://user-images.githubusercontent.com/6517388/111630063-3f5fa400-87f2-11eb-8376-b3d4f64c1627.png) So it seems as if starting a new R session for `tar_make()` takes a lot of time, which I don't understand. Am I missing something obvious here? Also, as written before, please say if you need a real reprex and I'll try to create one.
True
tar_make slow when callr_function not NULL - ## Prework * [x] Read and agree to the [code of conduct](https://ropensci.org/code-of-conduct/) and [contributing guidelines](https://github.com/ropensci/targets/blob/main/CONTRIBUTING.md). * [x] If there is [already a relevant issue](https://github.com/ropensci/targets/issues), whether open or closed, comment on the existing thread instead of posting a new issue. * [ ] Post a [minimal reproducible example](https://www.tidyverse.org/help/) like [this one](https://github.com/ropensci/targets/issues/256#issuecomment-754229683) so the maintainer can troubleshoot the problems you identify. A reproducible example is: * [ ] **Runnable**: post enough R code and data so any onlooker can create the error on their own computer. * [ ] **Minimal**: reduce runtime wherever possible and remove complicated details that are irrelevant to the issue at hand. * [ ] **Readable**: format your code according to the [tidyverse style guide](https://style.tidyverse.org/). ## Description `tar_make()` takes a lot of time to start up (several minutes). In comparison, `tar_make(callr_function = NULL)` starts almost without delay. See the benchmarks. ## Reproducible example Since I'm working with confidential data, creating a reprex is quite some work. Therefore, I hope the benchmarks already help. If they don't I will take the time to create a reprex. ## Benchmarks I tried to use `proffer`, but I got some errors. Also, as far as I understand, `proffer` should only work with `callr_function = NULL`. <details> ```r > proffer::pprof(tar_make()) * run target mifid_files v skip target ids v skip branch mifid_e5ccf3fc v skip branch mifid_276b0dbc v skip branch mifid_85ba955c v skip branch mifid_32c58c89 v skip branch ids_list_d984c7ca v skip branch ids_list_afadd32b v skip branch ids_list_1622817d v skip branch ids_list_5ff49e59 v skip branch id_counts_07b60a7a v skip branch id_counts_4ffd5edf v skip branch id_counts_6d3fac2a v skip branch id_counts_f10827a2 v skip branch id_counts_259d208a v skip branch id_counts_5bced0f6 v skip branch sdc_07b60a7a v skip branch sdc_4ffd5edf v skip branch sdc_6d3fac2a v skip branch sdc_f10827a2 v skip branch sdc_259d208a v skip branch sdc_5bced0f6 * end pipeline Error in rethrow_call(c_processx_exec, command, c(command, args), stdin, : Command '' not found @win/processx.c:994 (processx_exec) Type .Last.error.trace to see where the error occured > proffer::pprof(tar_make(callr_function = NULL)) * run target mifid_files v skip target ids v skip branch mifid_e5ccf3fc v skip branch mifid_276b0dbc v skip branch mifid_85ba955c v skip branch mifid_32c58c89 v skip branch ids_list_d984c7ca v skip branch ids_list_afadd32b v skip branch ids_list_1622817d v skip branch ids_list_5ff49e59 v skip branch id_counts_07b60a7a v skip branch id_counts_4ffd5edf v skip branch id_counts_6d3fac2a v skip branch id_counts_f10827a2 v skip branch id_counts_259d208a v skip branch id_counts_5bced0f6 v skip branch sdc_07b60a7a v skip branch sdc_4ffd5edf v skip branch sdc_6d3fac2a v skip branch sdc_f10827a2 v skip branch sdc_259d208a v skip branch sdc_5bced0f6 * end pipeline Error in rethrow_call(c_processx_exec, command, c(command, args), stdin, : Command '' not found @win/processx.c:994 (processx_exec) Type .Last.error.trace to see where the error occured > .Last.error.trace Stack trace: 1. proffer::pprof(tar_make(callr_function = NULL)) 2. proffer:::serve_pprof(pprof = pprof, host = host, port = port, ... 3. proffer:::serve_pprof_impl(args) 4. proffer:::with_safe_path(Sys.getenv("PROFFER_GRAPHVIZ_BIN"), ... 5. processx::process$new(command = pprof_path(), args = args, stdout = "|", ... 6. .subset2(public_bind_env, "initialize")(...) 7. processx:::process_initialize(self, private, command, args, stdin, ... 8. rethrow_call(c_processx_exec, command, c(command, args), stdin, ... x Command '' not found @win/processx.c:994 (processx_exec) ``` </details> Thus, I measured the time for `tar_make()` with `system.time()`: ```r > system.time(tar_make(callr_function = NULL)) * run target mifid_files v skip target ids v skip branch mifid_e5ccf3fc v skip branch mifid_276b0dbc v skip branch mifid_85ba955c v skip branch mifid_32c58c89 v skip branch ids_list_d984c7ca v skip branch ids_list_afadd32b v skip branch ids_list_1622817d v skip branch ids_list_5ff49e59 v skip branch id_counts_07b60a7a v skip branch id_counts_4ffd5edf v skip branch id_counts_6d3fac2a v skip branch id_counts_f10827a2 v skip branch id_counts_259d208a v skip branch id_counts_5bced0f6 v skip branch sdc_07b60a7a v skip branch sdc_4ffd5edf v skip branch sdc_6d3fac2a v skip branch sdc_f10827a2 v skip branch sdc_259d208a v skip branch sdc_5bced0f6 * end pipeline user system elapsed 0.20 0.39 10.23 > system.time(tar_make()) * run target mifid_files v skip target ids v skip branch mifid_e5ccf3fc v skip branch mifid_276b0dbc v skip branch mifid_85ba955c v skip branch mifid_32c58c89 v skip branch ids_list_d984c7ca v skip branch ids_list_afadd32b v skip branch ids_list_1622817d v skip branch ids_list_5ff49e59 v skip branch id_counts_07b60a7a v skip branch id_counts_4ffd5edf v skip branch id_counts_6d3fac2a v skip branch id_counts_f10827a2 v skip branch id_counts_259d208a v skip branch id_counts_5bced0f6 v skip branch sdc_07b60a7a v skip branch sdc_4ffd5edf v skip branch sdc_6d3fac2a v skip branch sdc_f10827a2 v skip branch sdc_259d208a v skip branch sdc_5bced0f6 * end pipeline user system elapsed 2.39 0.59 372.45 ``` I also profiled using `profvis`. Not sure if this helps, but see for yourself. For `profvis::profvis(tar_make(callr_function = NULL))`: ![image](https://user-images.githubusercontent.com/6517388/111629891-0fb09c00-87f2-11eb-98e7-608499177491.png) And for `profvis::profvis(tar_make())`: ![image](https://user-images.githubusercontent.com/6517388/111630063-3f5fa400-87f2-11eb-8376-b3d4f64c1627.png) So it seems as if starting a new R session for `tar_make()` takes a lot of time, which I don't understand. Am I missing something obvious here? Also, as written before, please say if you need a real reprex and I'll try to create one.
perf
tar make slow when callr function not null prework read and agree to the and if there is whether open or closed comment on the existing thread instead of posting a new issue post a like so the maintainer can troubleshoot the problems you identify a reproducible example is runnable post enough r code and data so any onlooker can create the error on their own computer minimal reduce runtime wherever possible and remove complicated details that are irrelevant to the issue at hand readable format your code according to the description tar make takes a lot of time to start up several minutes in comparison tar make callr function null starts almost without delay see the benchmarks reproducible example since i m working with confidential data creating a reprex is quite some work therefore i hope the benchmarks already help if they don t i will take the time to create a reprex benchmarks i tried to use proffer but i got some errors also as far as i understand proffer should only work with callr function null r proffer pprof tar make run target mifid files v skip target ids v skip branch mifid v skip branch mifid v skip branch mifid v skip branch mifid v skip branch ids list v skip branch ids list v skip branch ids list v skip branch ids list v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc end pipeline error in rethrow call c processx exec command c command args stdin command not found win processx c processx exec type last error trace to see where the error occured proffer pprof tar make callr function null run target mifid files v skip target ids v skip branch mifid v skip branch mifid v skip branch mifid v skip branch mifid v skip branch ids list v skip branch ids list v skip branch ids list v skip branch ids list v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc end pipeline error in rethrow call c processx exec command c command args stdin command not found win processx c processx exec type last error trace to see where the error occured last error trace stack trace proffer pprof tar make callr function null proffer serve pprof pprof pprof host host port port proffer serve pprof impl args proffer with safe path sys getenv proffer graphviz bin processx process new command pprof path args args stdout public bind env initialize processx process initialize self private command args stdin rethrow call c processx exec command c command args stdin x command not found win processx c processx exec thus i measured the time for tar make with system time r system time tar make callr function null run target mifid files v skip target ids v skip branch mifid v skip branch mifid v skip branch mifid v skip branch mifid v skip branch ids list v skip branch ids list v skip branch ids list v skip branch ids list v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc end pipeline user system elapsed system time tar make run target mifid files v skip target ids v skip branch mifid v skip branch mifid v skip branch mifid v skip branch mifid v skip branch ids list v skip branch ids list v skip branch ids list v skip branch ids list v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc end pipeline user system elapsed i also profiled using profvis not sure if this helps but see for yourself for profvis profvis tar make callr function null and for profvis profvis tar make so it seems as if starting a new r session for tar make takes a lot of time which i don t understand am i missing something obvious here also as written before please say if you need a real reprex and i ll try to create one
1
15,840
6,044,823,684
IssuesEvent
2017-06-12 07:27:46
LuaJIT/LuaJIT
https://api.github.com/repos/LuaJIT/LuaJIT
closed
Static library doesn't build for LuaJIT v2.0.5 on Windows
2.0 2.1 bug build system Windows
`AR` doesn't work correctly on `cmd.exe` on Windows. The problem is that `AR` is `ar rcus 2>/dev/null`. Since Windows doesn't have `/dev/null`, `AR` doesn't work. Excerpt from [neovim build log](https://ci.appveyor.com/project/neovim/neovim/build/13809/job/sljudgsg3ykaw4uj): ``` cd /d C:\projects\neovim\.deps\build\src\luajit && C:\msys64\mingw64\bin\mingw32-make.exe CC=C:/msys64/mingw64/bin/gcc.exe PREFIX=C:/projects/neovim/.deps/usr CFLAGS+=-DLUAJIT_DISABLE_JIT CFLAGS+=-DLUA_USE_APICHECK CFLAGS+=-DLUA_USE_ASSERT CCDEBUG+=-g BUILDMODE=static mingw32-make[3]: Entering directory 'C:/projects/neovim/.deps/build/src/luajit' "==== Building LuaJIT 2.0.5 ====" C:/msys64/mingw64/bin/mingw32-make -C src mingw32-make[4]: Entering directory 'C:/projects/neovim/.deps/build/src/luajit/src' "HOSTCC host/minilua.o" "HOSTLINK host/minilua.exe" "DYNASM host/buildvm_arch.h" "HOSTCC host/buildvm.o" "HOSTCC host/buildvm_asm.o" "HOSTCC host/buildvm_peobj.o" "HOSTCC host/buildvm_lib.o" "HOSTCC host/buildvm_fold.o" "HOSTLINK host/buildvm.exe" "BUILDVM lj_vm.o" "CC lj_gc.o" "BUILDVM lj_ffdef.h" "CC lj_err.o" "CC lj_char.o" "BUILDVM lj_bcdef.h" "CC lj_bc.o" "CC lj_obj.o" "CC lj_str.o" "CC lj_tab.o" "CC lj_func.o" "CC lj_udata.o" "CC lj_meta.o" "CC lj_debug.o" "CC lj_state.o" "CC lj_dispatch.o" "CC lj_vmevent.o" "CC lj_vmmath.o" "CC lj_strscan.o" "CC lj_api.o" "CC lj_lex.o" "CC lj_parse.o" "CC lj_bcread.o" "CC lj_bcwrite.o" "CC lj_load.o" "CC lj_ir.o" "CC lj_opt_mem.o" "BUILDVM lj_folddef.h" "CC lj_opt_fold.o" "CC lj_opt_narrow.o" "CC lj_opt_dce.o" "CC lj_opt_loop.o" "CC lj_opt_split.o" "CC lj_opt_sink.o" "CC lj_mcode.o" "CC lj_snap.o" "CC lj_record.o" "CC lj_crecord.o" "BUILDVM lj_recdef.h" "CC lj_ffrecord.o" "CC lj_asm.o" "CC lj_trace.o" "CC lj_gdbjit.o" "CC lj_ctype.o" "CC lj_cdata.o" "CC lj_cconv.o" "CC lj_ccall.o" "CC lj_ccallback.o" "CC lj_carith.o" "CC lj_clib.o" "CC lj_cparse.o" "CC lj_lib.o" "CC lj_alloc.o" "CC lib_aux.o" "BUILDVM lj_libdef.h" "CC lib_base.o" "CC lib_math.o" "CC lib_bit.o" "CC lib_string.o" "CC lib_table.o" "CC lib_io.o" "CC lib_os.o" "CC lib_package.o" "CC lib_debug.o" "CC lib_jit.o" "CC lib_ffi.o" "CC lib_init.o" "AR libluajit.a" The system cannot find the path specified. ```
1.0
Static library doesn't build for LuaJIT v2.0.5 on Windows - `AR` doesn't work correctly on `cmd.exe` on Windows. The problem is that `AR` is `ar rcus 2>/dev/null`. Since Windows doesn't have `/dev/null`, `AR` doesn't work. Excerpt from [neovim build log](https://ci.appveyor.com/project/neovim/neovim/build/13809/job/sljudgsg3ykaw4uj): ``` cd /d C:\projects\neovim\.deps\build\src\luajit && C:\msys64\mingw64\bin\mingw32-make.exe CC=C:/msys64/mingw64/bin/gcc.exe PREFIX=C:/projects/neovim/.deps/usr CFLAGS+=-DLUAJIT_DISABLE_JIT CFLAGS+=-DLUA_USE_APICHECK CFLAGS+=-DLUA_USE_ASSERT CCDEBUG+=-g BUILDMODE=static mingw32-make[3]: Entering directory 'C:/projects/neovim/.deps/build/src/luajit' "==== Building LuaJIT 2.0.5 ====" C:/msys64/mingw64/bin/mingw32-make -C src mingw32-make[4]: Entering directory 'C:/projects/neovim/.deps/build/src/luajit/src' "HOSTCC host/minilua.o" "HOSTLINK host/minilua.exe" "DYNASM host/buildvm_arch.h" "HOSTCC host/buildvm.o" "HOSTCC host/buildvm_asm.o" "HOSTCC host/buildvm_peobj.o" "HOSTCC host/buildvm_lib.o" "HOSTCC host/buildvm_fold.o" "HOSTLINK host/buildvm.exe" "BUILDVM lj_vm.o" "CC lj_gc.o" "BUILDVM lj_ffdef.h" "CC lj_err.o" "CC lj_char.o" "BUILDVM lj_bcdef.h" "CC lj_bc.o" "CC lj_obj.o" "CC lj_str.o" "CC lj_tab.o" "CC lj_func.o" "CC lj_udata.o" "CC lj_meta.o" "CC lj_debug.o" "CC lj_state.o" "CC lj_dispatch.o" "CC lj_vmevent.o" "CC lj_vmmath.o" "CC lj_strscan.o" "CC lj_api.o" "CC lj_lex.o" "CC lj_parse.o" "CC lj_bcread.o" "CC lj_bcwrite.o" "CC lj_load.o" "CC lj_ir.o" "CC lj_opt_mem.o" "BUILDVM lj_folddef.h" "CC lj_opt_fold.o" "CC lj_opt_narrow.o" "CC lj_opt_dce.o" "CC lj_opt_loop.o" "CC lj_opt_split.o" "CC lj_opt_sink.o" "CC lj_mcode.o" "CC lj_snap.o" "CC lj_record.o" "CC lj_crecord.o" "BUILDVM lj_recdef.h" "CC lj_ffrecord.o" "CC lj_asm.o" "CC lj_trace.o" "CC lj_gdbjit.o" "CC lj_ctype.o" "CC lj_cdata.o" "CC lj_cconv.o" "CC lj_ccall.o" "CC lj_ccallback.o" "CC lj_carith.o" "CC lj_clib.o" "CC lj_cparse.o" "CC lj_lib.o" "CC lj_alloc.o" "CC lib_aux.o" "BUILDVM lj_libdef.h" "CC lib_base.o" "CC lib_math.o" "CC lib_bit.o" "CC lib_string.o" "CC lib_table.o" "CC lib_io.o" "CC lib_os.o" "CC lib_package.o" "CC lib_debug.o" "CC lib_jit.o" "CC lib_ffi.o" "CC lib_init.o" "AR libluajit.a" The system cannot find the path specified. ```
non_perf
static library doesn t build for luajit on windows ar doesn t work correctly on cmd exe on windows the problem is that ar is ar rcus dev null since windows doesn t have dev null ar doesn t work excerpt from cd d c projects neovim deps build src luajit c bin make exe cc c bin gcc exe prefix c projects neovim deps usr cflags dluajit disable jit cflags dlua use apicheck cflags dlua use assert ccdebug g buildmode static make entering directory c projects neovim deps build src luajit building luajit c bin make c src make entering directory c projects neovim deps build src luajit src hostcc host minilua o hostlink host minilua exe dynasm host buildvm arch h hostcc host buildvm o hostcc host buildvm asm o hostcc host buildvm peobj o hostcc host buildvm lib o hostcc host buildvm fold o hostlink host buildvm exe buildvm lj vm o cc lj gc o buildvm lj ffdef h cc lj err o cc lj char o buildvm lj bcdef h cc lj bc o cc lj obj o cc lj str o cc lj tab o cc lj func o cc lj udata o cc lj meta o cc lj debug o cc lj state o cc lj dispatch o cc lj vmevent o cc lj vmmath o cc lj strscan o cc lj api o cc lj lex o cc lj parse o cc lj bcread o cc lj bcwrite o cc lj load o cc lj ir o cc lj opt mem o buildvm lj folddef h cc lj opt fold o cc lj opt narrow o cc lj opt dce o cc lj opt loop o cc lj opt split o cc lj opt sink o cc lj mcode o cc lj snap o cc lj record o cc lj crecord o buildvm lj recdef h cc lj ffrecord o cc lj asm o cc lj trace o cc lj gdbjit o cc lj ctype o cc lj cdata o cc lj cconv o cc lj ccall o cc lj ccallback o cc lj carith o cc lj clib o cc lj cparse o cc lj lib o cc lj alloc o cc lib aux o buildvm lj libdef h cc lib base o cc lib math o cc lib bit o cc lib string o cc lib table o cc lib io o cc lib os o cc lib package o cc lib debug o cc lib jit o cc lib ffi o cc lib init o ar libluajit a the system cannot find the path specified
0
35,805
17,267,354,675
IssuesEvent
2021-07-22 15:13:07
getsentry/sentry-cocoa
https://api.github.com/repos/getsentry/sentry-cocoa
closed
Name for SwiftUI auto generated transactions is long
Performance Product bug
### Environment iOSSwiftUI Sample app on this PR https://github.com/getsentry/sentry-cocoa/pull/1228 ### Steps to Reproduce Start iOSSwiftUI Sample app Wait until the transaction is sent to Sentry ### Expected Result A shorter nicer more descriptive transaction name. ### Actual Result The name of the transaction is `_TtGC7SwiftUI19UIHostingControllerGVS_15ModifiedContentVS_7AnyViewVS_12RootModifier__` <img width="724" alt="Screen Shot 2021-07-19 at 11 17 31" src="https://user-images.githubusercontent.com/2443292/126136172-446be568-b64d-4aa1-819a-866d1f02c928.png">
True
Name for SwiftUI auto generated transactions is long - ### Environment iOSSwiftUI Sample app on this PR https://github.com/getsentry/sentry-cocoa/pull/1228 ### Steps to Reproduce Start iOSSwiftUI Sample app Wait until the transaction is sent to Sentry ### Expected Result A shorter nicer more descriptive transaction name. ### Actual Result The name of the transaction is `_TtGC7SwiftUI19UIHostingControllerGVS_15ModifiedContentVS_7AnyViewVS_12RootModifier__` <img width="724" alt="Screen Shot 2021-07-19 at 11 17 31" src="https://user-images.githubusercontent.com/2443292/126136172-446be568-b64d-4aa1-819a-866d1f02c928.png">
perf
name for swiftui auto generated transactions is long environment iosswiftui sample app on this pr steps to reproduce start iosswiftui sample app wait until the transaction is sent to sentry expected result a shorter nicer more descriptive transaction name actual result the name of the transaction is img width alt screen shot at src
1
331,720
10,076,093,095
IssuesEvent
2019-07-24 15:31:04
Kennygunderman/SunDownShowdown
https://api.github.com/repos/Kennygunderman/SunDownShowdown
closed
Clear mobs from arena
enhancement priority
We need a way to clear mobs from the arena either before the showdown starts (to make room for adding the new mobs) or after it ends (this would mean setting a time on when the showdown is 'over'). To do accomplish this we can specify the location and size of the showdown arena. add cmd `/showdown arena set radius <radius size>` to set the location and size of the arena from where the user is standing. With the arena location and size set, we can clear mobs from this location by using something like: `world.getNearbyEntitys`to get all entities within the arena, then having logic to remove mob entities.
1.0
Clear mobs from arena - We need a way to clear mobs from the arena either before the showdown starts (to make room for adding the new mobs) or after it ends (this would mean setting a time on when the showdown is 'over'). To do accomplish this we can specify the location and size of the showdown arena. add cmd `/showdown arena set radius <radius size>` to set the location and size of the arena from where the user is standing. With the arena location and size set, we can clear mobs from this location by using something like: `world.getNearbyEntitys`to get all entities within the arena, then having logic to remove mob entities.
non_perf
clear mobs from arena we need a way to clear mobs from the arena either before the showdown starts to make room for adding the new mobs or after it ends this would mean setting a time on when the showdown is over to do accomplish this we can specify the location and size of the showdown arena add cmd showdown arena set radius to set the location and size of the arena from where the user is standing with the arena location and size set we can clear mobs from this location by using something like world getnearbyentitys to get all entities within the arena then having logic to remove mob entities
0
621,319
19,582,951,267
IssuesEvent
2022-01-05 00:41:43
Qiskit-Partners/qiskit-ibm
https://api.github.com/repos/Qiskit-Partners/qiskit-ibm
closed
FAIL: test_provider_backends
type: bug priority: high
<!-- ⚠️ If you do not respect this template, your issue will be closed --> <!-- ⚠️ Make sure to browse the opened and closed issues --> ### Information - **Qiskit IBM Provider version**: - **Python version**: - **Operating system**: ### What is the current behavior? 3 tests are failing due to a deprecation warning in terra-main runs. https://github.com/Qiskit/qiskit-ibmq-provider/runs/4377970666?check_suite_focus=true#step:5:4229 ``` DeprecationWarning: The `backends` provider attribute is deprecated. Please use `provider.backend` (singular) instead. You can continue to use `provider.backends()` to retrieve all backends. ``` ### Steps to reproduce the problem ### What is the expected behavior? ### Suggested solutions
1.0
FAIL: test_provider_backends - <!-- ⚠️ If you do not respect this template, your issue will be closed --> <!-- ⚠️ Make sure to browse the opened and closed issues --> ### Information - **Qiskit IBM Provider version**: - **Python version**: - **Operating system**: ### What is the current behavior? 3 tests are failing due to a deprecation warning in terra-main runs. https://github.com/Qiskit/qiskit-ibmq-provider/runs/4377970666?check_suite_focus=true#step:5:4229 ``` DeprecationWarning: The `backends` provider attribute is deprecated. Please use `provider.backend` (singular) instead. You can continue to use `provider.backends()` to retrieve all backends. ``` ### Steps to reproduce the problem ### What is the expected behavior? ### Suggested solutions
non_perf
fail test provider backends information qiskit ibm provider version python version operating system what is the current behavior tests are failing due to a deprecation warning in terra main runs deprecationwarning the backends provider attribute is deprecated please use provider backend singular instead you can continue to use provider backends to retrieve all backends steps to reproduce the problem what is the expected behavior suggested solutions
0
33,016
15,761,456,446
IssuesEvent
2021-03-31 10:00:25
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
opened
Performance tests: improve the deploy of the simulated agents
performance type/enhancement
The current implementation of this deployment adds a huge overhead in the pipeline because it is in charge of deploying the instances where the simulated agents will run, the toolset, and run the simulated agents. We must move this deployment to Fargate to get rid of the instances deployment and the installation of the toolset. Using Fargate we can define a Docker image that will run the simulated agents, making it completely transparent to the pipeline. **Tasks** - [ ] Create the ECS cluster. - [ ] Create a new Docker image to run the simulated agents. - [ ] Create a new ECS task to run it: - [ ] Create the task. - [ ] Analyze the behavior of the Docker and establish the best hardware resources for it. - [ ] Refactor the pipeline.
True
Performance tests: improve the deploy of the simulated agents - The current implementation of this deployment adds a huge overhead in the pipeline because it is in charge of deploying the instances where the simulated agents will run, the toolset, and run the simulated agents. We must move this deployment to Fargate to get rid of the instances deployment and the installation of the toolset. Using Fargate we can define a Docker image that will run the simulated agents, making it completely transparent to the pipeline. **Tasks** - [ ] Create the ECS cluster. - [ ] Create a new Docker image to run the simulated agents. - [ ] Create a new ECS task to run it: - [ ] Create the task. - [ ] Analyze the behavior of the Docker and establish the best hardware resources for it. - [ ] Refactor the pipeline.
perf
performance tests improve the deploy of the simulated agents the current implementation of this deployment adds a huge overhead in the pipeline because it is in charge of deploying the instances where the simulated agents will run the toolset and run the simulated agents we must move this deployment to fargate to get rid of the instances deployment and the installation of the toolset using fargate we can define a docker image that will run the simulated agents making it completely transparent to the pipeline tasks create the ecs cluster create a new docker image to run the simulated agents create a new ecs task to run it create the task analyze the behavior of the docker and establish the best hardware resources for it refactor the pipeline
1
6,091
8,951,501,866
IssuesEvent
2019-01-25 14:08:53
jasonblais/mattermost-community
https://api.github.com/repos/jasonblais/mattermost-community
opened
Add "PR Submitted" label for issues that already have a PR submitted
Contributor Journey Process
This could be automated for those PRs that reference a given help wanted issue.
1.0
Add "PR Submitted" label for issues that already have a PR submitted - This could be automated for those PRs that reference a given help wanted issue.
non_perf
add pr submitted label for issues that already have a pr submitted this could be automated for those prs that reference a given help wanted issue
0
102,167
12,748,735,159
IssuesEvent
2020-06-26 20:49:28
Hubvet8/hubvet-web
https://api.github.com/repos/Hubvet8/hubvet-web
opened
[Tarefa] Novo Breadcrumbs (Layout para novo posicionamento)
Módulo: Workspaces Tarefa UX Design
## Descrição Breve resumo do que é essa tarefa, descreva o trabalho a ser realizado - Deve ser feito a tela de exames conforme layout e épico Dados | Tipo ----|---- Texto 1 | String Número 2 | Integer,int Booleano 3 | boolean Monetáio 4 | Float Data 5 | datetime Hora 6 | Time Data e Hora | datetime Lista 7 | Select/combobox ## Ações - Selecionar - Deve aparecer o toolbar com a quantidade de itens selecionados e um botão para excluir todos os selecionados - Editar: - Deve chamar o modal/dialog para editar o Exame OBS: fazer após tela de cadastro estar concluída - Excluir: - Deve excluir o exame da mesma linha - Pesquisa: - Filtros sigla ou nome - [ ] Dados - [ ] Botões novo Exame e importar - [ ] Ações/Métodos - [ ] Selecionar: Deve aparecer o toolbar com a quantidade de itens selecionados e um botão para - [ ] excluir todos os selecionados - [ ] Editar: Deve chamar o modal/dialog para editar o Exame OBS: fazer após tela de cadastro estar concluída - [ ] Excluir: Deve excluir o exame da mesma linha - [ ] Botão Novo Exame: Deve abrir o modal para cadastrar novo Exame - [ ] Botão importar : Deve Abrir uma tela para escolher exames para importar exames - [ ] Pesquisa: - [ ] Pesquisar por sigla ou nome do exame Sugestões: Layout: ### API da tarefa Link:
1.0
[Tarefa] Novo Breadcrumbs (Layout para novo posicionamento) - ## Descrição Breve resumo do que é essa tarefa, descreva o trabalho a ser realizado - Deve ser feito a tela de exames conforme layout e épico Dados | Tipo ----|---- Texto 1 | String Número 2 | Integer,int Booleano 3 | boolean Monetáio 4 | Float Data 5 | datetime Hora 6 | Time Data e Hora | datetime Lista 7 | Select/combobox ## Ações - Selecionar - Deve aparecer o toolbar com a quantidade de itens selecionados e um botão para excluir todos os selecionados - Editar: - Deve chamar o modal/dialog para editar o Exame OBS: fazer após tela de cadastro estar concluída - Excluir: - Deve excluir o exame da mesma linha - Pesquisa: - Filtros sigla ou nome - [ ] Dados - [ ] Botões novo Exame e importar - [ ] Ações/Métodos - [ ] Selecionar: Deve aparecer o toolbar com a quantidade de itens selecionados e um botão para - [ ] excluir todos os selecionados - [ ] Editar: Deve chamar o modal/dialog para editar o Exame OBS: fazer após tela de cadastro estar concluída - [ ] Excluir: Deve excluir o exame da mesma linha - [ ] Botão Novo Exame: Deve abrir o modal para cadastrar novo Exame - [ ] Botão importar : Deve Abrir uma tela para escolher exames para importar exames - [ ] Pesquisa: - [ ] Pesquisar por sigla ou nome do exame Sugestões: Layout: ### API da tarefa Link:
non_perf
novo breadcrumbs layout para novo posicionamento descrição breve resumo do que é essa tarefa descreva o trabalho a ser realizado deve ser feito a tela de exames conforme layout e épico dados tipo texto string número integer int booleano boolean monetáio float data datetime hora time data e hora datetime lista select combobox ações selecionar deve aparecer o toolbar com a quantidade de itens selecionados e um botão para excluir todos os selecionados editar deve chamar o modal dialog para editar o exame obs fazer após tela de cadastro estar concluída excluir deve excluir o exame da mesma linha pesquisa filtros sigla ou nome dados botões novo exame e importar ações métodos selecionar deve aparecer o toolbar com a quantidade de itens selecionados e um botão para excluir todos os selecionados editar deve chamar o modal dialog para editar o exame obs fazer após tela de cadastro estar concluída excluir deve excluir o exame da mesma linha botão novo exame deve abrir o modal para cadastrar novo exame botão importar deve abrir uma tela para escolher exames para importar exames pesquisa pesquisar por sigla ou nome do exame sugestões layout api da tarefa link
0
70,268
23,086,210,479
IssuesEvent
2022-07-26 11:39:25
matrix-org/synapse
https://api.github.com/repos/matrix-org/synapse
closed
Faster joins: check for partial state when handling backfill
A-Federated-Join T-Defect
Suppose: * We have two chunks of room timeline, each with partial state. * We attempt to backfill backwards from the later chunk, and fill the gap (so we reach the earlier chunk) I think there is a bug here whereby we will incorrectly assume we have full state at the start of the backfilled chunk, which sounds messy. https://github.com/matrix-org/synapse/blob/7c6b2204d143550d81e5bf9612c4e69fe0866b4c/synapse/handlers/federation_event.py#L778-L785 Part of #12646
1.0
Faster joins: check for partial state when handling backfill - Suppose: * We have two chunks of room timeline, each with partial state. * We attempt to backfill backwards from the later chunk, and fill the gap (so we reach the earlier chunk) I think there is a bug here whereby we will incorrectly assume we have full state at the start of the backfilled chunk, which sounds messy. https://github.com/matrix-org/synapse/blob/7c6b2204d143550d81e5bf9612c4e69fe0866b4c/synapse/handlers/federation_event.py#L778-L785 Part of #12646
non_perf
faster joins check for partial state when handling backfill suppose we have two chunks of room timeline each with partial state we attempt to backfill backwards from the later chunk and fill the gap so we reach the earlier chunk i think there is a bug here whereby we will incorrectly assume we have full state at the start of the backfilled chunk which sounds messy part of
0
54,604
30,271,839,378
IssuesEvent
2023-07-07 15:57:53
playcanvas/engine
https://api.github.com/repos/playcanvas/engine
closed
Low frame rates on Android (Chrome 94) with WebGL 2
performance area: graphics
### Description Considerably lower frame rates when using WebGL 2 as compared to WebGL on Android (Chrome 94). Reproduced on Samsung Galaxy Tab S7+, Samsung Galaxy S8+ The screenshots below were taken on the Galaxy Tab. ### WebGL 2 ![Screenshot_20211001-142305_Chrome](https://user-images.githubusercontent.com/80072747/135623493-8a9561f3-6247-41d2-b9ad-bb21468a149e.jpg) ### WebGL ![Screenshot_20211001-142155_Chrome](https://user-images.githubusercontent.com/80072747/135623513-18703c5d-e422-43ad-93f7-7db880e2b321.jpg) Example project: https://playcanvas.com/project/836325/overview/webglwebgl-2-perf-test
True
Low frame rates on Android (Chrome 94) with WebGL 2 - ### Description Considerably lower frame rates when using WebGL 2 as compared to WebGL on Android (Chrome 94). Reproduced on Samsung Galaxy Tab S7+, Samsung Galaxy S8+ The screenshots below were taken on the Galaxy Tab. ### WebGL 2 ![Screenshot_20211001-142305_Chrome](https://user-images.githubusercontent.com/80072747/135623493-8a9561f3-6247-41d2-b9ad-bb21468a149e.jpg) ### WebGL ![Screenshot_20211001-142155_Chrome](https://user-images.githubusercontent.com/80072747/135623513-18703c5d-e422-43ad-93f7-7db880e2b321.jpg) Example project: https://playcanvas.com/project/836325/overview/webglwebgl-2-perf-test
perf
low frame rates on android chrome with webgl description considerably lower frame rates when using webgl as compared to webgl on android chrome reproduced on samsung galaxy tab samsung galaxy the screenshots below were taken on the galaxy tab webgl webgl example project
1
287,062
8,798,595,475
IssuesEvent
2018-12-24 08:45:44
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
support.microsoft.com - see bug description
browser-firefox priority-critical
<!-- @browser: Firefox 65.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update **Browser / Version**: Firefox 65.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: Access denied, this happens all the time with the Microsoft website searched through Google.com **Steps to Reproduce**: If I go to the Microsoft website directly (without searching through Google) there seems to be no problem... But problem always appears in a case like this: -> Searched on Google for "windows 10 update download" -> And clicked on the following link: De Update voor Windows 10, oktober 2018 downloaden https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update 31 okt. 2018 - Windows 10 waarschuwt u wanneer de update van oktober 2018 beschikbaar is voor uw apparaat. -> Then it showed "access denied" [![Screenshot Description](https://webcompat.com/uploads/2018/12/e0231ff2-3ff4-4fd4-97a8-45736c9749fb-thumb.jpeg)](https://webcompat.com/uploads/2018/12/e0231ff2-3ff4-4fd4-97a8-45736c9749fb.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181220174318</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Error: "The character encoding of the HTML document was not declared. The document will render with garbled text in some browser configurations if the document contains characters from outside the US-ASCII range. The character encoding of the page must be declared in the document or in the transfer protocol." {file: "https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
support.microsoft.com - see bug description - <!-- @browser: Firefox 65.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update **Browser / Version**: Firefox 65.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: Access denied, this happens all the time with the Microsoft website searched through Google.com **Steps to Reproduce**: If I go to the Microsoft website directly (without searching through Google) there seems to be no problem... But problem always appears in a case like this: -> Searched on Google for "windows 10 update download" -> And clicked on the following link: De Update voor Windows 10, oktober 2018 downloaden https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update 31 okt. 2018 - Windows 10 waarschuwt u wanneer de update van oktober 2018 beschikbaar is voor uw apparaat. -> Then it showed "access denied" [![Screenshot Description](https://webcompat.com/uploads/2018/12/e0231ff2-3ff4-4fd4-97a8-45736c9749fb-thumb.jpeg)](https://webcompat.com/uploads/2018/12/e0231ff2-3ff4-4fd4-97a8-45736c9749fb.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181220174318</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Error: "The character encoding of the HTML document was not declared. The document will render with garbled text in some browser configurations if the document contains characters from outside the US-ASCII range. The character encoding of the page must be declared in the document or in the transfer protocol." {file: "https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_perf
support microsoft com see bug description url browser version firefox operating system windows tested another browser yes problem type something else description access denied this happens all the time with the microsoft website searched through google com steps to reproduce if i go to the microsoft website directly without searching through google there seems to be no problem but problem always appears in a case like this searched on google for windows update download and clicked on the following link de update voor windows oktober downloaden okt windows waarschuwt u wanneer de update van oktober beschikbaar is voor uw apparaat then it showed access denied browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages from with ❤️
0
45,763
24,201,757,365
IssuesEvent
2022-09-24 17:14:22
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
Question: Does the ZFS kernel modul benefit from compiler option -march=znver2 or znver3?
Type: Performance Type: Question Component: Encryption
I do not know where to ask that question so I put it here as a bug report. Sorry for that. With the graysky2 kernel compiler patch (https://github.com/graysky2/kernel_compiler_patch) it is possible to compile the linux kernel with compiler optimizations for Ryzen processors for example (-march=znver2 or -march=znver3). This patch is already included in linux-xanmod and linux-zen. And if the kernel is compiled with -march=znver2 the zfs module is also compiled with that option. I tested that. Does that make any difference? Is that good or bad for the zfs performance or stability? I was wondering if encryption or compression would benefit from that.
True
Question: Does the ZFS kernel modul benefit from compiler option -march=znver2 or znver3? - I do not know where to ask that question so I put it here as a bug report. Sorry for that. With the graysky2 kernel compiler patch (https://github.com/graysky2/kernel_compiler_patch) it is possible to compile the linux kernel with compiler optimizations for Ryzen processors for example (-march=znver2 or -march=znver3). This patch is already included in linux-xanmod and linux-zen. And if the kernel is compiled with -march=znver2 the zfs module is also compiled with that option. I tested that. Does that make any difference? Is that good or bad for the zfs performance or stability? I was wondering if encryption or compression would benefit from that.
perf
question does the zfs kernel modul benefit from compiler option march or i do not know where to ask that question so i put it here as a bug report sorry for that with the kernel compiler patch it is possible to compile the linux kernel with compiler optimizations for ryzen processors for example march or march this patch is already included in linux xanmod and linux zen and if the kernel is compiled with march the zfs module is also compiled with that option i tested that does that make any difference is that good or bad for the zfs performance or stability i was wondering if encryption or compression would benefit from that
1
37,454
18,408,526,751
IssuesEvent
2021-10-13 00:39:04
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
opened
different output value in pytorch->onnx->tflite(int8 quantization)
type:performance
<em>Please make sure that this is an issue related to performance of TensorFlow. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:performance_template</em> **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 20.04 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: - TensorFlow installed from (source or binary): tensorflow:2.5.0-gpu docker - TensorFlow version (use command below): 2.5.0 - Python version: 3.6.9 - Bazel version (if compiling from source): - GCC/Compiler version (if compiling from source): - CUDA/cuDNN version: 11.2/ 8.1.0 - GPU model and memory: RTX 3090 You can collect some of this information using our environment capture [script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh) You can also obtain the TensorFlow version with: 1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"` 2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"` **Describe the current behavior** I convert resnet50 pytorch -> onnx -> tflite with int8 quantization. output value validation between pytorch <-> onnx, pytorch <-> pb, pytorch <-> tflite, pb <-> tflite input is same image with size 256, check output value "np.testing.assert_allclose(output1, output2, rtol=1e-3, atol=1e-05)" (using tflite interpreter only when i inference tflite "https://www.tensorflow.org/lite/guide/python?hl=ko") Max absolute difference: 0.00076199 in pytorch <-> onnx Max absolute difference: 0.00112534 in pytorch <-> pb Max absolute difference: 13.387602 in pytorch <-> tflite(quantized) Max absolute difference: 13.387438 in pb <-> tflite(quantized) it's same max absolute difference between tflite(no quantized) and something(pytorch, onnx, pb) ex) 0.0076~ in pytorch <-> tflite(no quant), 0.0011~ in pytorch <-> tflite(no quant) i don't know why occur this difference **Standalone code to reproduce the issue** Provide a reproducible test case that is the bare minimum necessary to generate the problem. If possible, please share a link to Colab/Jupyter/any notebook. **Other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. **pb to tflite log** 2021-10-13 09:18:56.162936: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0 2021-10-13 09:18:57.485452: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcuda.so.1 2021-10-13 09:18:57.511230: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.511916: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties: pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6 coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s 2021-10-13 09:18:57.511955: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0 2021-10-13 09:18:57.513717: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcublas.so.11 2021-10-13 09:18:57.513767: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcublasLt.so.11 2021-10-13 09:18:57.514354: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcufft.so.10 2021-10-13 09:18:57.514537: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcurand.so.10 2021-10-13 09:18:57.515198: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcusolver.so.11 2021-10-13 09:18:57.515720: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcusparse.so.11 2021-10-13 09:18:57.515866: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudnn.so.8 2021-10-13 09:18:57.515918: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.516398: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.516976: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0 2021-10-13 09:18:57.517199: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 FMA To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags. 2021-10-13 09:18:57.517766: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.518224: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties: pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6 coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s 2021-10-13 09:18:57.518272: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.518814: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.519243: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0 2021-10-13 09:18:57.519268: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0 2021-10-13 09:18:57.810376: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix: 2021-10-13 09:18:57.810410: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 0 2021-10-13 09:18:57.810420: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1277] 0: N 2021-10-13 09:18:57.810591: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.811162: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.811684: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.812186: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1418] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 21512 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 3090, pci bus id: 0000:01:00.0, compute capability: 8.6) 2021-10-13 09:18:58.498192: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:345] Ignored output_format. 2021-10-13 09:18:58.498225: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:348] Ignored drop_control_dependency. 2021-10-13 09:18:58.498234: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:354] Ignored change_concat_input_ranges. 2021-10-13 09:18:58.498881: I tensorflow/cc/saved_model/reader.cc:38] Reading SavedModel from: backbone_saved_model/ 2021-10-13 09:18:58.515289: I tensorflow/cc/saved_model/reader.cc:90] Reading meta graph with tags { serve } 2021-10-13 09:18:58.515331: I tensorflow/cc/saved_model/reader.cc:132] Reading SavedModel debug info (if present) from: backbone_saved_model/ 2021-10-13 09:18:58.515383: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix: 2021-10-13 09:18:58.515393: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 2021-10-13 09:18:58.527926: I tensorflow/cc/saved_model/loader.cc:206] Restoring SavedModel bundle. 2021-10-13 09:18:58.546224: I tensorflow/core/platform/profile_utils/cpu_utils.cc:114] CPU Frequency: 3699850000 Hz 2021-10-13 09:18:58.563849: I tensorflow/cc/saved_model/loader.cc:190] Running initialization op on SavedModel bundle at path: backbone_saved_model/ 2021-10-13 09:18:58.577967: I tensorflow/cc/saved_model/loader.cc:277] SavedModel load for tags { serve }; Status: success: OK. Took 79088 microseconds. 2021-10-13 09:18:58.657933: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:210] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable. 2021-10-13 09:18:58.675431: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.675985: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties: pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6 coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s 2021-10-13 09:18:58.676068: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.676635: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.677107: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0 2021-10-13 09:18:58.677148: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix: 2021-10-13 09:18:58.677157: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 0 2021-10-13 09:18:58.677165: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1277] 0: N 2021-10-13 09:18:58.677253: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.677779: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.678280: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1418] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 21512 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 3090, pci bus id: 0000:01:00.0, compute capability: 8.6) fully_quantize: 0, inference_type: 6, input_inference_type: 0, output_inference_type: 0
True
different output value in pytorch->onnx->tflite(int8 quantization) - <em>Please make sure that this is an issue related to performance of TensorFlow. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:performance_template</em> **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 20.04 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: - TensorFlow installed from (source or binary): tensorflow:2.5.0-gpu docker - TensorFlow version (use command below): 2.5.0 - Python version: 3.6.9 - Bazel version (if compiling from source): - GCC/Compiler version (if compiling from source): - CUDA/cuDNN version: 11.2/ 8.1.0 - GPU model and memory: RTX 3090 You can collect some of this information using our environment capture [script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh) You can also obtain the TensorFlow version with: 1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"` 2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"` **Describe the current behavior** I convert resnet50 pytorch -> onnx -> tflite with int8 quantization. output value validation between pytorch <-> onnx, pytorch <-> pb, pytorch <-> tflite, pb <-> tflite input is same image with size 256, check output value "np.testing.assert_allclose(output1, output2, rtol=1e-3, atol=1e-05)" (using tflite interpreter only when i inference tflite "https://www.tensorflow.org/lite/guide/python?hl=ko") Max absolute difference: 0.00076199 in pytorch <-> onnx Max absolute difference: 0.00112534 in pytorch <-> pb Max absolute difference: 13.387602 in pytorch <-> tflite(quantized) Max absolute difference: 13.387438 in pb <-> tflite(quantized) it's same max absolute difference between tflite(no quantized) and something(pytorch, onnx, pb) ex) 0.0076~ in pytorch <-> tflite(no quant), 0.0011~ in pytorch <-> tflite(no quant) i don't know why occur this difference **Standalone code to reproduce the issue** Provide a reproducible test case that is the bare minimum necessary to generate the problem. If possible, please share a link to Colab/Jupyter/any notebook. **Other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. **pb to tflite log** 2021-10-13 09:18:56.162936: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0 2021-10-13 09:18:57.485452: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcuda.so.1 2021-10-13 09:18:57.511230: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.511916: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties: pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6 coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s 2021-10-13 09:18:57.511955: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0 2021-10-13 09:18:57.513717: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcublas.so.11 2021-10-13 09:18:57.513767: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcublasLt.so.11 2021-10-13 09:18:57.514354: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcufft.so.10 2021-10-13 09:18:57.514537: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcurand.so.10 2021-10-13 09:18:57.515198: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcusolver.so.11 2021-10-13 09:18:57.515720: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcusparse.so.11 2021-10-13 09:18:57.515866: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudnn.so.8 2021-10-13 09:18:57.515918: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.516398: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.516976: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0 2021-10-13 09:18:57.517199: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 FMA To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags. 2021-10-13 09:18:57.517766: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.518224: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties: pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6 coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s 2021-10-13 09:18:57.518272: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.518814: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.519243: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0 2021-10-13 09:18:57.519268: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0 2021-10-13 09:18:57.810376: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix: 2021-10-13 09:18:57.810410: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 0 2021-10-13 09:18:57.810420: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1277] 0: N 2021-10-13 09:18:57.810591: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.811162: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.811684: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:57.812186: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1418] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 21512 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 3090, pci bus id: 0000:01:00.0, compute capability: 8.6) 2021-10-13 09:18:58.498192: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:345] Ignored output_format. 2021-10-13 09:18:58.498225: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:348] Ignored drop_control_dependency. 2021-10-13 09:18:58.498234: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:354] Ignored change_concat_input_ranges. 2021-10-13 09:18:58.498881: I tensorflow/cc/saved_model/reader.cc:38] Reading SavedModel from: backbone_saved_model/ 2021-10-13 09:18:58.515289: I tensorflow/cc/saved_model/reader.cc:90] Reading meta graph with tags { serve } 2021-10-13 09:18:58.515331: I tensorflow/cc/saved_model/reader.cc:132] Reading SavedModel debug info (if present) from: backbone_saved_model/ 2021-10-13 09:18:58.515383: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix: 2021-10-13 09:18:58.515393: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 2021-10-13 09:18:58.527926: I tensorflow/cc/saved_model/loader.cc:206] Restoring SavedModel bundle. 2021-10-13 09:18:58.546224: I tensorflow/core/platform/profile_utils/cpu_utils.cc:114] CPU Frequency: 3699850000 Hz 2021-10-13 09:18:58.563849: I tensorflow/cc/saved_model/loader.cc:190] Running initialization op on SavedModel bundle at path: backbone_saved_model/ 2021-10-13 09:18:58.577967: I tensorflow/cc/saved_model/loader.cc:277] SavedModel load for tags { serve }; Status: success: OK. Took 79088 microseconds. 2021-10-13 09:18:58.657933: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:210] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable. 2021-10-13 09:18:58.675431: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.675985: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties: pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6 coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s 2021-10-13 09:18:58.676068: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.676635: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.677107: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0 2021-10-13 09:18:58.677148: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix: 2021-10-13 09:18:58.677157: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 0 2021-10-13 09:18:58.677165: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1277] 0: N 2021-10-13 09:18:58.677253: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.677779: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero 2021-10-13 09:18:58.678280: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1418] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 21512 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 3090, pci bus id: 0000:01:00.0, compute capability: 8.6) fully_quantize: 0, inference_type: 6, input_inference_type: 0, output_inference_type: 0
perf
different output value in pytorch onnx tflite quantization please make sure that this is an issue related to performance of tensorflow as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag performance template system information have i written custom code as opposed to using a stock example script provided in tensorflow no os platform and distribution e g linux ubuntu ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow installed from source or binary tensorflow gpu docker tensorflow version use command below python version bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory rtx you can collect some of this information using our environment capture you can also obtain the tensorflow version with tf python c import tensorflow as tf print tf git version tf version tf python c import tensorflow as tf print tf version git version tf version version describe the current behavior i convert pytorch onnx tflite with quantization output value validation between pytorch onnx pytorch pb pytorch tflite pb tflite input is same image with size check output value np testing assert allclose rtol atol using tflite interpreter only when i inference tflite max absolute difference in pytorch onnx max absolute difference in pytorch pb max absolute difference in pytorch tflite quantized max absolute difference in pb tflite quantized it s same max absolute difference between tflite no quantized and something pytorch onnx pb ex in pytorch tflite no quant in pytorch tflite no quant i don t know why occur this difference standalone code to reproduce the issue provide a reproducible test case that is the bare minimum necessary to generate the problem if possible please share a link to colab jupyter any notebook other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached pb to tflite log i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcuda so i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc found device with properties pcibusid name nvidia geforce rtx computecapability coreclock corecount devicememorysize devicememorybandwidth s i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcublas so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcublaslt so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcufft so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcurand so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcusolver so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcusparse so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudnn so i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc adding visible gpu devices i tensorflow core platform cpu feature guard cc this tensorflow binary is optimized with oneapi deep neural network library onednn to use the following cpu instructions in performance critical operations fma to enable them in other operations rebuild tensorflow with the appropriate compiler flags i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc found device with properties pcibusid name nvidia geforce rtx computecapability coreclock corecount devicememorysize devicememorybandwidth s i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc adding visible gpu devices i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so i tensorflow core common runtime gpu gpu device cc device interconnect streamexecutor with strength edge matrix i tensorflow core common runtime gpu gpu device cc i tensorflow core common runtime gpu gpu device cc n i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc created tensorflow device job localhost replica task device gpu with mb memory physical gpu device name nvidia geforce rtx pci bus id compute capability w tensorflow compiler mlir lite python tf tfl flatbuffer helpers cc ignored output format w tensorflow compiler mlir lite python tf tfl flatbuffer helpers cc ignored drop control dependency w tensorflow compiler mlir lite python tf tfl flatbuffer helpers cc ignored change concat input ranges i tensorflow cc saved model reader cc reading savedmodel from backbone saved model i tensorflow cc saved model reader cc reading meta graph with tags serve i tensorflow cc saved model reader cc reading savedmodel debug info if present from backbone saved model i tensorflow core common runtime gpu gpu device cc device interconnect streamexecutor with strength edge matrix i tensorflow core common runtime gpu gpu device cc i tensorflow cc saved model loader cc restoring savedmodel bundle i tensorflow core platform profile utils cpu utils cc cpu frequency hz i tensorflow cc saved model loader cc running initialization op on savedmodel bundle at path backbone saved model i tensorflow cc saved model loader cc savedmodel load for tags serve status success ok took microseconds i tensorflow compiler mlir tensorflow utils dump mlir util cc disabling mlir crash reproducer set env var mlir crash reproducer directory to enable i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc found device with properties pcibusid name nvidia geforce rtx computecapability coreclock corecount devicememorysize devicememorybandwidth s i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc adding visible gpu devices i tensorflow core common runtime gpu gpu device cc device interconnect streamexecutor with strength edge matrix i tensorflow core common runtime gpu gpu device cc i tensorflow core common runtime gpu gpu device cc n i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc created tensorflow device job localhost replica task device gpu with mb memory physical gpu device name nvidia geforce rtx pci bus id compute capability fully quantize inference type input inference type output inference type
1
1,389
2,984,858,044
IssuesEvent
2015-07-18 12:12:17
akamsteeg/SwissArmyKnife
https://api.github.com/repos/akamsteeg/SwissArmyKnife
closed
string.Truncate() with suffix uses string.FormatWith(), instead of cheaper concatenation
Performance
`string.Truncate(int length, string suffix)` uses `string.FormatWith()` which in turn uses `string.Format()` internally. That's a bit wasteful, because a simple concatenation is far cheaper. `string.Truncate()` with `string.FormatWith()`: 106 ns. `string.Truncate()` with string concatenation: 64 ns.
True
string.Truncate() with suffix uses string.FormatWith(), instead of cheaper concatenation - `string.Truncate(int length, string suffix)` uses `string.FormatWith()` which in turn uses `string.Format()` internally. That's a bit wasteful, because a simple concatenation is far cheaper. `string.Truncate()` with `string.FormatWith()`: 106 ns. `string.Truncate()` with string concatenation: 64 ns.
perf
string truncate with suffix uses string formatwith instead of cheaper concatenation string truncate int length string suffix uses string formatwith which in turn uses string format internally that s a bit wasteful because a simple concatenation is far cheaper string truncate with string formatwith ns string truncate with string concatenation ns
1
29,754
14,256,177,664
IssuesEvent
2020-11-20 00:22:06
davidgarland/circa
https://api.github.com/repos/davidgarland/circa
closed
Differentiate between in-place and to-array higher-order functions.
enhancement memory performance
Right now: * foldl/foldr are to-value. * filter is in-place. * map is in-place. These are fairly sane defaults, but to-array variants of `filter` and `map`, as well as in-place variatns of `foldl/foldr`, would be nice to have for convenience and performance reasons depending upon situation. A naming convention differentiating between these (perhaps `_to` for "to-value"/"to-array" versions?) would be nice.
True
Differentiate between in-place and to-array higher-order functions. - Right now: * foldl/foldr are to-value. * filter is in-place. * map is in-place. These are fairly sane defaults, but to-array variants of `filter` and `map`, as well as in-place variatns of `foldl/foldr`, would be nice to have for convenience and performance reasons depending upon situation. A naming convention differentiating between these (perhaps `_to` for "to-value"/"to-array" versions?) would be nice.
perf
differentiate between in place and to array higher order functions right now foldl foldr are to value filter is in place map is in place these are fairly sane defaults but to array variants of filter and map as well as in place variatns of foldl foldr would be nice to have for convenience and performance reasons depending upon situation a naming convention differentiating between these perhaps to for to value to array versions would be nice
1
5,271
4,864,245,574
IssuesEvent
2016-11-14 17:27:48
elmsln/elmsln
https://api.github.com/repos/elmsln/elmsln
closed
Disable aggregation for better performance
scale / performance
Weird but true. There's no gain from aggregating resources.
True
Disable aggregation for better performance - Weird but true. There's no gain from aggregating resources.
perf
disable aggregation for better performance weird but true there s no gain from aggregating resources
1
9,099
6,760,875,616
IssuesEvent
2017-10-24 22:21:07
Scratchcat1/AATC
https://api.github.com/repos/Scratchcat1/AATC
opened
Monitor viewer render inefficiency
performance
https://github.com/Scratchcat1/AATC/blob/dd43ecaff710c77337c95ffbc70f30e791883068/AATC_Monitor_Viewer.py#L94-L95 These lines need not be calculated if object is not to be drawn.
True
Monitor viewer render inefficiency - https://github.com/Scratchcat1/AATC/blob/dd43ecaff710c77337c95ffbc70f30e791883068/AATC_Monitor_Viewer.py#L94-L95 These lines need not be calculated if object is not to be drawn.
perf
monitor viewer render inefficiency these lines need not be calculated if object is not to be drawn
1
27,070
13,171,778,431
IssuesEvent
2020-08-11 17:15:43
Giveth/giveth-dapp
https://api.github.com/repos/Giveth/giveth-dapp
closed
Loading times are extremely high on the dapp
bug performance
Loading times shouldn't be as high specially on the main page
True
Loading times are extremely high on the dapp - Loading times shouldn't be as high specially on the main page
perf
loading times are extremely high on the dapp loading times shouldn t be as high specially on the main page
1
48,151
25,381,221,055
IssuesEvent
2022-11-21 17:42:15
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
kv: pebbleReadOnly constructed in replicaRaftStorage.Entries even when reading from entry cache
C-performance A-kv-replication T-kv-replication
I noticed in a recent round of benchmarking that we construct a `pebbleReadOnly` in `replicaRaftStorage.Entries` even in the vastly common case where all entries are served out of the raft entry cache: https://github.com/cockroachdb/cockroach/blob/2a041d61d021f563828597fe9961f54727419c1d/pkg/kv/kvserver/replica_raftstorage.go#L102 This is wasteful. I would have made the fix of pushing this into `entries` myself, but there's was just enough friction due to `term`'s [use of the `reader`](https://github.com/cockroachdb/cockroach/blob/2a041d61d021f563828597fe9961f54727419c1d/pkg/kv/kvserver/replica_raftstorage.go#L296) that I figured I'd leave it to the experts. Is it worth asking whether this code needs to operate on a LSM snapshot? There should not be concurrent mutation to the state it is looking at.
True
kv: pebbleReadOnly constructed in replicaRaftStorage.Entries even when reading from entry cache - I noticed in a recent round of benchmarking that we construct a `pebbleReadOnly` in `replicaRaftStorage.Entries` even in the vastly common case where all entries are served out of the raft entry cache: https://github.com/cockroachdb/cockroach/blob/2a041d61d021f563828597fe9961f54727419c1d/pkg/kv/kvserver/replica_raftstorage.go#L102 This is wasteful. I would have made the fix of pushing this into `entries` myself, but there's was just enough friction due to `term`'s [use of the `reader`](https://github.com/cockroachdb/cockroach/blob/2a041d61d021f563828597fe9961f54727419c1d/pkg/kv/kvserver/replica_raftstorage.go#L296) that I figured I'd leave it to the experts. Is it worth asking whether this code needs to operate on a LSM snapshot? There should not be concurrent mutation to the state it is looking at.
perf
kv pebblereadonly constructed in replicaraftstorage entries even when reading from entry cache i noticed in a recent round of benchmarking that we construct a pebblereadonly in replicaraftstorage entries even in the vastly common case where all entries are served out of the raft entry cache this is wasteful i would have made the fix of pushing this into entries myself but there s was just enough friction due to term s that i figured i d leave it to the experts is it worth asking whether this code needs to operate on a lsm snapshot there should not be concurrent mutation to the state it is looking at
1
3,611
2,773,523,446
IssuesEvent
2015-05-03 18:24:31
RexOps/Rex
https://api.github.com/repos/RexOps/Rex
closed
PATH environment problem
documentation
Hi , in /etc/profile , I export some path to PATH environment , so I use source_global_profile(1) command to make sure my run command will work fine. But this does not work , because rex reset PATH environment after sourcing /etc/profile. the workaround for me is sourcing again before run command run(". /etc/profile >/dev/null 2>&1 ; mvn -version"); == part of /etc/profile == export M2_HOME=/usr/local/apache-maven export M2=$M2_HOME/bin export PATH=$M2:$PATH == part of log == [2015-05-03 16:27:06] DEBUG - SSH/executing: . /etc/profile >/dev/null 2>&1 ; LC_ALL=C ; export LC_ALL; PATH=/bin:/sbin:/usr/bin:/usr/sbin:/usr/local/bin:/usr/local/sbin:/usr/pkg/bin:/usr/pkg/sbin; export PATH; mvn -version
1.0
PATH environment problem - Hi , in /etc/profile , I export some path to PATH environment , so I use source_global_profile(1) command to make sure my run command will work fine. But this does not work , because rex reset PATH environment after sourcing /etc/profile. the workaround for me is sourcing again before run command run(". /etc/profile >/dev/null 2>&1 ; mvn -version"); == part of /etc/profile == export M2_HOME=/usr/local/apache-maven export M2=$M2_HOME/bin export PATH=$M2:$PATH == part of log == [2015-05-03 16:27:06] DEBUG - SSH/executing: . /etc/profile >/dev/null 2>&1 ; LC_ALL=C ; export LC_ALL; PATH=/bin:/sbin:/usr/bin:/usr/sbin:/usr/local/bin:/usr/local/sbin:/usr/pkg/bin:/usr/pkg/sbin; export PATH; mvn -version
non_perf
path environment problem hi in etc profile i export some path to path environment so i use source global profile command to make sure my run command will work fine but this does not work because rex reset path environment after sourcing etc profile the workaround for me is sourcing again before run command run etc profile dev null mvn version part of etc profile export home usr local apache maven export home bin export path path part of log debug ssh executing etc profile dev null lc all c export lc all path bin sbin usr bin usr sbin usr local bin usr local sbin usr pkg bin usr pkg sbin export path mvn version
0
38,084
8,408,510,994
IssuesEvent
2018-10-12 02:04:36
robot-lab/judyst-link-analysis
https://api.github.com/repos/robot-lab/judyst-link-analysis
closed
Выделить webcrawler в компоненту
code maintenance task
Нужно перенести компонент сбора данных в репозиторий judyst-web-crawler
1.0
Выделить webcrawler в компоненту - Нужно перенести компонент сбора данных в репозиторий judyst-web-crawler
non_perf
выделить webcrawler в компоненту нужно перенести компонент сбора данных в репозиторий judyst web crawler
0
159,619
20,085,863,570
IssuesEvent
2022-02-05 01:03:11
doc-ai/tensorflow-ios-framework
https://api.github.com/repos/doc-ai/tensorflow-ios-framework
opened
CVE-2020-26266 (Medium) detected in tensorflowv2.0.2
security vulnerability
## CVE-2020-26266 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflowv2.0.2</b></p></summary> <p> <p>An Open Source Machine Learning Framework for Everyone</p> <p>Library home page: <a href=https://github.com/tensorflow/tensorflow.git>https://github.com/tensorflow/tensorflow.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In affected versions of TensorFlow under certain cases a saved model can trigger use of uninitialized values during code execution. This is caused by having tensor buffers be filled with the default value of the type but forgetting to default initialize the quantized floating point types in Eigen. This is fixed in versions 1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, and 2.4.0. <p>Publish Date: 2020-12-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26266>CVE-2020-26266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qhxx-j73r-qpm2">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qhxx-j73r-qpm2</a></p> <p>Release Date: 2020-12-10</p> <p>Fix Resolution: tensorflow-1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, 2.4.0,tensorflow-cpu-2.1.3, 2.2.2, 2.3.2, 2.4.0,tensorflow-gpu-1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, 2.4.0</p> </p> </details> <p></p>
True
CVE-2020-26266 (Medium) detected in tensorflowv2.0.2 - ## CVE-2020-26266 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflowv2.0.2</b></p></summary> <p> <p>An Open Source Machine Learning Framework for Everyone</p> <p>Library home page: <a href=https://github.com/tensorflow/tensorflow.git>https://github.com/tensorflow/tensorflow.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In affected versions of TensorFlow under certain cases a saved model can trigger use of uninitialized values during code execution. This is caused by having tensor buffers be filled with the default value of the type but forgetting to default initialize the quantized floating point types in Eigen. This is fixed in versions 1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, and 2.4.0. <p>Publish Date: 2020-12-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26266>CVE-2020-26266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qhxx-j73r-qpm2">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qhxx-j73r-qpm2</a></p> <p>Release Date: 2020-12-10</p> <p>Fix Resolution: tensorflow-1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, 2.4.0,tensorflow-cpu-2.1.3, 2.2.2, 2.3.2, 2.4.0,tensorflow-gpu-1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, 2.4.0</p> </p> </details> <p></p>
non_perf
cve medium detected in cve medium severity vulnerability vulnerable library an open source machine learning framework for everyone library home page a href found in base branch master vulnerable source files tensorflow ios framework tensorflow framework headers third party unsupported eigen src fixedpoint fixedpointtypes h tensorflow ios framework tensorflow framework headers third party unsupported eigen src fixedpoint fixedpointtypes h tensorflow ios framework tensorflow framework headers third party unsupported eigen src fixedpoint fixedpointtypes h vulnerability details in affected versions of tensorflow under certain cases a saved model can trigger use of uninitialized values during code execution this is caused by having tensor buffers be filled with the default value of the type but forgetting to default initialize the quantized floating point types in eigen this is fixed in versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu
0
45,112
18,412,318,165
IssuesEvent
2021-10-13 07:36:31
craftbuy/uptime
https://api.github.com/repos/craftbuy/uptime
closed
🛑 CraftBuy.ru Services is down
status craft-buy-ru-services
In [`3e9b722`](https://github.com/craftbuy/uptime/commit/3e9b722fda208fd98fe607008fcad6a54eb85567 ), CraftBuy.ru Services ($CRAFTBUY_SERVICES) was **down**: - HTTP code: 0 - Response time: 0 ms
1.0
🛑 CraftBuy.ru Services is down - In [`3e9b722`](https://github.com/craftbuy/uptime/commit/3e9b722fda208fd98fe607008fcad6a54eb85567 ), CraftBuy.ru Services ($CRAFTBUY_SERVICES) was **down**: - HTTP code: 0 - Response time: 0 ms
non_perf
🛑 craftbuy ru services is down in craftbuy ru services craftbuy services was down http code response time ms
0
234,633
25,880,477,572
IssuesEvent
2022-12-14 10:54:45
rsoreq/WebGoat
https://api.github.com/repos/rsoreq/WebGoat
closed
CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - autoclosed
security vulnerability
## CVE-2020-7733 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary> <p>Lightweight JavaScript-based user-agent string parser</p> <p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p> <p>Path to dependency file: /docs/package.json</p> <p>Path to vulnerable library: /docs/node_modules/ua-parser-js/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA. <p>Publish Date: 2020-09-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7733>CVE-2020-7733</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p> <p>Release Date: 2020-09-16</p> <p>Fix Resolution (ua-parser-js): 0.7.22</p> <p>Direct dependency fix Resolution (browser-sync): 2.26.9</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - autoclosed - ## CVE-2020-7733 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary> <p>Lightweight JavaScript-based user-agent string parser</p> <p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p> <p>Path to dependency file: /docs/package.json</p> <p>Path to vulnerable library: /docs/node_modules/ua-parser-js/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA. <p>Publish Date: 2020-09-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7733>CVE-2020-7733</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p> <p>Release Date: 2020-09-16</p> <p>Fix Resolution (ua-parser-js): 0.7.22</p> <p>Direct dependency fix Resolution (browser-sync): 2.26.9</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_perf
cve high detected in ua parser js tgz autoclosed cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file docs package json path to vulnerable library docs node modules ua parser js package json dependency hierarchy browser sync tgz root library x ua parser js tgz vulnerable library found in base branch develop vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos via the regex for redmi phones and mi pad tablets ua publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ua parser js direct dependency fix resolution browser sync rescue worker helmet automatic remediation is available for this issue
0
58,413
8,257,538,104
IssuesEvent
2018-09-13 05:42:12
Microsoft/pxt-microbit
https://api.github.com/repos/Microsoft/pxt-microbit
closed
"Tutorials" & "Games" except "Examples" don't have brief introduction on home page
documentation homescreen
**Repro steps:** 1. Navigate to https://makecode.microbit.org/beta 2. Scroll down the scroll bar and click examples under "Tutorials" & "Games" one by one 3. Observe the opened window **Expected result:** Each tutorial and game should have a brief introduction ![compare911](https://user-images.githubusercontent.com/23466737/45353698-22790300-b5ee-11e8-9c68-adc9051e3c6b.PNG) **Actual result:** "Tutorials" & "Games" except "Examples" don't have brief introduction ![issue91104](https://user-images.githubusercontent.com/23466737/45353681-17be6e00-b5ee-11e8-839e-5f2e3a6a89ed.gif)
1.0
"Tutorials" & "Games" except "Examples" don't have brief introduction on home page - **Repro steps:** 1. Navigate to https://makecode.microbit.org/beta 2. Scroll down the scroll bar and click examples under "Tutorials" & "Games" one by one 3. Observe the opened window **Expected result:** Each tutorial and game should have a brief introduction ![compare911](https://user-images.githubusercontent.com/23466737/45353698-22790300-b5ee-11e8-9c68-adc9051e3c6b.PNG) **Actual result:** "Tutorials" & "Games" except "Examples" don't have brief introduction ![issue91104](https://user-images.githubusercontent.com/23466737/45353681-17be6e00-b5ee-11e8-839e-5f2e3a6a89ed.gif)
non_perf
tutorials games except examples don t have brief introduction on home page repro steps navigate to scroll down the scroll bar and click examples under tutorials games one by one observe the opened window expected result each tutorial and game should have a brief introduction actual result tutorials games except examples don t have brief introduction
0
137,986
18,769,555,189
IssuesEvent
2021-11-06 15:30:31
samqws-marketing/box_box-ui-elements
https://api.github.com/repos/samqws-marketing/box_box-ui-elements
opened
WS-2019-0339 (High) detected in bin-links-1.1.3.tgz, bin-links-1.1.2.tgz
security vulnerability
## WS-2019-0339 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bin-links-1.1.3.tgz</b>, <b>bin-links-1.1.2.tgz</b></p></summary> <p> <details><summary><b>bin-links-1.1.3.tgz</b></p></summary> <p>JavaScript package binary linker</p> <p>Library home page: <a href="https://registry.npmjs.org/bin-links/-/bin-links-1.1.3.tgz">https://registry.npmjs.org/bin-links/-/bin-links-1.1.3.tgz</a></p> <p>Path to dependency file: box_box-ui-elements/package.json</p> <p>Path to vulnerable library: box_box-ui-elements/node_modules/bin-links/package.json</p> <p> Dependency Hierarchy: - semantic-release-16.0.2.tgz (Root Library) - npm-6.0.0.tgz - npm-6.13.1.tgz - :x: **bin-links-1.1.3.tgz** (Vulnerable Library) </details> <details><summary><b>bin-links-1.1.2.tgz</b></p></summary> <p>JavaScript package binary linker</p> <p>Library home page: <a href="https://registry.npmjs.org/bin-links/-/bin-links-1.1.2.tgz">https://registry.npmjs.org/bin-links/-/bin-links-1.1.2.tgz</a></p> <p>Path to dependency file: box_box-ui-elements/package.json</p> <p>Path to vulnerable library: box_box-ui-elements/node_modules/bin-links/package.json</p> <p> Dependency Hierarchy: - storybook-5.3.9.tgz (Root Library) - cli-5.3.9.tgz - pkg-add-deps-0.1.0.tgz - libnpm-2.0.1.tgz - :x: **bin-links-1.1.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In bin-links, versions prior to v1.1.6 are vulnerable to a Global 'node_modules' Binary Overwrite. It fails to prevent globally-installed binaries to be overwritten by other package installs. <p>Publish Date: 2019-12-11 <p>URL: <a href=https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69>WS-2019-0339</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69">https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69</a></p> <p>Release Date: 2019-12-17</p> <p>Fix Resolution: bin-links - 1.1.6</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"bin-links","packageVersion":"1.1.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"semantic-release:16.0.2;@semantic-release/npm:6.0.0;npm:6.13.1;bin-links:1.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bin-links - 1.1.6"},{"packageType":"javascript/Node.js","packageName":"bin-links","packageVersion":"1.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"storybook:5.3.9;@storybook/cli:5.3.9;pkg-add-deps:0.1.0;libnpm:2.0.1;bin-links:1.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bin-links - 1.1.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0339","vulnerabilityDetails":"In bin-links, versions prior to v1.1.6 are vulnerable to a Global \u0027node_modules\u0027 Binary Overwrite. It fails to prevent globally-installed binaries to be overwritten by other package installs.","vulnerabilityUrl":"https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0339 (High) detected in bin-links-1.1.3.tgz, bin-links-1.1.2.tgz - ## WS-2019-0339 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bin-links-1.1.3.tgz</b>, <b>bin-links-1.1.2.tgz</b></p></summary> <p> <details><summary><b>bin-links-1.1.3.tgz</b></p></summary> <p>JavaScript package binary linker</p> <p>Library home page: <a href="https://registry.npmjs.org/bin-links/-/bin-links-1.1.3.tgz">https://registry.npmjs.org/bin-links/-/bin-links-1.1.3.tgz</a></p> <p>Path to dependency file: box_box-ui-elements/package.json</p> <p>Path to vulnerable library: box_box-ui-elements/node_modules/bin-links/package.json</p> <p> Dependency Hierarchy: - semantic-release-16.0.2.tgz (Root Library) - npm-6.0.0.tgz - npm-6.13.1.tgz - :x: **bin-links-1.1.3.tgz** (Vulnerable Library) </details> <details><summary><b>bin-links-1.1.2.tgz</b></p></summary> <p>JavaScript package binary linker</p> <p>Library home page: <a href="https://registry.npmjs.org/bin-links/-/bin-links-1.1.2.tgz">https://registry.npmjs.org/bin-links/-/bin-links-1.1.2.tgz</a></p> <p>Path to dependency file: box_box-ui-elements/package.json</p> <p>Path to vulnerable library: box_box-ui-elements/node_modules/bin-links/package.json</p> <p> Dependency Hierarchy: - storybook-5.3.9.tgz (Root Library) - cli-5.3.9.tgz - pkg-add-deps-0.1.0.tgz - libnpm-2.0.1.tgz - :x: **bin-links-1.1.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In bin-links, versions prior to v1.1.6 are vulnerable to a Global 'node_modules' Binary Overwrite. It fails to prevent globally-installed binaries to be overwritten by other package installs. <p>Publish Date: 2019-12-11 <p>URL: <a href=https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69>WS-2019-0339</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69">https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69</a></p> <p>Release Date: 2019-12-17</p> <p>Fix Resolution: bin-links - 1.1.6</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"bin-links","packageVersion":"1.1.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"semantic-release:16.0.2;@semantic-release/npm:6.0.0;npm:6.13.1;bin-links:1.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bin-links - 1.1.6"},{"packageType":"javascript/Node.js","packageName":"bin-links","packageVersion":"1.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"storybook:5.3.9;@storybook/cli:5.3.9;pkg-add-deps:0.1.0;libnpm:2.0.1;bin-links:1.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bin-links - 1.1.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0339","vulnerabilityDetails":"In bin-links, versions prior to v1.1.6 are vulnerable to a Global \u0027node_modules\u0027 Binary Overwrite. It fails to prevent globally-installed binaries to be overwritten by other package installs.","vulnerabilityUrl":"https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_perf
ws high detected in bin links tgz bin links tgz ws high severity vulnerability vulnerable libraries bin links tgz bin links tgz bin links tgz javascript package binary linker library home page a href path to dependency file box box ui elements package json path to vulnerable library box box ui elements node modules bin links package json dependency hierarchy semantic release tgz root library npm tgz npm tgz x bin links tgz vulnerable library bin links tgz javascript package binary linker library home page a href path to dependency file box box ui elements package json path to vulnerable library box box ui elements node modules bin links package json dependency hierarchy storybook tgz root library cli tgz pkg add deps tgz libnpm tgz x bin links tgz vulnerable library found in head commit a href found in base branch master vulnerability details in bin links versions prior to are vulnerable to a global node modules binary overwrite it fails to prevent globally installed binaries to be overwritten by other package installs publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bin links isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree semantic release semantic release npm npm bin links isminimumfixversionavailable true minimumfixversion bin links packagetype javascript node js packagename bin links packageversion packagefilepaths istransitivedependency true dependencytree storybook storybook cli pkg add deps libnpm bin links isminimumfixversionavailable true minimumfixversion bin links basebranches vulnerabilityidentifier ws vulnerabilitydetails in bin links versions prior to are vulnerable to a global modules binary overwrite it fails to prevent globally installed binaries to be overwritten by other package installs vulnerabilityurl
0
7,613
3,105,356,852
IssuesEvent
2015-08-31 20:32:56
softlayer/sl-ember-components
https://api.github.com/repos/softlayer/sl-ember-components
opened
Evaluate "selectionDirection" property
0 - Backlog documentation question research sl-textarea
The `selectionDirection` attribute of a `textarea` element can be set but according to the spec - http://www.w3.org/TR/html5/forms.html#dom-textarea/input-selectiondirection - this value does not actually do anything from a "setting" perspective, though it does contain a value during text selection to be used in a "getter" perspective. Do we want to adjust the description of this property at #/demos/sl-textarea to align with this? This then leads into the next item which is this attribute is set inconsistently across browsers. Chrome on OS X at least does not have this value populated. Given this (and other browsers should be tested too) do we even want to expose this property? What role do we want this library to play? Normalize across things or provide support pass-though and those environments that do will?
1.0
Evaluate "selectionDirection" property - The `selectionDirection` attribute of a `textarea` element can be set but according to the spec - http://www.w3.org/TR/html5/forms.html#dom-textarea/input-selectiondirection - this value does not actually do anything from a "setting" perspective, though it does contain a value during text selection to be used in a "getter" perspective. Do we want to adjust the description of this property at #/demos/sl-textarea to align with this? This then leads into the next item which is this attribute is set inconsistently across browsers. Chrome on OS X at least does not have this value populated. Given this (and other browsers should be tested too) do we even want to expose this property? What role do we want this library to play? Normalize across things or provide support pass-though and those environments that do will?
non_perf
evaluate selectiondirection property the selectiondirection attribute of a textarea element can be set but according to the spec this value does not actually do anything from a setting perspective though it does contain a value during text selection to be used in a getter perspective do we want to adjust the description of this property at demos sl textarea to align with this this then leads into the next item which is this attribute is set inconsistently across browsers chrome on os x at least does not have this value populated given this and other browsers should be tested too do we even want to expose this property what role do we want this library to play normalize across things or provide support pass though and those environments that do will
0
44,759
23,757,210,725
IssuesEvent
2022-09-01 05:07:47
datafuselabs/databend
https://api.github.com/repos/datafuselabs/databend
closed
optimizer_statistics_exact is missing in new planner
C-performance A-planner
**Summary** `optimizer_statistics_exact` works in old planner, but it's missing in new planner
True
optimizer_statistics_exact is missing in new planner - **Summary** `optimizer_statistics_exact` works in old planner, but it's missing in new planner
perf
optimizer statistics exact is missing in new planner summary optimizer statistics exact works in old planner but it s missing in new planner
1
222,274
24,697,454,546
IssuesEvent
2022-10-19 13:09:09
rsoreq/kendo-ui-core
https://api.github.com/repos/rsoreq/kendo-ui-core
opened
CVE-2022-3517 (High) detected in multiple libraries
security vulnerability
## CVE-2022-3517 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimatch-0.2.14.tgz</b>, <b>minimatch-2.0.10.tgz</b>, <b>minimatch-3.0.4.tgz</b>, <b>minimatch-1.0.0.tgz</b>, <b>minimatch-0.3.0.tgz</b></p></summary> <p> <details><summary><b>minimatch-0.2.14.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.2.14.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.2.14.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/globule/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - amd-optimize-0.6.1.tgz (Root Library) - vinyl-fs-0.3.14.tgz - glob-watcher-0.0.6.tgz - gaze-0.5.2.tgz - globule-0.1.0.tgz - :x: **minimatch-0.2.14.tgz** (Vulnerable Library) </details> <details><summary><b>minimatch-2.0.10.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-2.0.10.tgz">https://registry.npmjs.org/minimatch/-/minimatch-2.0.10.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - glob-5.0.15.tgz (Root Library) - :x: **minimatch-2.0.10.tgz** (Vulnerable Library) </details> <details><summary><b>minimatch-3.0.4.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/karma/node_modules/minimatch/package.json,/node_modules/accord/node_modules/minimatch/package.json,/docs-aspnet/node_modules/minimatch/package.json,/node_modules/gulp-match/node_modules/minimatch/package.json,/node_modules/mocha/node_modules/minimatch/package.json,/docs/node_modules/minimatch/package.json,/node_modules/rimraf/node_modules/minimatch/package.json,/node_modules/multimatch/node_modules/minimatch/package.json,/node_modules/gulp-jshint/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - gulp-if-2.0.2.tgz (Root Library) - gulp-match-1.1.0.tgz - :x: **minimatch-3.0.4.tgz** (Vulnerable Library) </details> <details><summary><b>minimatch-1.0.0.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-1.0.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-1.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/gulp-ignore/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - gulp-ignore-1.2.1.tgz (Root Library) - gulp-match-0.2.1.tgz - :x: **minimatch-1.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>minimatch-0.3.0.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/gulp-subset-process/node_modules/minimatch/package.json,/node_modules/cli/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - jshint-2.8.0.tgz (Root Library) - cli-0.6.6.tgz - glob-3.2.11.tgz - :x: **minimatch-0.3.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/rsoreq/kendo-ui-core/commit/62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad">62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service. <p>Publish Date: 2022-10-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-17</p> <p>Fix Resolution: minimatch - 3.0.5</p> </p> </details> <p></p>
True
CVE-2022-3517 (High) detected in multiple libraries - ## CVE-2022-3517 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimatch-0.2.14.tgz</b>, <b>minimatch-2.0.10.tgz</b>, <b>minimatch-3.0.4.tgz</b>, <b>minimatch-1.0.0.tgz</b>, <b>minimatch-0.3.0.tgz</b></p></summary> <p> <details><summary><b>minimatch-0.2.14.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.2.14.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.2.14.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/globule/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - amd-optimize-0.6.1.tgz (Root Library) - vinyl-fs-0.3.14.tgz - glob-watcher-0.0.6.tgz - gaze-0.5.2.tgz - globule-0.1.0.tgz - :x: **minimatch-0.2.14.tgz** (Vulnerable Library) </details> <details><summary><b>minimatch-2.0.10.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-2.0.10.tgz">https://registry.npmjs.org/minimatch/-/minimatch-2.0.10.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - glob-5.0.15.tgz (Root Library) - :x: **minimatch-2.0.10.tgz** (Vulnerable Library) </details> <details><summary><b>minimatch-3.0.4.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/karma/node_modules/minimatch/package.json,/node_modules/accord/node_modules/minimatch/package.json,/docs-aspnet/node_modules/minimatch/package.json,/node_modules/gulp-match/node_modules/minimatch/package.json,/node_modules/mocha/node_modules/minimatch/package.json,/docs/node_modules/minimatch/package.json,/node_modules/rimraf/node_modules/minimatch/package.json,/node_modules/multimatch/node_modules/minimatch/package.json,/node_modules/gulp-jshint/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - gulp-if-2.0.2.tgz (Root Library) - gulp-match-1.1.0.tgz - :x: **minimatch-3.0.4.tgz** (Vulnerable Library) </details> <details><summary><b>minimatch-1.0.0.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-1.0.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-1.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/gulp-ignore/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - gulp-ignore-1.2.1.tgz (Root Library) - gulp-match-0.2.1.tgz - :x: **minimatch-1.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>minimatch-0.3.0.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/gulp-subset-process/node_modules/minimatch/package.json,/node_modules/cli/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - jshint-2.8.0.tgz (Root Library) - cli-0.6.6.tgz - glob-3.2.11.tgz - :x: **minimatch-0.3.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/rsoreq/kendo-ui-core/commit/62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad">62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service. <p>Publish Date: 2022-10-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-17</p> <p>Fix Resolution: minimatch - 3.0.5</p> </p> </details> <p></p>
non_perf
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries minimatch tgz minimatch tgz minimatch tgz minimatch tgz minimatch tgz minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules globule node modules minimatch package json dependency hierarchy amd optimize tgz root library vinyl fs tgz glob watcher tgz gaze tgz globule tgz x minimatch tgz vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules minimatch package json dependency hierarchy glob tgz root library x minimatch tgz vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules karma node modules minimatch package json node modules accord node modules minimatch package json docs aspnet node modules minimatch package json node modules gulp match node modules minimatch package json node modules mocha node modules minimatch package json docs node modules minimatch package json node modules rimraf node modules minimatch package json node modules multimatch node modules minimatch package json node modules gulp jshint node modules minimatch package json dependency hierarchy gulp if tgz root library gulp match tgz x minimatch tgz vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules gulp ignore node modules minimatch package json dependency hierarchy gulp ignore tgz root library gulp match tgz x minimatch tgz vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules gulp subset process node modules minimatch package json node modules cli node modules minimatch package json dependency hierarchy jshint tgz root library cli tgz glob tgz x minimatch tgz vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability was found in the minimatch package this flaw allows a regular expression denial of service redos when calling the braceexpand function with specific arguments resulting in a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimatch
0
91,897
18,737,234,364
IssuesEvent
2021-11-04 09:17:31
cosmos/ibc-go
https://api.github.com/repos/cosmos/ibc-go
closed
Nitpicks: ICA Audit (TrySendTxFlow)
good first issue code-hygiene 27-interchain-accounts audit-ica
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Summary of Issue The following nits arose as part of the audit. - [x] TrySendTx (relay.go) channel not found should return for which port id - [ ] ~~Return nil instead of []byte{} in all returns for keeper/relay.go~~ - [x] Require that we pass in an array of sdk.Msg instead of single sdk.Msg (keeper/keeper.go) - [ ] ~~Add channel/port id in capability not found error (createOutgoingPacket - keeper/relay.go)~~ - [x] Add channel/port id in get next send sequence not found error (createOutgoingPacket - keeper/relay.go) - [x] Potentially panic on errors that indicate bugs in code? (keeper/keeper.go) - [x] ErrUnkownPacketData - ErrUnknownDataType (errors.go) - [x] relay.go: AuthenticateTx: Get Interchain address first then loop through signers and return an error if the expected signer is not ICA address - [x] relay.go: ExectureTx reduce code by returning when error occurs on executeMsg. Add comment for how cache context is functioning (atomic execution) - [ ] ~~Fix error type and wrapping in module.go AcknowledgePacket~~ ____ #### For Admin Use - [ ] Not duplicate issue - [ ] Appropriate labels applied - [ ] Appropriate contributors tagged/assigned
1.0
Nitpicks: ICA Audit (TrySendTxFlow) - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Summary of Issue The following nits arose as part of the audit. - [x] TrySendTx (relay.go) channel not found should return for which port id - [ ] ~~Return nil instead of []byte{} in all returns for keeper/relay.go~~ - [x] Require that we pass in an array of sdk.Msg instead of single sdk.Msg (keeper/keeper.go) - [ ] ~~Add channel/port id in capability not found error (createOutgoingPacket - keeper/relay.go)~~ - [x] Add channel/port id in get next send sequence not found error (createOutgoingPacket - keeper/relay.go) - [x] Potentially panic on errors that indicate bugs in code? (keeper/keeper.go) - [x] ErrUnkownPacketData - ErrUnknownDataType (errors.go) - [x] relay.go: AuthenticateTx: Get Interchain address first then loop through signers and return an error if the expected signer is not ICA address - [x] relay.go: ExectureTx reduce code by returning when error occurs on executeMsg. Add comment for how cache context is functioning (atomic execution) - [ ] ~~Fix error type and wrapping in module.go AcknowledgePacket~~ ____ #### For Admin Use - [ ] Not duplicate issue - [ ] Appropriate labels applied - [ ] Appropriate contributors tagged/assigned
non_perf
nitpicks ica audit trysendtxflow ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ summary of issue the following nits arose as part of the audit trysendtx relay go channel not found should return for which port id return nil instead of byte in all returns for keeper relay go require that we pass in an array of sdk msg instead of single sdk msg keeper keeper go add channel port id in capability not found error createoutgoingpacket keeper relay go add channel port id in get next send sequence not found error createoutgoingpacket keeper relay go potentially panic on errors that indicate bugs in code keeper keeper go errunkownpacketdata errunknowndatatype errors go relay go authenticatetx get interchain address first then loop through signers and return an error if the expected signer is not ica address relay go execturetx reduce code by returning when error occurs on executemsg add comment for how cache context is functioning atomic execution fix error type and wrapping in module go acknowledgepacket for admin use not duplicate issue appropriate labels applied appropriate contributors tagged assigned
0
221,732
17,026,196,841
IssuesEvent
2021-07-03 15:11:54
WebDevJ/ski
https://api.github.com/repos/WebDevJ/ski
closed
Update README
documentation
* Comments; what was done, what wasn't, features added & known bugs If any. * Provide a way to view the completed code and run it; local environment install/run instructions.
1.0
Update README - * Comments; what was done, what wasn't, features added & known bugs If any. * Provide a way to view the completed code and run it; local environment install/run instructions.
non_perf
update readme comments what was done what wasn t features added known bugs if any provide a way to view the completed code and run it local environment install run instructions
0
44,483
23,651,188,800
IssuesEvent
2022-08-26 06:49:28
clj-commons/aleph
https://api.github.com/repos/clj-commons/aleph
closed
Improve wrap-future
performance good first issue
This was uncovered in the process of reviewing #608. See my comments there for more. In short, `wrap-future` checks `.isSuccess` on a future, when it really means to check `.isDone`. If it succeeded, it creates an immediate success-deferred. If not, it attaches handlers. In the case of a failed/canceled future, it does a lot of unnecessary work before immediately invoking the handler: fn thread frame binding, handler object creation, and scheduling the handlers to run (potentially) on another thread. This is also kind of confusing because we're checking the wrong condition. To do: - [ ] Switch to `.isDone` from `.isSuccess` in `wrap-future` - [ ] If `.isDone`, choose between returning success- and error-deferreds based on the future's status. For consistency, we can reuse the `operation-complete` handler, but add a flag to tell it to use/ignore executors (so we can run it on the current thread) - [ ] Verify all tests still work, `test-classloader` in particular
True
Improve wrap-future - This was uncovered in the process of reviewing #608. See my comments there for more. In short, `wrap-future` checks `.isSuccess` on a future, when it really means to check `.isDone`. If it succeeded, it creates an immediate success-deferred. If not, it attaches handlers. In the case of a failed/canceled future, it does a lot of unnecessary work before immediately invoking the handler: fn thread frame binding, handler object creation, and scheduling the handlers to run (potentially) on another thread. This is also kind of confusing because we're checking the wrong condition. To do: - [ ] Switch to `.isDone` from `.isSuccess` in `wrap-future` - [ ] If `.isDone`, choose between returning success- and error-deferreds based on the future's status. For consistency, we can reuse the `operation-complete` handler, but add a flag to tell it to use/ignore executors (so we can run it on the current thread) - [ ] Verify all tests still work, `test-classloader` in particular
perf
improve wrap future this was uncovered in the process of reviewing see my comments there for more in short wrap future checks issuccess on a future when it really means to check isdone if it succeeded it creates an immediate success deferred if not it attaches handlers in the case of a failed canceled future it does a lot of unnecessary work before immediately invoking the handler fn thread frame binding handler object creation and scheduling the handlers to run potentially on another thread this is also kind of confusing because we re checking the wrong condition to do switch to isdone from issuccess in wrap future if isdone choose between returning success and error deferreds based on the future s status for consistency we can reuse the operation complete handler but add a flag to tell it to use ignore executors so we can run it on the current thread verify all tests still work test classloader in particular
1
121,419
17,651,781,223
IssuesEvent
2021-08-20 14:07:12
snowdensb/spring-security-oauth
https://api.github.com/repos/snowdensb/spring-security-oauth
opened
CVE-2020-10673 (High) detected in jackson-databind-2.8.11.3.jar, jackson-databind-2.6.6.jar
security vulnerability
## CVE-2020-10673 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.11.3.jar</b>, <b>jackson-databind-2.6.6.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.8.11.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.22.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.6.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.6/jackson-databind-2.6.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.3.5.RELEASE.jar (Root Library) - :x: **jackson-databind-2.6.6.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/snowdensb/spring-security-oauth/commit/62c6485ab8fc644a76523cfb55746e549555797c">62c6485ab8fc644a76523cfb55746e549555797c</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus). <p>Publish Date: 2020-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p> <p>Release Date: 2020-03-18</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.3","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.22.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.6","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.3.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-10673","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-10673 (High) detected in jackson-databind-2.8.11.3.jar, jackson-databind-2.6.6.jar - ## CVE-2020-10673 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.11.3.jar</b>, <b>jackson-databind-2.6.6.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.8.11.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.22.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.6.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.6/jackson-databind-2.6.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.3.5.RELEASE.jar (Root Library) - :x: **jackson-databind-2.6.6.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/snowdensb/spring-security-oauth/commit/62c6485ab8fc644a76523cfb55746e549555797c">62c6485ab8fc644a76523cfb55746e549555797c</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus). <p>Publish Date: 2020-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p> <p>Release Date: 2020-03-18</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.3","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.22.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.6","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.3.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-10673","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_perf
cve high detected in jackson databind jar jackson databind jar cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus vulnerabilityurl
0
211,511
16,447,875,419
IssuesEvent
2021-05-20 22:16:20
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
Minor documentation bugs in quick-start section
area/documentation
Quick Start -> Java -> YCQL 4.6 section POM file section is missing Java source and target versions Fix whitespace in dependencies section Quick Start -> Java -> YSQL JDBC section Maven command needs additional target: package Quick Start -> Java -> YSQL Spring Data section Fix path for source
1.0
Minor documentation bugs in quick-start section - Quick Start -> Java -> YCQL 4.6 section POM file section is missing Java source and target versions Fix whitespace in dependencies section Quick Start -> Java -> YSQL JDBC section Maven command needs additional target: package Quick Start -> Java -> YSQL Spring Data section Fix path for source
non_perf
minor documentation bugs in quick start section quick start java ycql section pom file section is missing java source and target versions fix whitespace in dependencies section quick start java ysql jdbc section maven command needs additional target package quick start java ysql spring data section fix path for source
0
36,382
17,650,209,064
IssuesEvent
2021-08-20 12:11:31
socialsensingbot/frontend
https://api.github.com/repos/socialsensingbot/frontend
opened
Export Tweets to CSV is inefficent
internal performance
This can now be done entirely in SQL with a single call to the REST API. This should be done prior to 2.2 release.
True
Export Tweets to CSV is inefficent - This can now be done entirely in SQL with a single call to the REST API. This should be done prior to 2.2 release.
perf
export tweets to csv is inefficent this can now be done entirely in sql with a single call to the rest api this should be done prior to release
1
314,482
9,597,762,876
IssuesEvent
2019-05-09 22:22:55
CypherpunkArmory/UserLAnd
https://api.github.com/repos/CypherpunkArmory/UserLAnd
closed
Open Ubuntu Lxde but Debian starts
app request enhancement low priority
Samsung Note 9 installed Debian first with lxde and started using "start lxde &" At some point starting debian began opening lxde automatically (all tight vnc). (In the past Used to use gnu root with xsdl but i see it is not supported and tight vnc works fine and more easily. ) I then installed ubuntu file system with xfce4 and started it using command line to start xfce i forget the command something like startxfce4& or something. It worked but firefox was black. At some point the lxde and xfce showed up on userland home screen. I deleted ubuntu and reinstalled with lxde to see if firefox would run. Now i cannot get ubuntu to run lxde gui. When i click lxde or xfce from userland menu only debian opens. IS there a way to select the file system that will be opened with lxde or xfce userland home button? IS there a way to start lxde in ubuntu command line? Is installing lxde/xfce in terminal still necessary with the home screen buttons? O notice if i click xfce debian will run in xfce even tho i installed only lxde in debian command line in the beginning. Thank you for your time.
1.0
Open Ubuntu Lxde but Debian starts - Samsung Note 9 installed Debian first with lxde and started using "start lxde &" At some point starting debian began opening lxde automatically (all tight vnc). (In the past Used to use gnu root with xsdl but i see it is not supported and tight vnc works fine and more easily. ) I then installed ubuntu file system with xfce4 and started it using command line to start xfce i forget the command something like startxfce4& or something. It worked but firefox was black. At some point the lxde and xfce showed up on userland home screen. I deleted ubuntu and reinstalled with lxde to see if firefox would run. Now i cannot get ubuntu to run lxde gui. When i click lxde or xfce from userland menu only debian opens. IS there a way to select the file system that will be opened with lxde or xfce userland home button? IS there a way to start lxde in ubuntu command line? Is installing lxde/xfce in terminal still necessary with the home screen buttons? O notice if i click xfce debian will run in xfce even tho i installed only lxde in debian command line in the beginning. Thank you for your time.
non_perf
open ubuntu lxde but debian starts samsung note installed debian first with lxde and started using start lxde at some point starting debian began opening lxde automatically all tight vnc in the past used to use gnu root with xsdl but i see it is not supported and tight vnc works fine and more easily i then installed ubuntu file system with and started it using command line to start xfce i forget the command something like or something it worked but firefox was black at some point the lxde and xfce showed up on userland home screen i deleted ubuntu and reinstalled with lxde to see if firefox would run now i cannot get ubuntu to run lxde gui when i click lxde or xfce from userland menu only debian opens is there a way to select the file system that will be opened with lxde or xfce userland home button is there a way to start lxde in ubuntu command line is installing lxde xfce in terminal still necessary with the home screen buttons o notice if i click xfce debian will run in xfce even tho i installed only lxde in debian command line in the beginning thank you for your time
0
153,713
24,175,424,453
IssuesEvent
2022-09-23 00:44:11
enigmalea/enigmalea
https://api.github.com/repos/enigmalea/enigmalea
closed
Explore new color palettes
design
I'm not in love with the current color palettes for either light or dark mode. I need to explore more options.
1.0
Explore new color palettes - I'm not in love with the current color palettes for either light or dark mode. I need to explore more options.
non_perf
explore new color palettes i m not in love with the current color palettes for either light or dark mode i need to explore more options
0
23,353
4,932,397,177
IssuesEvent
2016-11-28 13:33:41
Jumpscale/jscockpit
https://api.github.com/repos/Jumpscale/jscockpit
opened
cockpit-doc: Instances Walkthrough
type_documentation type_feature
## GOAL: Explain what the **Instances** page is all about ## DESCRIPTION: Placeholder: https://github.com/Jumpscale/jscockpit/blob/master/docs/walkthrough/Instances/Instances.md
1.0
cockpit-doc: Instances Walkthrough - ## GOAL: Explain what the **Instances** page is all about ## DESCRIPTION: Placeholder: https://github.com/Jumpscale/jscockpit/blob/master/docs/walkthrough/Instances/Instances.md
non_perf
cockpit doc instances walkthrough goal explain what the instances page is all about description placeholder
0