Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3
values | title stringlengths 1 970 | labels stringlengths 4 625 | body stringlengths 3 247k | index stringclasses 9
values | text_combine stringlengths 96 247k | label stringclasses 2
values | text stringlengths 96 218k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
27,222 | 7,924,329,450 | IssuesEvent | 2018-07-05 16:34:27 | WH2-WEC/eom_marienburg | https://api.github.com/repos/WH2-WEC/eom_marienburg | opened | Create mutual protection script | Feature Building | If declared on by Empire factions, gain bonus with bret and vice versa.
| 1.0 | Create mutual protection script - If declared on by Empire factions, gain bonus with bret and vice versa.
| non_perf | create mutual protection script if declared on by empire factions gain bonus with bret and vice versa | 0 |
15,724 | 9,029,940,983 | IssuesEvent | 2019-02-08 01:09:51 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | AttributeError: module 'tensorflow.compat' has no attribute 'v1' | comp:apis type:bug/performance | I ran cnn_mnist.py on my machine but got "AttributeError: module 'tensorflow.compat' has no attribute 'v1'". can anyone solve this problem?
| True | AttributeError: module 'tensorflow.compat' has no attribute 'v1' - I ran cnn_mnist.py on my machine but got "AttributeError: module 'tensorflow.compat' has no attribute 'v1'". can anyone solve this problem?
| perf | attributeerror module tensorflow compat has no attribute i ran cnn mnist py on my machine but got attributeerror module tensorflow compat has no attribute can anyone solve this problem | 1 |
256,969 | 19,480,534,275 | IssuesEvent | 2021-12-25 06:45:04 | TuSimple/naive-ui | https://api.github.com/repos/TuSimple/naive-ui | closed | frFR locale update | documentation | <!-- generated by issue-helper DO NOT REMOVE __BUG__ -->
### TuSimple/naive-ui version (版本)
2.23.1
### Vue version (Vue 版本)
3.2.26
### Browser and its version (浏览器及其版本)
N/A
### System and its version (系统及其版本)
N/A
### Node version (Node 版本)
### Reappearance link (重现链接)
https://github.com/TuSimple/naive-ui/blob/main/src/locales/common/frFR.ts
### Reappearance steps (重现步骤)
Choose frFr locale
### Expected results (期望的结果)
Updated frFR locale translation
### Actual results (实际的结果)
Outdated frFR locale translation
### Remarks (补充说明)
Here is the updated frFr locale translation:
```js
import type { NLocale } from './enUS'
const frFR: NLocale = {
name: 'fr-FR',
global: {
undo: 'Défaire',
redo: 'Refaire',
confirm: 'Confirmer'
},
Popconfirm: {
positiveText: 'Confirmer',
negativeText: 'Annuler'
},
Cascader: {
placeholder: 'Sélectionner',
loading: 'Chargement',
loadingRequiredMessage: (label: string): string =>
`Charger tous les enfants de ${label} avant de le sélectionner`
},
Time: {
dateFormat: 'dd/MM/yyyy',
dateTimeFormat: 'dd/MM/yyyy HH:mm:ss'
},
DatePicker: {
yearFormat: 'yyyy',
monthFormat: 'MMM',
dayFormat: 'eeeeee',
yearTypeFormat: 'yyyy',
monthTypeFormat: 'MM/yyyy',
dateFormat: 'dd/MM/yyyy',
dateTimeFormat: 'dd/MM/yyyy HH:mm:ss',
quarterFormat: 'qqq yyyy',
clear: 'Effacer',
now: 'Maintenant',
confirm: 'Confirmer',
selectTime: "Sélectionner l'heure",
selectDate: 'Sélectionner la date',
datePlaceholder: 'Sélectionner la date',
datetimePlaceholder: "Sélectionner la date et l'heure",
monthPlaceholder: 'Sélectionner le mois',
yearPlaceholder: "Sélectionner l'année",
quarterPlaceholder: 'Sélectionner le trimestre',
startDatePlaceholder: 'Date de début',
endDatePlaceholder: 'Date de fin',
startDatetimePlaceholder: 'Date et heure de début',
endDatetimePlaceholder: 'Date et heure de fin',
monthBeforeYear: true,
firstDayOfWeek: 1,
today: "Aujourd'hui"
},
DataTable: {
checkTableAll: 'Sélectionner tout',
uncheckTableAll: 'Désélectionner tout',
confirm: 'Confirmer',
clear: 'Effacer'
},
Transfer: {
sourceTitle: 'Source',
targetTitle: 'Cible'
},
Empty: {
description: 'Aucune donnée'
},
Select: {
placeholder: 'Sélectionner'
},
TimePicker: {
placeholder: "Sélectionner l'heure",
positiveText: 'OK',
negativeText: 'Annuler',
now: 'Maintenant'
},
Pagination: {
goto: 'Aller à',
selectionSuffix: 'page'
},
DynamicTags: {
add: 'Ajouter'
},
Log: {
loading: 'Chargement'
},
Input: {
placeholder: 'Saisir'
},
InputNumber: {
placeholder: 'Saisir'
},
DynamicInput: {
create: 'Créer'
},
ThemeEditor: {
title: 'Editeur de thème',
clearAllVars: 'Effacer toutes les variables',
clearSearch: 'Effacer la recherche',
filterCompName: 'Filtrer par nom de composant',
filterVarName: 'Filtrer par nom de variable',
import: 'Importer',
export: 'Exporter',
restore: 'Réinitialiser'
}
}
export default frFR
```
<!-- generated by issue-helper DO NOT REMOVE __BUG__ --> | 1.0 | frFR locale update - <!-- generated by issue-helper DO NOT REMOVE __BUG__ -->
### TuSimple/naive-ui version (版本)
2.23.1
### Vue version (Vue 版本)
3.2.26
### Browser and its version (浏览器及其版本)
N/A
### System and its version (系统及其版本)
N/A
### Node version (Node 版本)
### Reappearance link (重现链接)
https://github.com/TuSimple/naive-ui/blob/main/src/locales/common/frFR.ts
### Reappearance steps (重现步骤)
Choose frFr locale
### Expected results (期望的结果)
Updated frFR locale translation
### Actual results (实际的结果)
Outdated frFR locale translation
### Remarks (补充说明)
Here is the updated frFr locale translation:
```js
import type { NLocale } from './enUS'
const frFR: NLocale = {
name: 'fr-FR',
global: {
undo: 'Défaire',
redo: 'Refaire',
confirm: 'Confirmer'
},
Popconfirm: {
positiveText: 'Confirmer',
negativeText: 'Annuler'
},
Cascader: {
placeholder: 'Sélectionner',
loading: 'Chargement',
loadingRequiredMessage: (label: string): string =>
`Charger tous les enfants de ${label} avant de le sélectionner`
},
Time: {
dateFormat: 'dd/MM/yyyy',
dateTimeFormat: 'dd/MM/yyyy HH:mm:ss'
},
DatePicker: {
yearFormat: 'yyyy',
monthFormat: 'MMM',
dayFormat: 'eeeeee',
yearTypeFormat: 'yyyy',
monthTypeFormat: 'MM/yyyy',
dateFormat: 'dd/MM/yyyy',
dateTimeFormat: 'dd/MM/yyyy HH:mm:ss',
quarterFormat: 'qqq yyyy',
clear: 'Effacer',
now: 'Maintenant',
confirm: 'Confirmer',
selectTime: "Sélectionner l'heure",
selectDate: 'Sélectionner la date',
datePlaceholder: 'Sélectionner la date',
datetimePlaceholder: "Sélectionner la date et l'heure",
monthPlaceholder: 'Sélectionner le mois',
yearPlaceholder: "Sélectionner l'année",
quarterPlaceholder: 'Sélectionner le trimestre',
startDatePlaceholder: 'Date de début',
endDatePlaceholder: 'Date de fin',
startDatetimePlaceholder: 'Date et heure de début',
endDatetimePlaceholder: 'Date et heure de fin',
monthBeforeYear: true,
firstDayOfWeek: 1,
today: "Aujourd'hui"
},
DataTable: {
checkTableAll: 'Sélectionner tout',
uncheckTableAll: 'Désélectionner tout',
confirm: 'Confirmer',
clear: 'Effacer'
},
Transfer: {
sourceTitle: 'Source',
targetTitle: 'Cible'
},
Empty: {
description: 'Aucune donnée'
},
Select: {
placeholder: 'Sélectionner'
},
TimePicker: {
placeholder: "Sélectionner l'heure",
positiveText: 'OK',
negativeText: 'Annuler',
now: 'Maintenant'
},
Pagination: {
goto: 'Aller à',
selectionSuffix: 'page'
},
DynamicTags: {
add: 'Ajouter'
},
Log: {
loading: 'Chargement'
},
Input: {
placeholder: 'Saisir'
},
InputNumber: {
placeholder: 'Saisir'
},
DynamicInput: {
create: 'Créer'
},
ThemeEditor: {
title: 'Editeur de thème',
clearAllVars: 'Effacer toutes les variables',
clearSearch: 'Effacer la recherche',
filterCompName: 'Filtrer par nom de composant',
filterVarName: 'Filtrer par nom de variable',
import: 'Importer',
export: 'Exporter',
restore: 'Réinitialiser'
}
}
export default frFR
```
<!-- generated by issue-helper DO NOT REMOVE __BUG__ --> | non_perf | frfr locale update tusimple naive ui version 版本 vue version vue 版本 browser and its version 浏览器及其版本 n a system and its version 系统及其版本 n a node version node 版本 reappearance link 重现链接 reappearance steps 重现步骤 choose frfr locale expected results 期望的结果 updated frfr locale translation actual results 实际的结果 outdated frfr locale translation remarks 补充说明 here is the updated frfr locale translation js import type nlocale from enus const frfr nlocale name fr fr global undo défaire redo refaire confirm confirmer popconfirm positivetext confirmer negativetext annuler cascader placeholder sélectionner loading chargement loadingrequiredmessage label string string charger tous les enfants de label avant de le sélectionner time dateformat dd mm yyyy datetimeformat dd mm yyyy hh mm ss datepicker yearformat yyyy monthformat mmm dayformat eeeeee yeartypeformat yyyy monthtypeformat mm yyyy dateformat dd mm yyyy datetimeformat dd mm yyyy hh mm ss quarterformat qqq yyyy clear effacer now maintenant confirm confirmer selecttime sélectionner l heure selectdate sélectionner la date dateplaceholder sélectionner la date datetimeplaceholder sélectionner la date et l heure monthplaceholder sélectionner le mois yearplaceholder sélectionner l année quarterplaceholder sélectionner le trimestre startdateplaceholder date de début enddateplaceholder date de fin startdatetimeplaceholder date et heure de début enddatetimeplaceholder date et heure de fin monthbeforeyear true firstdayofweek today aujourd hui datatable checktableall sélectionner tout unchecktableall désélectionner tout confirm confirmer clear effacer transfer sourcetitle source targettitle cible empty description aucune donnée select placeholder sélectionner timepicker placeholder sélectionner l heure positivetext ok negativetext annuler now maintenant pagination goto aller à selectionsuffix page dynamictags add ajouter log loading chargement input placeholder saisir inputnumber placeholder saisir dynamicinput create créer themeeditor title editeur de thème clearallvars effacer toutes les variables clearsearch effacer la recherche filtercompname filtrer par nom de composant filtervarname filtrer par nom de variable import importer export exporter restore réinitialiser export default frfr | 0 |
73,443 | 9,665,463,481 | IssuesEvent | 2019-05-21 08:34:55 | kubernetes/test-infra | https://api.github.com/repos/kubernetes/test-infra | closed | Running kubetest with precompiled kubectl | area/kubetest kind/bug kind/documentation lifecycle/stale sig/contributor-experience | Currently `kubetest` fails with
> It looks as if you don't have a compiled kubectl binary
even when kubectl is in `client/bin/kubectl` (and in $PATH, but that is secondary issue).
This can be fixed user-side by putting the precompiled binary into `_output/bin` or other paths defined by https://github.com/kubernetes/kubernetes/blob/d0439d417b0563d44c65b6e400e2070964dea7d1/cluster/clientbin.sh#L86
Is there a reason kube-test requires a locally compiled binary? With the new containerized kubernetes build, it is quite hard to build a custom container with kubetest as it would require docker in docker during image build (which is not currently possible).
This is a followup to #8698
/sig contributor-experience
/kind documentation
/area kubetest | 1.0 | Running kubetest with precompiled kubectl - Currently `kubetest` fails with
> It looks as if you don't have a compiled kubectl binary
even when kubectl is in `client/bin/kubectl` (and in $PATH, but that is secondary issue).
This can be fixed user-side by putting the precompiled binary into `_output/bin` or other paths defined by https://github.com/kubernetes/kubernetes/blob/d0439d417b0563d44c65b6e400e2070964dea7d1/cluster/clientbin.sh#L86
Is there a reason kube-test requires a locally compiled binary? With the new containerized kubernetes build, it is quite hard to build a custom container with kubetest as it would require docker in docker during image build (which is not currently possible).
This is a followup to #8698
/sig contributor-experience
/kind documentation
/area kubetest | non_perf | running kubetest with precompiled kubectl currently kubetest fails with it looks as if you don t have a compiled kubectl binary even when kubectl is in client bin kubectl and in path but that is secondary issue this can be fixed user side by putting the precompiled binary into output bin or other paths defined by is there a reason kube test requires a locally compiled binary with the new containerized kubernetes build it is quite hard to build a custom container with kubetest as it would require docker in docker during image build which is not currently possible this is a followup to sig contributor experience kind documentation area kubetest | 0 |
52,155 | 27,402,939,712 | IssuesEvent | 2023-03-01 03:03:05 | dermotduffy/frigate-hass-card | https://api.github.com/repos/dermotduffy/frigate-hass-card | closed | Desktop so much faster than mobile. Why? | performance | First of all; Thank you for this incredible work! What a card!
So my issues; When loading my camera page 6 cams + a snapshot gallery on chrome on my macbook, the page loads fully in about 2 seconds. When viewing the same page on my mobile (companion app on Android) the page takes minimal 5 seconds and in most cases up to 10.
Is there anything I can do to optimize?
HASS runs on a NUC <10% CPU. Frigate runs on another micro PC (J4125+GPU+Coral) at about 25% CPU. View provider in the card is the first option (homeassistan) but I have not done anything with webrtc yet.
Sorry if I posted this in the wrong category | True | Desktop so much faster than mobile. Why? - First of all; Thank you for this incredible work! What a card!
So my issues; When loading my camera page 6 cams + a snapshot gallery on chrome on my macbook, the page loads fully in about 2 seconds. When viewing the same page on my mobile (companion app on Android) the page takes minimal 5 seconds and in most cases up to 10.
Is there anything I can do to optimize?
HASS runs on a NUC <10% CPU. Frigate runs on another micro PC (J4125+GPU+Coral) at about 25% CPU. View provider in the card is the first option (homeassistan) but I have not done anything with webrtc yet.
Sorry if I posted this in the wrong category | perf | desktop so much faster than mobile why first of all thank you for this incredible work what a card so my issues when loading my camera page cams a snapshot gallery on chrome on my macbook the page loads fully in about seconds when viewing the same page on my mobile companion app on android the page takes minimal seconds and in most cases up to is there anything i can do to optimize hass runs on a nuc cpu frigate runs on another micro pc gpu coral at about cpu view provider in the card is the first option homeassistan but i have not done anything with webrtc yet sorry if i posted this in the wrong category | 1 |
52,067 | 27,360,916,180 | IssuesEvent | 2023-02-27 15:48:55 | jupyterlab/jupyterlab | https://api.github.com/repos/jupyterlab/jupyterlab | closed | JupyterLab requests /api/nbconvert twice during load | bug pkg:notebook tag:Performance good first issue | ## Description
JupyterLab during webpage load performs requests to /api/nbconvert twice.
## Reproduce
1. Open Chrome DevTools, switch to Network tab
2. Reload the page with JupyterLab
3. Find /api/nbconvert requests (you could filter by 'nbconvert')
4. See there are two simultanous requests
## Expected behavior
I probably would expect one request not two. The /api/nbconvert response time can be relatively slow (is quite slow in my full, non 'jlab-test' environment for some reason).
## Context
- Operating System and version: CentOS
- Browser and version: Edge 108.0.1462
- JupyterLab version: 3.5.2
<details><summary>Troubleshoot Output</summary>
<pre>
$ jupyter troubleshoot
$PATH:
/home/kungurov/.conda/envs/jlab-test/bin
/usr/frog/scicomp/pythonds/conda/condabin
/usr/frog/scicomp/pythonds/v1.2/bin
/usr/frog/scicomp/pythonds/latex/bin/x86_64-linux
/usr/frog/scicomp/pythonds/conda/condabin
/usr/local/bin
/usr/bin
/usr/local/sbin
/usr/sbin
/home/kungurov/bin
sys.path:
/home/kungurov/.conda/envs/jlab-test/bin
/home/kungurov/.conda/envs/jlab-test/lib/python311.zip
/home/kungurov/.conda/envs/jlab-test/lib/python3.11
/home/kungurov/.conda/envs/jlab-test/lib/python3.11/lib-dynload
/home/kungurov/.conda/envs/jlab-test/lib/python3.11/site-packages
sys.executable:
/home/kungurov/.conda/envs/jlab-test/bin/python
sys.version:
3.11.0 | packaged by conda-forge | (main, Jan 14 2023, 12:27:40) [GCC 11.3.0]
platform.platform():
Linux-3.10.0-1160.76.1.el7.x86_64-x86_64-with-glibc2.17
which -a jupyter:
/home/kungurov/.conda/envs/jlab-test/bin/jupyter
pip list:
Package Version
----------------------------- -----------
anyio 3.6.2
argon2-cffi 21.3.0
argon2-cffi-bindings 21.2.0
asttokens 2.2.1
attrs 22.2.0
Babel 2.11.0
backcall 0.2.0
backports.functools-lru-cache 1.6.4
beautifulsoup4 4.11.1
bleach 5.0.1
brotlipy 0.7.0
certifi 2022.12.7
cffi 1.15.1
charset-normalizer 2.1.1
comm 0.1.2
cryptography 39.0.0
debugpy 1.6.5
decorator 5.1.1
defusedxml 0.7.1
entrypoints 0.4
executing 1.2.0
fastjsonschema 2.16.2
flit_core 3.8.0
idna 3.4
importlib-metadata 6.0.0
importlib-resources 5.10.2
ipykernel 6.20.2
ipython 8.8.0
ipython-genutils 0.2.0
jedi 0.18.2
Jinja2 3.1.2
json5 0.9.5
jsonschema 4.17.3
jupyter_client 7.4.9
jupyter_core 5.1.3
jupyter-events 0.6.3
jupyter_server 2.1.0
jupyter_server_terminals 0.4.4
jupyterlab 3.5.2
jupyterlab-pygments 0.2.2
jupyterlab_server 2.19.0
MarkupSafe 2.1.2
matplotlib-inline 0.1.6
mistune 2.0.4
nbclassic 0.4.8
nbclient 0.7.2
nbconvert 7.2.8
nbformat 5.7.3
nest-asyncio 1.5.6
notebook 6.5.2
notebook_shim 0.2.2
packaging 23.0
pandocfilters 1.5.0
parso 0.8.3
pexpect 4.8.0
pickleshare 0.7.5
pip 22.3.1
pkgutil_resolve_name 1.3.10
platformdirs 2.6.2
prometheus-client 0.15.0
prompt-toolkit 3.0.36
psutil 5.9.4
ptyprocess 0.7.0
pure-eval 0.2.2
pycparser 2.21
Pygments 2.14.0
pyOpenSSL 23.0.0
pyrsistent 0.19.3
PySocks 1.7.1
python-dateutil 2.8.2
python-json-logger 2.0.4
pytz 2022.7.1
PyYAML 6.0
pyzmq 25.0.0
requests 2.28.2
rfc3339-validator 0.1.4
rfc3986-validator 0.1.1
Send2Trash 1.8.0
setuptools 66.1.1
six 1.16.0
sniffio 1.3.0
soupsieve 2.3.2.post1
stack-data 0.6.2
terminado 0.17.1
tinycss2 1.2.1
tomli 2.0.1
tornado 6.2
traitlets 5.8.1
typing_extensions 4.4.0
urllib3 1.26.14
wcwidth 0.2.6
webencodings 0.5.1
websocket-client 1.4.2
wheel 0.38.4
zipp 3.11.0
conda list:
# packages in environment at /home/kungurov/.conda/envs/jlab-test:
#
# Name Version Build Channel
_libgcc_mutex 0.1 conda_forge defaults
_openmp_mutex 4.5 2_gnu defaults
anyio 3.6.2 pyhd8ed1ab_0 defaults
argon2-cffi 21.3.0 pyhd8ed1ab_0 defaults
argon2-cffi-bindings 21.2.0 py311hd4cff14_3 defaults
asttokens 2.2.1 pyhd8ed1ab_0 defaults
attrs 22.2.0 pyh71513ae_0 defaults
babel 2.11.0 pyhd8ed1ab_0 defaults
backcall 0.2.0 pyh9f0ad1d_0 defaults
backports 1.0 pyhd8ed1ab_3 defaults
backports.functools_lru_cache 1.6.4 pyhd8ed1ab_0 defaults
beautifulsoup4 4.11.1 pyha770c72_0 defaults
bleach 5.0.1 pyhd8ed1ab_0 defaults
brotlipy 0.7.0 py311hd4cff14_1005 defaults
bzip2 1.0.8 h7f98852_4 defaults
ca-certificates 2022.12.7 ha878542_0 defaults
certifi 2022.12.7 pyhd8ed1ab_0 defaults
cffi 1.15.1 py311h409f033_3 defaults
charset-normalizer 2.1.1 pyhd8ed1ab_0 defaults
comm 0.1.2 pyhd8ed1ab_0 defaults
cryptography 39.0.0 py311h9b4c7bb_0 defaults
debugpy 1.6.5 py311hcafe171_0 defaults
decorator 5.1.1 pyhd8ed1ab_0 defaults
defusedxml 0.7.1 pyhd8ed1ab_0 defaults
entrypoints 0.4 pyhd8ed1ab_0 defaults
executing 1.2.0 pyhd8ed1ab_0 defaults
flit-core 3.8.0 pyhd8ed1ab_0 defaults
idna 3.4 pyhd8ed1ab_0 defaults
importlib-metadata 6.0.0 pyha770c72_0 defaults
importlib_resources 5.10.2 pyhd8ed1ab_0 defaults
ipykernel 6.20.2 pyh210e3f2_0 defaults
ipython 8.8.0 pyh41d4057_0 defaults
ipython_genutils 0.2.0 py_1 defaults
jedi 0.18.2 pyhd8ed1ab_0 defaults
jinja2 3.1.2 pyhd8ed1ab_1 defaults
json5 0.9.5 pyh9f0ad1d_0 defaults
jsonschema 4.17.3 pyhd8ed1ab_0 defaults
jupyter_client 7.4.9 pyhd8ed1ab_0 defaults
jupyter_core 5.1.3 py311h38be061_0 defaults
jupyter_events 0.6.3 pyhd8ed1ab_0 defaults
jupyter_server 2.1.0 pyhd8ed1ab_0 defaults
jupyter_server_terminals 0.4.4 pyhd8ed1ab_1 defaults
jupyterlab 3.5.2 pyhd8ed1ab_0 defaults
jupyterlab_pygments 0.2.2 pyhd8ed1ab_0 defaults
jupyterlab_server 2.19.0 pyhd8ed1ab_0 defaults
ld_impl_linux-64 2.39 hcc3a1bd_1 defaults
libffi 3.4.2 h7f98852_5 defaults
libgcc-ng 12.2.0 h65d4601_19 defaults
libgomp 12.2.0 h65d4601_19 defaults
libnsl 2.0.0 h7f98852_0 defaults
libsodium 1.0.18 h36c2ea0_1 defaults
libsqlite 3.40.0 h753d276_0 defaults
libstdcxx-ng 12.2.0 h46fd767_19 defaults
libuuid 2.32.1 h7f98852_1000 defaults
libzlib 1.2.13 h166bdaf_4 defaults
markupsafe 2.1.2 py311h2582759_0 defaults
matplotlib-inline 0.1.6 pyhd8ed1ab_0 defaults
mistune 2.0.4 pyhd8ed1ab_0 defaults
nbclassic 0.4.8 pyhd8ed1ab_0 defaults
nbclient 0.7.2 pyhd8ed1ab_0 defaults
nbconvert 7.2.8 pyhd8ed1ab_0 defaults
nbconvert-core 7.2.8 pyhd8ed1ab_0 defaults
nbconvert-pandoc 7.2.8 pyhd8ed1ab_0 defaults
nbformat 5.7.3 pyhd8ed1ab_0 defaults
ncurses 6.3 h27087fc_1 defaults
nest-asyncio 1.5.6 pyhd8ed1ab_0 defaults
notebook 6.5.2 pyha770c72_1 defaults
notebook-shim 0.2.2 pyhd8ed1ab_0 defaults
openssl 3.0.7 h0b41bf4_1 defaults
packaging 23.0 pyhd8ed1ab_0 defaults
pandoc 2.19.2 h32600fe_1 defaults
pandocfilters 1.5.0 pyhd8ed1ab_0 defaults
parso 0.8.3 pyhd8ed1ab_0 defaults
pexpect 4.8.0 pyh1a96a4e_2 defaults
pickleshare 0.7.5 py_1003 defaults
pip 22.3.1 pyhd8ed1ab_0 defaults
pkgutil-resolve-name 1.3.10 pyhd8ed1ab_0 defaults
platformdirs 2.6.2 pyhd8ed1ab_0 defaults
prometheus_client 0.15.0 pyhd8ed1ab_0 defaults
prompt-toolkit 3.0.36 pyha770c72_0 defaults
psutil 5.9.4 py311hd4cff14_0 defaults
ptyprocess 0.7.0 pyhd3deb0d_0 defaults
pure_eval 0.2.2 pyhd8ed1ab_0 defaults
pycparser 2.21 pyhd8ed1ab_0 defaults
pygments 2.14.0 pyhd8ed1ab_0 defaults
pyopenssl 23.0.0 pyhd8ed1ab_0 defaults
pyrsistent 0.19.3 py311h2582759_0 defaults
pysocks 1.7.1 pyha2e5f31_6 defaults
python 3.11.0 he550d4f_1_cpython defaults
python-dateutil 2.8.2 pyhd8ed1ab_0 defaults
python-fastjsonschema 2.16.2 pyhd8ed1ab_0 defaults
python-json-logger 2.0.4 pyhd8ed1ab_0 defaults
python_abi 3.11 3_cp311 defaults
pytz 2022.7.1 pyhd8ed1ab_0 defaults
pyyaml 6.0 py311hd4cff14_5 defaults
pyzmq 25.0.0 py311hd6ccaeb_0 defaults
readline 8.1.2 h0f457ee_0 defaults
requests 2.28.2 pyhd8ed1ab_0 defaults
rfc3339-validator 0.1.4 pyhd8ed1ab_0 defaults
rfc3986-validator 0.1.1 pyh9f0ad1d_0 defaults
send2trash 1.8.0 pyhd8ed1ab_0 defaults
setuptools 66.1.1 pyhd8ed1ab_0 defaults
six 1.16.0 pyh6c4a22f_0 defaults
sniffio 1.3.0 pyhd8ed1ab_0 defaults
soupsieve 2.3.2.post1 pyhd8ed1ab_0 defaults
stack_data 0.6.2 pyhd8ed1ab_0 defaults
terminado 0.17.1 pyh41d4057_0 defaults
tinycss2 1.2.1 pyhd8ed1ab_0 defaults
tk 8.6.12 h27826a3_0 defaults
tomli 2.0.1 pyhd8ed1ab_0 defaults
tornado 6.2 py311hd4cff14_1 defaults
traitlets 5.8.1 pyhd8ed1ab_0 defaults
typing-extensions 4.4.0 hd8ed1ab_0 defaults
typing_extensions 4.4.0 pyha770c72_0 defaults
tzdata 2022g h191b570_0 defaults
urllib3 1.26.14 pyhd8ed1ab_0 defaults
wcwidth 0.2.6 pyhd8ed1ab_0 defaults
webencodings 0.5.1 py_1 defaults
websocket-client 1.4.2 pyhd8ed1ab_0 defaults
wheel 0.38.4 pyhd8ed1ab_0 defaults
xz 5.2.6 h166bdaf_0 defaults
yaml 0.2.5 h7f98852_2 defaults
zeromq 4.3.4 h9c3ff4c_1 defaults
zipp 3.11.0 pyhd8ed1ab_0 defaults
conda env:
name: jlab-test
channels:
- defaults
dependencies:
- _libgcc_mutex=0.1=conda_forge
- _openmp_mutex=4.5=2_gnu
- anyio=3.6.2=pyhd8ed1ab_0
- argon2-cffi=21.3.0=pyhd8ed1ab_0
- argon2-cffi-bindings=21.2.0=py311hd4cff14_3
- asttokens=2.2.1=pyhd8ed1ab_0
- attrs=22.2.0=pyh71513ae_0
- babel=2.11.0=pyhd8ed1ab_0
- backcall=0.2.0=pyh9f0ad1d_0
- backports=1.0=pyhd8ed1ab_3
- backports.functools_lru_cache=1.6.4=pyhd8ed1ab_0
- beautifulsoup4=4.11.1=pyha770c72_0
- bleach=5.0.1=pyhd8ed1ab_0
- brotlipy=0.7.0=py311hd4cff14_1005
- bzip2=1.0.8=h7f98852_4
- ca-certificates=2022.12.7=ha878542_0
- certifi=2022.12.7=pyhd8ed1ab_0
- cffi=1.15.1=py311h409f033_3
- charset-normalizer=2.1.1=pyhd8ed1ab_0
- comm=0.1.2=pyhd8ed1ab_0
- cryptography=39.0.0=py311h9b4c7bb_0
- debugpy=1.6.5=py311hcafe171_0
- decorator=5.1.1=pyhd8ed1ab_0
- defusedxml=0.7.1=pyhd8ed1ab_0
- entrypoints=0.4=pyhd8ed1ab_0
- executing=1.2.0=pyhd8ed1ab_0
- flit-core=3.8.0=pyhd8ed1ab_0
- idna=3.4=pyhd8ed1ab_0
- importlib-metadata=6.0.0=pyha770c72_0
- importlib_resources=5.10.2=pyhd8ed1ab_0
- ipykernel=6.20.2=pyh210e3f2_0
- ipython=8.8.0=pyh41d4057_0
- ipython_genutils=0.2.0=py_1
- jedi=0.18.2=pyhd8ed1ab_0
- jinja2=3.1.2=pyhd8ed1ab_1
- json5=0.9.5=pyh9f0ad1d_0
- jsonschema=4.17.3=pyhd8ed1ab_0
- jupyter_client=7.4.9=pyhd8ed1ab_0
- jupyter_core=5.1.3=py311h38be061_0
- jupyter_events=0.6.3=pyhd8ed1ab_0
- jupyter_server=2.1.0=pyhd8ed1ab_0
- jupyter_server_terminals=0.4.4=pyhd8ed1ab_1
- jupyterlab=3.5.2=pyhd8ed1ab_0
- jupyterlab_pygments=0.2.2=pyhd8ed1ab_0
- jupyterlab_server=2.19.0=pyhd8ed1ab_0
- ld_impl_linux-64=2.39=hcc3a1bd_1
- libffi=3.4.2=h7f98852_5
- libgcc-ng=12.2.0=h65d4601_19
- libgomp=12.2.0=h65d4601_19
- libnsl=2.0.0=h7f98852_0
- libsodium=1.0.18=h36c2ea0_1
- libsqlite=3.40.0=h753d276_0
- libstdcxx-ng=12.2.0=h46fd767_19
- libuuid=2.32.1=h7f98852_1000
- libzlib=1.2.13=h166bdaf_4
- markupsafe=2.1.2=py311h2582759_0
- matplotlib-inline=0.1.6=pyhd8ed1ab_0
- mistune=2.0.4=pyhd8ed1ab_0
- nbclassic=0.4.8=pyhd8ed1ab_0
- nbclient=0.7.2=pyhd8ed1ab_0
- nbconvert=7.2.8=pyhd8ed1ab_0
- nbconvert-core=7.2.8=pyhd8ed1ab_0
- nbconvert-pandoc=7.2.8=pyhd8ed1ab_0
- nbformat=5.7.3=pyhd8ed1ab_0
- ncurses=6.3=h27087fc_1
- nest-asyncio=1.5.6=pyhd8ed1ab_0
- notebook=6.5.2=pyha770c72_1
- notebook-shim=0.2.2=pyhd8ed1ab_0
- openssl=3.0.7=h0b41bf4_1
- packaging=23.0=pyhd8ed1ab_0
- pandoc=2.19.2=h32600fe_1
- pandocfilters=1.5.0=pyhd8ed1ab_0
- parso=0.8.3=pyhd8ed1ab_0
- pexpect=4.8.0=pyh1a96a4e_2
- pickleshare=0.7.5=py_1003
- pip=22.3.1=pyhd8ed1ab_0
- pkgutil-resolve-name=1.3.10=pyhd8ed1ab_0
- platformdirs=2.6.2=pyhd8ed1ab_0
- prometheus_client=0.15.0=pyhd8ed1ab_0
- prompt-toolkit=3.0.36=pyha770c72_0
- psutil=5.9.4=py311hd4cff14_0
- ptyprocess=0.7.0=pyhd3deb0d_0
- pure_eval=0.2.2=pyhd8ed1ab_0
- pycparser=2.21=pyhd8ed1ab_0
- pygments=2.14.0=pyhd8ed1ab_0
- pyopenssl=23.0.0=pyhd8ed1ab_0
- pyrsistent=0.19.3=py311h2582759_0
- pysocks=1.7.1=pyha2e5f31_6
- python=3.11.0=he550d4f_1_cpython
- python-dateutil=2.8.2=pyhd8ed1ab_0
- python-fastjsonschema=2.16.2=pyhd8ed1ab_0
- python-json-logger=2.0.4=pyhd8ed1ab_0
- python_abi=3.11=3_cp311
- pytz=2022.7.1=pyhd8ed1ab_0
- pyyaml=6.0=py311hd4cff14_5
- pyzmq=25.0.0=py311hd6ccaeb_0
- readline=8.1.2=h0f457ee_0
- requests=2.28.2=pyhd8ed1ab_0
- rfc3339-validator=0.1.4=pyhd8ed1ab_0
- rfc3986-validator=0.1.1=pyh9f0ad1d_0
- send2trash=1.8.0=pyhd8ed1ab_0
- setuptools=66.1.1=pyhd8ed1ab_0
- six=1.16.0=pyh6c4a22f_0
- sniffio=1.3.0=pyhd8ed1ab_0
- soupsieve=2.3.2.post1=pyhd8ed1ab_0
- stack_data=0.6.2=pyhd8ed1ab_0
- terminado=0.17.1=pyh41d4057_0
- tinycss2=1.2.1=pyhd8ed1ab_0
- tk=8.6.12=h27826a3_0
- tomli=2.0.1=pyhd8ed1ab_0
- tornado=6.2=py311hd4cff14_1
- traitlets=5.8.1=pyhd8ed1ab_0
- typing-extensions=4.4.0=hd8ed1ab_0
- typing_extensions=4.4.0=pyha770c72_0
- tzdata=2022g=h191b570_0
- urllib3=1.26.14=pyhd8ed1ab_0
- wcwidth=0.2.6=pyhd8ed1ab_0
- webencodings=0.5.1=py_1
- websocket-client=1.4.2=pyhd8ed1ab_0
- wheel=0.38.4=pyhd8ed1ab_0
- xz=5.2.6=h166bdaf_0
- yaml=0.2.5=h7f98852_2
- zeromq=4.3.4=h9c3ff4c_1
- zipp=3.11.0=pyhd8ed1ab_0
prefix: /home/kungurov/.conda/envs/jlab-test
</pre>
</details>
I think it's because getExportFormats() called twice in the code (see below).
index.js, line 340:
```
void services.nbconvert.getExportFormats().then(response => {
if (response) {
const formatLabels = Private.getFormatLabels(translato);
// Convert export list to palette and menu items.
```
index.js, line 510:
```
void services.nbconvert.getExportFormats().then(response => {
if (response) {
/**
... ... ...
*/
const rawFormatExclude = [
``` | True | JupyterLab requests /api/nbconvert twice during load - ## Description
JupyterLab during webpage load performs requests to /api/nbconvert twice.
## Reproduce
1. Open Chrome DevTools, switch to Network tab
2. Reload the page with JupyterLab
3. Find /api/nbconvert requests (you could filter by 'nbconvert')
4. See there are two simultanous requests
## Expected behavior
I probably would expect one request not two. The /api/nbconvert response time can be relatively slow (is quite slow in my full, non 'jlab-test' environment for some reason).
## Context
- Operating System and version: CentOS
- Browser and version: Edge 108.0.1462
- JupyterLab version: 3.5.2
<details><summary>Troubleshoot Output</summary>
<pre>
$ jupyter troubleshoot
$PATH:
/home/kungurov/.conda/envs/jlab-test/bin
/usr/frog/scicomp/pythonds/conda/condabin
/usr/frog/scicomp/pythonds/v1.2/bin
/usr/frog/scicomp/pythonds/latex/bin/x86_64-linux
/usr/frog/scicomp/pythonds/conda/condabin
/usr/local/bin
/usr/bin
/usr/local/sbin
/usr/sbin
/home/kungurov/bin
sys.path:
/home/kungurov/.conda/envs/jlab-test/bin
/home/kungurov/.conda/envs/jlab-test/lib/python311.zip
/home/kungurov/.conda/envs/jlab-test/lib/python3.11
/home/kungurov/.conda/envs/jlab-test/lib/python3.11/lib-dynload
/home/kungurov/.conda/envs/jlab-test/lib/python3.11/site-packages
sys.executable:
/home/kungurov/.conda/envs/jlab-test/bin/python
sys.version:
3.11.0 | packaged by conda-forge | (main, Jan 14 2023, 12:27:40) [GCC 11.3.0]
platform.platform():
Linux-3.10.0-1160.76.1.el7.x86_64-x86_64-with-glibc2.17
which -a jupyter:
/home/kungurov/.conda/envs/jlab-test/bin/jupyter
pip list:
Package Version
----------------------------- -----------
anyio 3.6.2
argon2-cffi 21.3.0
argon2-cffi-bindings 21.2.0
asttokens 2.2.1
attrs 22.2.0
Babel 2.11.0
backcall 0.2.0
backports.functools-lru-cache 1.6.4
beautifulsoup4 4.11.1
bleach 5.0.1
brotlipy 0.7.0
certifi 2022.12.7
cffi 1.15.1
charset-normalizer 2.1.1
comm 0.1.2
cryptography 39.0.0
debugpy 1.6.5
decorator 5.1.1
defusedxml 0.7.1
entrypoints 0.4
executing 1.2.0
fastjsonschema 2.16.2
flit_core 3.8.0
idna 3.4
importlib-metadata 6.0.0
importlib-resources 5.10.2
ipykernel 6.20.2
ipython 8.8.0
ipython-genutils 0.2.0
jedi 0.18.2
Jinja2 3.1.2
json5 0.9.5
jsonschema 4.17.3
jupyter_client 7.4.9
jupyter_core 5.1.3
jupyter-events 0.6.3
jupyter_server 2.1.0
jupyter_server_terminals 0.4.4
jupyterlab 3.5.2
jupyterlab-pygments 0.2.2
jupyterlab_server 2.19.0
MarkupSafe 2.1.2
matplotlib-inline 0.1.6
mistune 2.0.4
nbclassic 0.4.8
nbclient 0.7.2
nbconvert 7.2.8
nbformat 5.7.3
nest-asyncio 1.5.6
notebook 6.5.2
notebook_shim 0.2.2
packaging 23.0
pandocfilters 1.5.0
parso 0.8.3
pexpect 4.8.0
pickleshare 0.7.5
pip 22.3.1
pkgutil_resolve_name 1.3.10
platformdirs 2.6.2
prometheus-client 0.15.0
prompt-toolkit 3.0.36
psutil 5.9.4
ptyprocess 0.7.0
pure-eval 0.2.2
pycparser 2.21
Pygments 2.14.0
pyOpenSSL 23.0.0
pyrsistent 0.19.3
PySocks 1.7.1
python-dateutil 2.8.2
python-json-logger 2.0.4
pytz 2022.7.1
PyYAML 6.0
pyzmq 25.0.0
requests 2.28.2
rfc3339-validator 0.1.4
rfc3986-validator 0.1.1
Send2Trash 1.8.0
setuptools 66.1.1
six 1.16.0
sniffio 1.3.0
soupsieve 2.3.2.post1
stack-data 0.6.2
terminado 0.17.1
tinycss2 1.2.1
tomli 2.0.1
tornado 6.2
traitlets 5.8.1
typing_extensions 4.4.0
urllib3 1.26.14
wcwidth 0.2.6
webencodings 0.5.1
websocket-client 1.4.2
wheel 0.38.4
zipp 3.11.0
conda list:
# packages in environment at /home/kungurov/.conda/envs/jlab-test:
#
# Name Version Build Channel
_libgcc_mutex 0.1 conda_forge defaults
_openmp_mutex 4.5 2_gnu defaults
anyio 3.6.2 pyhd8ed1ab_0 defaults
argon2-cffi 21.3.0 pyhd8ed1ab_0 defaults
argon2-cffi-bindings 21.2.0 py311hd4cff14_3 defaults
asttokens 2.2.1 pyhd8ed1ab_0 defaults
attrs 22.2.0 pyh71513ae_0 defaults
babel 2.11.0 pyhd8ed1ab_0 defaults
backcall 0.2.0 pyh9f0ad1d_0 defaults
backports 1.0 pyhd8ed1ab_3 defaults
backports.functools_lru_cache 1.6.4 pyhd8ed1ab_0 defaults
beautifulsoup4 4.11.1 pyha770c72_0 defaults
bleach 5.0.1 pyhd8ed1ab_0 defaults
brotlipy 0.7.0 py311hd4cff14_1005 defaults
bzip2 1.0.8 h7f98852_4 defaults
ca-certificates 2022.12.7 ha878542_0 defaults
certifi 2022.12.7 pyhd8ed1ab_0 defaults
cffi 1.15.1 py311h409f033_3 defaults
charset-normalizer 2.1.1 pyhd8ed1ab_0 defaults
comm 0.1.2 pyhd8ed1ab_0 defaults
cryptography 39.0.0 py311h9b4c7bb_0 defaults
debugpy 1.6.5 py311hcafe171_0 defaults
decorator 5.1.1 pyhd8ed1ab_0 defaults
defusedxml 0.7.1 pyhd8ed1ab_0 defaults
entrypoints 0.4 pyhd8ed1ab_0 defaults
executing 1.2.0 pyhd8ed1ab_0 defaults
flit-core 3.8.0 pyhd8ed1ab_0 defaults
idna 3.4 pyhd8ed1ab_0 defaults
importlib-metadata 6.0.0 pyha770c72_0 defaults
importlib_resources 5.10.2 pyhd8ed1ab_0 defaults
ipykernel 6.20.2 pyh210e3f2_0 defaults
ipython 8.8.0 pyh41d4057_0 defaults
ipython_genutils 0.2.0 py_1 defaults
jedi 0.18.2 pyhd8ed1ab_0 defaults
jinja2 3.1.2 pyhd8ed1ab_1 defaults
json5 0.9.5 pyh9f0ad1d_0 defaults
jsonschema 4.17.3 pyhd8ed1ab_0 defaults
jupyter_client 7.4.9 pyhd8ed1ab_0 defaults
jupyter_core 5.1.3 py311h38be061_0 defaults
jupyter_events 0.6.3 pyhd8ed1ab_0 defaults
jupyter_server 2.1.0 pyhd8ed1ab_0 defaults
jupyter_server_terminals 0.4.4 pyhd8ed1ab_1 defaults
jupyterlab 3.5.2 pyhd8ed1ab_0 defaults
jupyterlab_pygments 0.2.2 pyhd8ed1ab_0 defaults
jupyterlab_server 2.19.0 pyhd8ed1ab_0 defaults
ld_impl_linux-64 2.39 hcc3a1bd_1 defaults
libffi 3.4.2 h7f98852_5 defaults
libgcc-ng 12.2.0 h65d4601_19 defaults
libgomp 12.2.0 h65d4601_19 defaults
libnsl 2.0.0 h7f98852_0 defaults
libsodium 1.0.18 h36c2ea0_1 defaults
libsqlite 3.40.0 h753d276_0 defaults
libstdcxx-ng 12.2.0 h46fd767_19 defaults
libuuid 2.32.1 h7f98852_1000 defaults
libzlib 1.2.13 h166bdaf_4 defaults
markupsafe 2.1.2 py311h2582759_0 defaults
matplotlib-inline 0.1.6 pyhd8ed1ab_0 defaults
mistune 2.0.4 pyhd8ed1ab_0 defaults
nbclassic 0.4.8 pyhd8ed1ab_0 defaults
nbclient 0.7.2 pyhd8ed1ab_0 defaults
nbconvert 7.2.8 pyhd8ed1ab_0 defaults
nbconvert-core 7.2.8 pyhd8ed1ab_0 defaults
nbconvert-pandoc 7.2.8 pyhd8ed1ab_0 defaults
nbformat 5.7.3 pyhd8ed1ab_0 defaults
ncurses 6.3 h27087fc_1 defaults
nest-asyncio 1.5.6 pyhd8ed1ab_0 defaults
notebook 6.5.2 pyha770c72_1 defaults
notebook-shim 0.2.2 pyhd8ed1ab_0 defaults
openssl 3.0.7 h0b41bf4_1 defaults
packaging 23.0 pyhd8ed1ab_0 defaults
pandoc 2.19.2 h32600fe_1 defaults
pandocfilters 1.5.0 pyhd8ed1ab_0 defaults
parso 0.8.3 pyhd8ed1ab_0 defaults
pexpect 4.8.0 pyh1a96a4e_2 defaults
pickleshare 0.7.5 py_1003 defaults
pip 22.3.1 pyhd8ed1ab_0 defaults
pkgutil-resolve-name 1.3.10 pyhd8ed1ab_0 defaults
platformdirs 2.6.2 pyhd8ed1ab_0 defaults
prometheus_client 0.15.0 pyhd8ed1ab_0 defaults
prompt-toolkit 3.0.36 pyha770c72_0 defaults
psutil 5.9.4 py311hd4cff14_0 defaults
ptyprocess 0.7.0 pyhd3deb0d_0 defaults
pure_eval 0.2.2 pyhd8ed1ab_0 defaults
pycparser 2.21 pyhd8ed1ab_0 defaults
pygments 2.14.0 pyhd8ed1ab_0 defaults
pyopenssl 23.0.0 pyhd8ed1ab_0 defaults
pyrsistent 0.19.3 py311h2582759_0 defaults
pysocks 1.7.1 pyha2e5f31_6 defaults
python 3.11.0 he550d4f_1_cpython defaults
python-dateutil 2.8.2 pyhd8ed1ab_0 defaults
python-fastjsonschema 2.16.2 pyhd8ed1ab_0 defaults
python-json-logger 2.0.4 pyhd8ed1ab_0 defaults
python_abi 3.11 3_cp311 defaults
pytz 2022.7.1 pyhd8ed1ab_0 defaults
pyyaml 6.0 py311hd4cff14_5 defaults
pyzmq 25.0.0 py311hd6ccaeb_0 defaults
readline 8.1.2 h0f457ee_0 defaults
requests 2.28.2 pyhd8ed1ab_0 defaults
rfc3339-validator 0.1.4 pyhd8ed1ab_0 defaults
rfc3986-validator 0.1.1 pyh9f0ad1d_0 defaults
send2trash 1.8.0 pyhd8ed1ab_0 defaults
setuptools 66.1.1 pyhd8ed1ab_0 defaults
six 1.16.0 pyh6c4a22f_0 defaults
sniffio 1.3.0 pyhd8ed1ab_0 defaults
soupsieve 2.3.2.post1 pyhd8ed1ab_0 defaults
stack_data 0.6.2 pyhd8ed1ab_0 defaults
terminado 0.17.1 pyh41d4057_0 defaults
tinycss2 1.2.1 pyhd8ed1ab_0 defaults
tk 8.6.12 h27826a3_0 defaults
tomli 2.0.1 pyhd8ed1ab_0 defaults
tornado 6.2 py311hd4cff14_1 defaults
traitlets 5.8.1 pyhd8ed1ab_0 defaults
typing-extensions 4.4.0 hd8ed1ab_0 defaults
typing_extensions 4.4.0 pyha770c72_0 defaults
tzdata 2022g h191b570_0 defaults
urllib3 1.26.14 pyhd8ed1ab_0 defaults
wcwidth 0.2.6 pyhd8ed1ab_0 defaults
webencodings 0.5.1 py_1 defaults
websocket-client 1.4.2 pyhd8ed1ab_0 defaults
wheel 0.38.4 pyhd8ed1ab_0 defaults
xz 5.2.6 h166bdaf_0 defaults
yaml 0.2.5 h7f98852_2 defaults
zeromq 4.3.4 h9c3ff4c_1 defaults
zipp 3.11.0 pyhd8ed1ab_0 defaults
conda env:
name: jlab-test
channels:
- defaults
dependencies:
- _libgcc_mutex=0.1=conda_forge
- _openmp_mutex=4.5=2_gnu
- anyio=3.6.2=pyhd8ed1ab_0
- argon2-cffi=21.3.0=pyhd8ed1ab_0
- argon2-cffi-bindings=21.2.0=py311hd4cff14_3
- asttokens=2.2.1=pyhd8ed1ab_0
- attrs=22.2.0=pyh71513ae_0
- babel=2.11.0=pyhd8ed1ab_0
- backcall=0.2.0=pyh9f0ad1d_0
- backports=1.0=pyhd8ed1ab_3
- backports.functools_lru_cache=1.6.4=pyhd8ed1ab_0
- beautifulsoup4=4.11.1=pyha770c72_0
- bleach=5.0.1=pyhd8ed1ab_0
- brotlipy=0.7.0=py311hd4cff14_1005
- bzip2=1.0.8=h7f98852_4
- ca-certificates=2022.12.7=ha878542_0
- certifi=2022.12.7=pyhd8ed1ab_0
- cffi=1.15.1=py311h409f033_3
- charset-normalizer=2.1.1=pyhd8ed1ab_0
- comm=0.1.2=pyhd8ed1ab_0
- cryptography=39.0.0=py311h9b4c7bb_0
- debugpy=1.6.5=py311hcafe171_0
- decorator=5.1.1=pyhd8ed1ab_0
- defusedxml=0.7.1=pyhd8ed1ab_0
- entrypoints=0.4=pyhd8ed1ab_0
- executing=1.2.0=pyhd8ed1ab_0
- flit-core=3.8.0=pyhd8ed1ab_0
- idna=3.4=pyhd8ed1ab_0
- importlib-metadata=6.0.0=pyha770c72_0
- importlib_resources=5.10.2=pyhd8ed1ab_0
- ipykernel=6.20.2=pyh210e3f2_0
- ipython=8.8.0=pyh41d4057_0
- ipython_genutils=0.2.0=py_1
- jedi=0.18.2=pyhd8ed1ab_0
- jinja2=3.1.2=pyhd8ed1ab_1
- json5=0.9.5=pyh9f0ad1d_0
- jsonschema=4.17.3=pyhd8ed1ab_0
- jupyter_client=7.4.9=pyhd8ed1ab_0
- jupyter_core=5.1.3=py311h38be061_0
- jupyter_events=0.6.3=pyhd8ed1ab_0
- jupyter_server=2.1.0=pyhd8ed1ab_0
- jupyter_server_terminals=0.4.4=pyhd8ed1ab_1
- jupyterlab=3.5.2=pyhd8ed1ab_0
- jupyterlab_pygments=0.2.2=pyhd8ed1ab_0
- jupyterlab_server=2.19.0=pyhd8ed1ab_0
- ld_impl_linux-64=2.39=hcc3a1bd_1
- libffi=3.4.2=h7f98852_5
- libgcc-ng=12.2.0=h65d4601_19
- libgomp=12.2.0=h65d4601_19
- libnsl=2.0.0=h7f98852_0
- libsodium=1.0.18=h36c2ea0_1
- libsqlite=3.40.0=h753d276_0
- libstdcxx-ng=12.2.0=h46fd767_19
- libuuid=2.32.1=h7f98852_1000
- libzlib=1.2.13=h166bdaf_4
- markupsafe=2.1.2=py311h2582759_0
- matplotlib-inline=0.1.6=pyhd8ed1ab_0
- mistune=2.0.4=pyhd8ed1ab_0
- nbclassic=0.4.8=pyhd8ed1ab_0
- nbclient=0.7.2=pyhd8ed1ab_0
- nbconvert=7.2.8=pyhd8ed1ab_0
- nbconvert-core=7.2.8=pyhd8ed1ab_0
- nbconvert-pandoc=7.2.8=pyhd8ed1ab_0
- nbformat=5.7.3=pyhd8ed1ab_0
- ncurses=6.3=h27087fc_1
- nest-asyncio=1.5.6=pyhd8ed1ab_0
- notebook=6.5.2=pyha770c72_1
- notebook-shim=0.2.2=pyhd8ed1ab_0
- openssl=3.0.7=h0b41bf4_1
- packaging=23.0=pyhd8ed1ab_0
- pandoc=2.19.2=h32600fe_1
- pandocfilters=1.5.0=pyhd8ed1ab_0
- parso=0.8.3=pyhd8ed1ab_0
- pexpect=4.8.0=pyh1a96a4e_2
- pickleshare=0.7.5=py_1003
- pip=22.3.1=pyhd8ed1ab_0
- pkgutil-resolve-name=1.3.10=pyhd8ed1ab_0
- platformdirs=2.6.2=pyhd8ed1ab_0
- prometheus_client=0.15.0=pyhd8ed1ab_0
- prompt-toolkit=3.0.36=pyha770c72_0
- psutil=5.9.4=py311hd4cff14_0
- ptyprocess=0.7.0=pyhd3deb0d_0
- pure_eval=0.2.2=pyhd8ed1ab_0
- pycparser=2.21=pyhd8ed1ab_0
- pygments=2.14.0=pyhd8ed1ab_0
- pyopenssl=23.0.0=pyhd8ed1ab_0
- pyrsistent=0.19.3=py311h2582759_0
- pysocks=1.7.1=pyha2e5f31_6
- python=3.11.0=he550d4f_1_cpython
- python-dateutil=2.8.2=pyhd8ed1ab_0
- python-fastjsonschema=2.16.2=pyhd8ed1ab_0
- python-json-logger=2.0.4=pyhd8ed1ab_0
- python_abi=3.11=3_cp311
- pytz=2022.7.1=pyhd8ed1ab_0
- pyyaml=6.0=py311hd4cff14_5
- pyzmq=25.0.0=py311hd6ccaeb_0
- readline=8.1.2=h0f457ee_0
- requests=2.28.2=pyhd8ed1ab_0
- rfc3339-validator=0.1.4=pyhd8ed1ab_0
- rfc3986-validator=0.1.1=pyh9f0ad1d_0
- send2trash=1.8.0=pyhd8ed1ab_0
- setuptools=66.1.1=pyhd8ed1ab_0
- six=1.16.0=pyh6c4a22f_0
- sniffio=1.3.0=pyhd8ed1ab_0
- soupsieve=2.3.2.post1=pyhd8ed1ab_0
- stack_data=0.6.2=pyhd8ed1ab_0
- terminado=0.17.1=pyh41d4057_0
- tinycss2=1.2.1=pyhd8ed1ab_0
- tk=8.6.12=h27826a3_0
- tomli=2.0.1=pyhd8ed1ab_0
- tornado=6.2=py311hd4cff14_1
- traitlets=5.8.1=pyhd8ed1ab_0
- typing-extensions=4.4.0=hd8ed1ab_0
- typing_extensions=4.4.0=pyha770c72_0
- tzdata=2022g=h191b570_0
- urllib3=1.26.14=pyhd8ed1ab_0
- wcwidth=0.2.6=pyhd8ed1ab_0
- webencodings=0.5.1=py_1
- websocket-client=1.4.2=pyhd8ed1ab_0
- wheel=0.38.4=pyhd8ed1ab_0
- xz=5.2.6=h166bdaf_0
- yaml=0.2.5=h7f98852_2
- zeromq=4.3.4=h9c3ff4c_1
- zipp=3.11.0=pyhd8ed1ab_0
prefix: /home/kungurov/.conda/envs/jlab-test
</pre>
</details>
I think it's because getExportFormats() called twice in the code (see below).
index.js, line 340:
```
void services.nbconvert.getExportFormats().then(response => {
if (response) {
const formatLabels = Private.getFormatLabels(translato);
// Convert export list to palette and menu items.
```
index.js, line 510:
```
void services.nbconvert.getExportFormats().then(response => {
if (response) {
/**
... ... ...
*/
const rawFormatExclude = [
``` | perf | jupyterlab requests api nbconvert twice during load description jupyterlab during webpage load performs requests to api nbconvert twice reproduce open chrome devtools switch to network tab reload the page with jupyterlab find api nbconvert requests you could filter by nbconvert see there are two simultanous requests expected behavior i probably would expect one request not two the api nbconvert response time can be relatively slow is quite slow in my full non jlab test environment for some reason context operating system and version centos browser and version edge jupyterlab version troubleshoot output jupyter troubleshoot path home kungurov conda envs jlab test bin usr frog scicomp pythonds conda condabin usr frog scicomp pythonds bin usr frog scicomp pythonds latex bin linux usr frog scicomp pythonds conda condabin usr local bin usr bin usr local sbin usr sbin home kungurov bin sys path home kungurov conda envs jlab test bin home kungurov conda envs jlab test lib zip home kungurov conda envs jlab test lib home kungurov conda envs jlab test lib lib dynload home kungurov conda envs jlab test lib site packages sys executable home kungurov conda envs jlab test bin python sys version packaged by conda forge main jan platform platform linux with which a jupyter home kungurov conda envs jlab test bin jupyter pip list package version anyio cffi cffi bindings asttokens attrs babel backcall backports functools lru cache bleach brotlipy certifi cffi charset normalizer comm cryptography debugpy decorator defusedxml entrypoints executing fastjsonschema flit core idna importlib metadata importlib resources ipykernel ipython ipython genutils jedi jsonschema jupyter client jupyter core jupyter events jupyter server jupyter server terminals jupyterlab jupyterlab pygments jupyterlab server markupsafe matplotlib inline mistune nbclassic nbclient nbconvert nbformat nest asyncio notebook notebook shim packaging pandocfilters parso pexpect pickleshare pip pkgutil resolve name platformdirs prometheus client prompt toolkit psutil ptyprocess pure eval pycparser pygments pyopenssl pyrsistent pysocks python dateutil python json logger pytz pyyaml pyzmq requests validator validator setuptools six sniffio soupsieve stack data terminado tomli tornado traitlets typing extensions wcwidth webencodings websocket client wheel zipp conda list packages in environment at home kungurov conda envs jlab test name version build channel libgcc mutex conda forge defaults openmp mutex gnu defaults anyio defaults cffi defaults cffi bindings defaults asttokens defaults attrs defaults babel defaults backcall defaults backports defaults backports functools lru cache defaults defaults bleach defaults brotlipy defaults defaults ca certificates defaults certifi defaults cffi defaults charset normalizer defaults comm defaults cryptography defaults debugpy defaults decorator defaults defusedxml defaults entrypoints defaults executing defaults flit core defaults idna defaults importlib metadata defaults importlib resources defaults ipykernel defaults ipython defaults ipython genutils py defaults jedi defaults defaults defaults jsonschema defaults jupyter client defaults jupyter core defaults jupyter events defaults jupyter server defaults jupyter server terminals defaults jupyterlab defaults jupyterlab pygments defaults jupyterlab server defaults ld impl linux defaults libffi defaults libgcc ng defaults libgomp defaults libnsl defaults libsodium defaults libsqlite defaults libstdcxx ng defaults libuuid defaults libzlib defaults markupsafe defaults matplotlib inline defaults mistune defaults nbclassic defaults nbclient defaults nbconvert defaults nbconvert core defaults nbconvert pandoc defaults nbformat defaults ncurses defaults nest asyncio defaults notebook defaults notebook shim defaults openssl defaults packaging defaults pandoc defaults pandocfilters defaults parso defaults pexpect defaults pickleshare py defaults pip defaults pkgutil resolve name defaults platformdirs defaults prometheus client defaults prompt toolkit defaults psutil defaults ptyprocess defaults pure eval defaults pycparser defaults pygments defaults pyopenssl defaults pyrsistent defaults pysocks defaults python cpython defaults python dateutil defaults python fastjsonschema defaults python json logger defaults python abi defaults pytz defaults pyyaml defaults pyzmq defaults readline defaults requests defaults validator defaults validator defaults defaults setuptools defaults six defaults sniffio defaults soupsieve defaults stack data defaults terminado defaults defaults tk defaults tomli defaults tornado defaults traitlets defaults typing extensions defaults typing extensions defaults tzdata defaults defaults wcwidth defaults webencodings py defaults websocket client defaults wheel defaults xz defaults yaml defaults zeromq defaults zipp defaults conda env name jlab test channels defaults dependencies libgcc mutex conda forge openmp mutex gnu anyio cffi cffi bindings asttokens attrs babel backcall backports backports functools lru cache bleach brotlipy ca certificates certifi cffi charset normalizer comm cryptography debugpy decorator defusedxml entrypoints executing flit core idna importlib metadata importlib resources ipykernel ipython ipython genutils py jedi jsonschema jupyter client jupyter core jupyter events jupyter server jupyter server terminals jupyterlab jupyterlab pygments jupyterlab server ld impl linux libffi libgcc ng libgomp libnsl libsodium libsqlite libstdcxx ng libuuid libzlib markupsafe matplotlib inline mistune nbclassic nbclient nbconvert nbconvert core nbconvert pandoc nbformat ncurses nest asyncio notebook notebook shim openssl packaging pandoc pandocfilters parso pexpect pickleshare py pip pkgutil resolve name platformdirs prometheus client prompt toolkit psutil ptyprocess pure eval pycparser pygments pyopenssl pyrsistent pysocks python cpython python dateutil python fastjsonschema python json logger python abi pytz pyyaml pyzmq readline requests validator validator setuptools six sniffio soupsieve stack data terminado tk tomli tornado traitlets typing extensions typing extensions tzdata wcwidth webencodings py websocket client wheel xz yaml zeromq zipp prefix home kungurov conda envs jlab test i think it s because getexportformats called twice in the code see below index js line void services nbconvert getexportformats then response if response const formatlabels private getformatlabels translato convert export list to palette and menu items index js line void services nbconvert getexportformats then response if response const rawformatexclude | 1 |
121,899 | 16,050,131,277 | IssuesEvent | 2021-04-22 18:04:15 | t10d/kingdom-python-server | https://api.github.com/repos/t10d/kingdom-python-server | opened | Adapt authorization module to work with current entrypoints | crucial design change refactor | Analogous to #20, but for entrypoint integrations.
## Goal
**One liner**: Integrate authorization module services to current implementations of middleware & directives.
1. Adapt authentication middleware and context propagation
2. Adapt authorization middleware and proper scope results handling on query resolvers
3. Proper error handling
| 1.0 | Adapt authorization module to work with current entrypoints - Analogous to #20, but for entrypoint integrations.
## Goal
**One liner**: Integrate authorization module services to current implementations of middleware & directives.
1. Adapt authentication middleware and context propagation
2. Adapt authorization middleware and proper scope results handling on query resolvers
3. Proper error handling
| non_perf | adapt authorization module to work with current entrypoints analogous to but for entrypoint integrations goal one liner integrate authorization module services to current implementations of middleware directives adapt authentication middleware and context propagation adapt authorization middleware and proper scope results handling on query resolvers proper error handling | 0 |
308,979 | 9,459,738,633 | IssuesEvent | 2019-04-17 09:14:30 | WallarooLabs/wallaroo | https://api.github.com/repos/WallarooLabs/wallaroo | closed | Add ability to migrate a source to a different worker. | enhancement priority: low | Currently a source cannot be moved from a worker. In order to allow shrink-to-fit on any worker, we need to be able to migrate a source from one worker to the other.
We may come up with different solutions for push and pull sources, because whatever connects to a push source will need to be informed about having to reconnect to a different place, or some other trickery will have to be done to make sure we move the connection to the right place. For pull sources, this seems much easier. | 1.0 | Add ability to migrate a source to a different worker. - Currently a source cannot be moved from a worker. In order to allow shrink-to-fit on any worker, we need to be able to migrate a source from one worker to the other.
We may come up with different solutions for push and pull sources, because whatever connects to a push source will need to be informed about having to reconnect to a different place, or some other trickery will have to be done to make sure we move the connection to the right place. For pull sources, this seems much easier. | non_perf | add ability to migrate a source to a different worker currently a source cannot be moved from a worker in order to allow shrink to fit on any worker we need to be able to migrate a source from one worker to the other we may come up with different solutions for push and pull sources because whatever connects to a push source will need to be informed about having to reconnect to a different place or some other trickery will have to be done to make sure we move the connection to the right place for pull sources this seems much easier | 0 |
407,678 | 27,624,984,463 | IssuesEvent | 2023-03-10 05:37:41 | binance/binance-connector-dotnet | https://api.github.com/repos/binance/binance-connector-dotnet | closed | Documentation: BSwap.Swap not working from USDT to ETH (binance-connector-dotnet) | documentation Reviewing | ### Description
Hello, i am not sure what i am doing wrong, i am trying to make a swap from USDT to ETH but it gives me an "Exception of type 'Binance.Common.BinanceClientException' error. USDT to BUSD works fine. Any help will be appreciated, thank you.
CODE:
var bSwap = new BSwap(binanceUrl, binanceApiKey, binanceApiSecret);
var result = await bSwap.Swap("USDT", "ETH", 5.0m);
print(result); | 1.0 | Documentation: BSwap.Swap not working from USDT to ETH (binance-connector-dotnet) - ### Description
Hello, i am not sure what i am doing wrong, i am trying to make a swap from USDT to ETH but it gives me an "Exception of type 'Binance.Common.BinanceClientException' error. USDT to BUSD works fine. Any help will be appreciated, thank you.
CODE:
var bSwap = new BSwap(binanceUrl, binanceApiKey, binanceApiSecret);
var result = await bSwap.Swap("USDT", "ETH", 5.0m);
print(result); | non_perf | documentation bswap swap not working from usdt to eth binance connector dotnet description hello i am not sure what i am doing wrong i am trying to make a swap from usdt to eth but it gives me an exception of type binance common binanceclientexception error usdt to busd works fine any help will be appreciated thank you code var bswap new bswap binanceurl binanceapikey binanceapisecret var result await bswap swap usdt eth print result | 0 |
575,524 | 17,033,844,413 | IssuesEvent | 2021-07-05 02:25:20 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | opened | solvers: Ipopt segfaults when a problem has a quadratic objects + a Lorentz cone constraint | component: mathematical program priority: medium team: manipulation type: bug | This runs fine on Mosek, but causes IPOPT to segfault:
```
from pydrake.all import MathematicalProgram, Solve
prog = MathematicalProgram()
x = prog.NewContinuousVariables(2)
prog.AddLorentzConeConstraint(x)
prog.AddQuadraticCost(x.dot(x))
result = Solve(prog)
print(result.is_success())
print(result.get_solver_id().name())
``` | 1.0 | solvers: Ipopt segfaults when a problem has a quadratic objects + a Lorentz cone constraint - This runs fine on Mosek, but causes IPOPT to segfault:
```
from pydrake.all import MathematicalProgram, Solve
prog = MathematicalProgram()
x = prog.NewContinuousVariables(2)
prog.AddLorentzConeConstraint(x)
prog.AddQuadraticCost(x.dot(x))
result = Solve(prog)
print(result.is_success())
print(result.get_solver_id().name())
``` | non_perf | solvers ipopt segfaults when a problem has a quadratic objects a lorentz cone constraint this runs fine on mosek but causes ipopt to segfault from pydrake all import mathematicalprogram solve prog mathematicalprogram x prog newcontinuousvariables prog addlorentzconeconstraint x prog addquadraticcost x dot x result solve prog print result is success print result get solver id name | 0 |
367,050 | 10,833,483,517 | IssuesEvent | 2019-11-11 13:02:32 | CESARBR/knot-gateway-webui | https://api.github.com/repos/CESARBR/knot-gateway-webui | opened | After removing device, the WebUI shows an error | bug priority: low | **Expected Behavior**:
No error should be shown after removing a device.
**Current Behavior**:
The message `An unexpected error occurred` is showed after removing a device.
**Context (Environment):**
The WebUI should be configured with a user.
**Steps to reproduce:**
- Add a device
- Remove the device and wait for a few seconds.
- The message `An unexpected error occurred` is showed.
**Possible solution:**
Avoid fetching data from devices removed. | 1.0 | After removing device, the WebUI shows an error - **Expected Behavior**:
No error should be shown after removing a device.
**Current Behavior**:
The message `An unexpected error occurred` is showed after removing a device.
**Context (Environment):**
The WebUI should be configured with a user.
**Steps to reproduce:**
- Add a device
- Remove the device and wait for a few seconds.
- The message `An unexpected error occurred` is showed.
**Possible solution:**
Avoid fetching data from devices removed. | non_perf | after removing device the webui shows an error expected behavior no error should be shown after removing a device current behavior the message an unexpected error occurred is showed after removing a device context environment the webui should be configured with a user steps to reproduce add a device remove the device and wait for a few seconds the message an unexpected error occurred is showed possible solution avoid fetching data from devices removed | 0 |
21,115 | 10,574,280,617 | IssuesEvent | 2019-10-07 13:44:02 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | pgwire: create an audit log for user logins | A-security A-sql-pgwire C-bug C-enhancement | Requested on forum: https://forum.cockroachlabs.com/t/how-to-audit-login-attempts/3182/2
We need/want a log of success/failed login attempts and also user logouts (or conn dropped).
This is currently unavailable via the existing logging and audit facilities. | True | pgwire: create an audit log for user logins - Requested on forum: https://forum.cockroachlabs.com/t/how-to-audit-login-attempts/3182/2
We need/want a log of success/failed login attempts and also user logouts (or conn dropped).
This is currently unavailable via the existing logging and audit facilities. | non_perf | pgwire create an audit log for user logins requested on forum we need want a log of success failed login attempts and also user logouts or conn dropped this is currently unavailable via the existing logging and audit facilities | 0 |
19,154 | 10,326,083,407 | IssuesEvent | 2019-09-01 23:13:32 | modin-project/modin | https://api.github.com/repos/modin-project/modin | closed | UserWarning: `Series.__getstate__` defaulting to pandas implementation. | Performance 🚀 bug :bug: | <!--
General questions should be asked on the mailing list modin-dev@googlegroups.com.
Before submitting an issue, please fill out the following form.
-->
### System information
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**:Ubuntu 18.04.1 LTS
- **Modin installed from (source or binary)**:binary
- **Modin version**:0.5.4
- **Python version**:3.6.8
- **Exact command to reproduce**:
The ‘merged’ is a DataFrame that 'modin.pandas.concat(list)' function generates. The column of ‘ticker’ is a series that the type of element is an integer.
```
merged['ticker'] = merged['ticker'].astype(str)
```
<!--
You can obtain the Modin version with
python -c "import modin; print(modin.__version__)"
-->
### Describe the problem
<!-- Describe the problem clearly here. -->
Then raise a warning:
```
UserWarning: `Series.__getstate__` defaulting to pandas implementation.
```
The program is stuck and don't go on.
### Source code / logs
<!-- Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. Try to provide a reproducible test case that is the bare minimum necessary to generate the problem. -->
| True | UserWarning: `Series.__getstate__` defaulting to pandas implementation. - <!--
General questions should be asked on the mailing list modin-dev@googlegroups.com.
Before submitting an issue, please fill out the following form.
-->
### System information
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**:Ubuntu 18.04.1 LTS
- **Modin installed from (source or binary)**:binary
- **Modin version**:0.5.4
- **Python version**:3.6.8
- **Exact command to reproduce**:
The ‘merged’ is a DataFrame that 'modin.pandas.concat(list)' function generates. The column of ‘ticker’ is a series that the type of element is an integer.
```
merged['ticker'] = merged['ticker'].astype(str)
```
<!--
You can obtain the Modin version with
python -c "import modin; print(modin.__version__)"
-->
### Describe the problem
<!-- Describe the problem clearly here. -->
Then raise a warning:
```
UserWarning: `Series.__getstate__` defaulting to pandas implementation.
```
The program is stuck and don't go on.
### Source code / logs
<!-- Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. Try to provide a reproducible test case that is the bare minimum necessary to generate the problem. -->
| perf | userwarning series getstate defaulting to pandas implementation general questions should be asked on the mailing list modin dev googlegroups com before submitting an issue please fill out the following form system information os platform and distribution e g linux ubuntu ubuntu lts modin installed from source or binary binary modin version python version exact command to reproduce the ‘merged’ is a dataframe that modin pandas concat list function generates the column of ‘ticker’ is a series that the type of element is an integer merged merged astype str you can obtain the modin version with python c import modin print modin version describe the problem then raise a warning userwarning series getstate defaulting to pandas implementation the program is stuck and don t go on source code logs | 1 |
3,728 | 2,906,269,741 | IssuesEvent | 2015-06-19 08:55:08 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | User Registration - New User Account Activation=Admin. Admin does not need to confirm activation | No Code Attached Yet | #### Steps to reproduce the issue
Create a new Joomla site, with one Super User (automatic). Version 3.4.1
Allow users to register:
Users > User Manager > Options
Make the following changes:
Allow User Registration = Yes
Send Password = No
New User Account Activation = Admin
On the front end, Create an Account (using Create an Account link)
Enter in your name, password and a valid email address. Submit the registration.
Verify on the backend (under user manager) that user has been added but that "Enabled" and "Activated" are OFF/Disable.
Check your email for the registration email. Click the link in the email to verify the account. You should be directed to the Joomla site with a message that your email address has been verified and once an administrator approves it you can login.
Verify in the backend that the new user shows "Activated" but not "Enabled".
#### Expected result
You should see that the user is "Activated" but not "Enabled". The user should not be able to login.
#### Actual result
The user is both "Activated and Enabled". No admin intervention was required.
#### System information (as much as possible)
Joomla Site hosted on Cloudaccess.net.
Version 3.4.1
#### Additional comments
Recap of existing forum posts for this issue:
Perfect description of the issue (but, can happen if you only have one admin)
http://forum.joomla.org/viewtopic.php?f=708&t=879213&p=3281053&hilit=user+activation#p3281053
and another one (v 3.4)
http://forum.joomla.org/viewtopic.php?f=719&t=880887
http://forum.joomla.org/viewtopic.php?f=708&t=856769&hilit=user+activation
not exactly the issue
https://github.com/joomla/joomla-cms/issues/4376
And a post I responded to, but unsure that it is being looked into further (and this does not require that two admin accounts be present for it to be an issue)
http://forum.joomla.org/viewtopic.php?f=719&t=880887
| 1.0 | User Registration - New User Account Activation=Admin. Admin does not need to confirm activation - #### Steps to reproduce the issue
Create a new Joomla site, with one Super User (automatic). Version 3.4.1
Allow users to register:
Users > User Manager > Options
Make the following changes:
Allow User Registration = Yes
Send Password = No
New User Account Activation = Admin
On the front end, Create an Account (using Create an Account link)
Enter in your name, password and a valid email address. Submit the registration.
Verify on the backend (under user manager) that user has been added but that "Enabled" and "Activated" are OFF/Disable.
Check your email for the registration email. Click the link in the email to verify the account. You should be directed to the Joomla site with a message that your email address has been verified and once an administrator approves it you can login.
Verify in the backend that the new user shows "Activated" but not "Enabled".
#### Expected result
You should see that the user is "Activated" but not "Enabled". The user should not be able to login.
#### Actual result
The user is both "Activated and Enabled". No admin intervention was required.
#### System information (as much as possible)
Joomla Site hosted on Cloudaccess.net.
Version 3.4.1
#### Additional comments
Recap of existing forum posts for this issue:
Perfect description of the issue (but, can happen if you only have one admin)
http://forum.joomla.org/viewtopic.php?f=708&t=879213&p=3281053&hilit=user+activation#p3281053
and another one (v 3.4)
http://forum.joomla.org/viewtopic.php?f=719&t=880887
http://forum.joomla.org/viewtopic.php?f=708&t=856769&hilit=user+activation
not exactly the issue
https://github.com/joomla/joomla-cms/issues/4376
And a post I responded to, but unsure that it is being looked into further (and this does not require that two admin accounts be present for it to be an issue)
http://forum.joomla.org/viewtopic.php?f=719&t=880887
| non_perf | user registration new user account activation admin admin does not need to confirm activation steps to reproduce the issue create a new joomla site with one super user automatic version allow users to register users user manager options make the following changes allow user registration yes send password no new user account activation admin on the front end create an account using create an account link enter in your name password and a valid email address submit the registration verify on the backend under user manager that user has been added but that enabled and activated are off disable check your email for the registration email click the link in the email to verify the account you should be directed to the joomla site with a message that your email address has been verified and once an administrator approves it you can login verify in the backend that the new user shows activated but not enabled expected result you should see that the user is activated but not enabled the user should not be able to login actual result the user is both activated and enabled no admin intervention was required system information as much as possible joomla site hosted on cloudaccess net version additional comments recap of existing forum posts for this issue perfect description of the issue but can happen if you only have one admin and another one v not exactly the issue and a post i responded to but unsure that it is being looked into further and this does not require that two admin accounts be present for it to be an issue | 0 |
143,025 | 21,932,070,775 | IssuesEvent | 2022-05-23 10:37:20 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | closed | Explore adding site width and padding to global styles | Needs Design Feedback Global Styles | This iteration is inspired by feedback from those exploring patterns and also the visual work explored in full site editing by @shaunandrews. Global styles have a suggested feature enabling to increase or decrease the typography scale. This could easily lead to lines with a few words or too many for comfortable reading. A great additional option could be to allow a site width setting and padding around that site.
I am aware this won't be easy for all themes, so just proposing it here to work around some issues that might come up. I also know that this is being explored in templates, so it might be decided it belongs there instead.
## Option A: textual
This version uses text to indicate the width and padding. It's also worth noting all illustrations show 'layout', which I am not set on and could change. Similar, these use the sidebar as that is what we are using for v1 global styles, this could be comfortable in the toolbar as other iterations happen.
<img width="1257" alt="global-i1" src="https://user-images.githubusercontent.com/253067/76330175-7056f000-62e5-11ea-9441-1365d790855e.png">
It is worth noting that with the iterations to the interface, we don't have a small input with increasing/decreasing arrows. I explored a little what this could look like with a click in the state showing arrows:
<img width="242" alt="Frame 1" src="https://user-images.githubusercontent.com/253067/76330376-ad22e700-62e5-11ea-8292-dddd6464c586.png">
@jasmussen looping you in to get input for interface iterations.
Here it is without explicit arrows:
<img width="1257" alt="global-i2" src="https://user-images.githubusercontent.com/253067/76330409-b6ac4f00-62e5-11ea-96b8-9f75e247af9a.png">
## Option B : visual
This took a large dose of full site editing inspiration, I brought in some visuals used there to explore what that could look like. I think the iconography might need exploring but for me, the visual has a certain something extra I really like.
<img width="1257" alt="global-i3" src="https://user-images.githubusercontent.com/253067/76329911-122a0d00-62e5-11ea-8154-d9402f7026cb.png">
## Feedback
I would love general feedback as this is an early idea and bringing in concepts from elsewhere. That said, it would be good to get specific responses around:
- Does this belong in global styles along with templates?
- Do you prefer visual or textual? | 1.0 | Explore adding site width and padding to global styles - This iteration is inspired by feedback from those exploring patterns and also the visual work explored in full site editing by @shaunandrews. Global styles have a suggested feature enabling to increase or decrease the typography scale. This could easily lead to lines with a few words or too many for comfortable reading. A great additional option could be to allow a site width setting and padding around that site.
I am aware this won't be easy for all themes, so just proposing it here to work around some issues that might come up. I also know that this is being explored in templates, so it might be decided it belongs there instead.
## Option A: textual
This version uses text to indicate the width and padding. It's also worth noting all illustrations show 'layout', which I am not set on and could change. Similar, these use the sidebar as that is what we are using for v1 global styles, this could be comfortable in the toolbar as other iterations happen.
<img width="1257" alt="global-i1" src="https://user-images.githubusercontent.com/253067/76330175-7056f000-62e5-11ea-9441-1365d790855e.png">
It is worth noting that with the iterations to the interface, we don't have a small input with increasing/decreasing arrows. I explored a little what this could look like with a click in the state showing arrows:
<img width="242" alt="Frame 1" src="https://user-images.githubusercontent.com/253067/76330376-ad22e700-62e5-11ea-8292-dddd6464c586.png">
@jasmussen looping you in to get input for interface iterations.
Here it is without explicit arrows:
<img width="1257" alt="global-i2" src="https://user-images.githubusercontent.com/253067/76330409-b6ac4f00-62e5-11ea-96b8-9f75e247af9a.png">
## Option B : visual
This took a large dose of full site editing inspiration, I brought in some visuals used there to explore what that could look like. I think the iconography might need exploring but for me, the visual has a certain something extra I really like.
<img width="1257" alt="global-i3" src="https://user-images.githubusercontent.com/253067/76329911-122a0d00-62e5-11ea-8154-d9402f7026cb.png">
## Feedback
I would love general feedback as this is an early idea and bringing in concepts from elsewhere. That said, it would be good to get specific responses around:
- Does this belong in global styles along with templates?
- Do you prefer visual or textual? | non_perf | explore adding site width and padding to global styles this iteration is inspired by feedback from those exploring patterns and also the visual work explored in full site editing by shaunandrews global styles have a suggested feature enabling to increase or decrease the typography scale this could easily lead to lines with a few words or too many for comfortable reading a great additional option could be to allow a site width setting and padding around that site i am aware this won t be easy for all themes so just proposing it here to work around some issues that might come up i also know that this is being explored in templates so it might be decided it belongs there instead option a textual this version uses text to indicate the width and padding it s also worth noting all illustrations show layout which i am not set on and could change similar these use the sidebar as that is what we are using for global styles this could be comfortable in the toolbar as other iterations happen img width alt global src it is worth noting that with the iterations to the interface we don t have a small input with increasing decreasing arrows i explored a little what this could look like with a click in the state showing arrows img width alt frame src jasmussen looping you in to get input for interface iterations here it is without explicit arrows img width alt global src option b visual this took a large dose of full site editing inspiration i brought in some visuals used there to explore what that could look like i think the iconography might need exploring but for me the visual has a certain something extra i really like img width alt global src feedback i would love general feedback as this is an early idea and bringing in concepts from elsewhere that said it would be good to get specific responses around does this belong in global styles along with templates do you prefer visual or textual | 0 |
47,157 | 24,894,757,522 | IssuesEvent | 2022-10-28 14:53:03 | ualbertalib/jupiter | https://api.github.com/repos/ualbertalib/jupiter | opened | Google detected Largest Contentful Paint | performance | **Describe the bug**
Google systems have detected slow LCP (Largest Contentful Paint) on your website. LCP is an important, user-centric metric for measuring perceived page load speed. This might result in lower search rankings.
An example page with slow loading speed on your site: https://era.library.ualberta.ca/items/ed56e5c2-4392-4137-9996-e9be26c9872d
Visit the Core Web Vitals report linked below to see more affected URLs on your site.
You can resolve this issue by improving components on your pages that may cause high LCP. Please review the articles below for more information.
* https://search.google.com/search-console/core-web-vitals/drilldown?resource_id=https%3A%2F%2Fera.library.ualberta.ca%2F&item_key=CAMQAhgD&hl=en-CA
* https://search.google.com/search-console/core-web-vitals?resource_id=https://era.library.ualberta.ca/&utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA
* https://developers.google.com/search/docs/appearance/page-experience?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA
* https://web.dev/optimize-lcp/?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA
* https://support.google.com/webmasters/community/?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA
**To Reproduce**
Visit a page like https://era.library.ualberta.ca/items/ed56e5c2-4392-4137-9996-e9be26c9872d
**Expected behavior**
Pages take less than 4s.
**Screenshots**


| True | Google detected Largest Contentful Paint - **Describe the bug**
Google systems have detected slow LCP (Largest Contentful Paint) on your website. LCP is an important, user-centric metric for measuring perceived page load speed. This might result in lower search rankings.
An example page with slow loading speed on your site: https://era.library.ualberta.ca/items/ed56e5c2-4392-4137-9996-e9be26c9872d
Visit the Core Web Vitals report linked below to see more affected URLs on your site.
You can resolve this issue by improving components on your pages that may cause high LCP. Please review the articles below for more information.
* https://search.google.com/search-console/core-web-vitals/drilldown?resource_id=https%3A%2F%2Fera.library.ualberta.ca%2F&item_key=CAMQAhgD&hl=en-CA
* https://search.google.com/search-console/core-web-vitals?resource_id=https://era.library.ualberta.ca/&utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA
* https://developers.google.com/search/docs/appearance/page-experience?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA
* https://web.dev/optimize-lcp/?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA
* https://support.google.com/webmasters/community/?utm_source=wnc_20225878&utm_medium=gamma&utm_campaign=wnc_20225878&utm_content=msg_110599462&hl=en-CA
**To Reproduce**
Visit a page like https://era.library.ualberta.ca/items/ed56e5c2-4392-4137-9996-e9be26c9872d
**Expected behavior**
Pages take less than 4s.
**Screenshots**


| perf | google detected largest contentful paint describe the bug google systems have detected slow lcp largest contentful paint on your website lcp is an important user centric metric for measuring perceived page load speed this might result in lower search rankings an example page with slow loading speed on your site visit the core web vitals report linked below to see more affected urls on your site you can resolve this issue by improving components on your pages that may cause high lcp please review the articles below for more information to reproduce visit a page like expected behavior pages take less than screenshots | 1 |
243,203 | 18,677,925,029 | IssuesEvent | 2021-10-31 21:43:58 | eclipse-ee4j/cargotracker | https://api.github.com/repos/eclipse-ee4j/cargotracker | closed | New migration section in the documentation | documentation Priority: Minor | Create a section in the documentation regarding the changes that has been done for migrating from Java EE 7 to Jakarta EE 8, and also in the future for the next versions.
This might help also other folks trying to migrate their projects
@alexmario74, do you mind keeping track of the changes / issues you are facing? If @m-reza-rahman thinks could be a good idea adding the new section. | 1.0 | New migration section in the documentation - Create a section in the documentation regarding the changes that has been done for migrating from Java EE 7 to Jakarta EE 8, and also in the future for the next versions.
This might help also other folks trying to migrate their projects
@alexmario74, do you mind keeping track of the changes / issues you are facing? If @m-reza-rahman thinks could be a good idea adding the new section. | non_perf | new migration section in the documentation create a section in the documentation regarding the changes that has been done for migrating from java ee to jakarta ee and also in the future for the next versions this might help also other folks trying to migrate their projects do you mind keeping track of the changes issues you are facing if m reza rahman thinks could be a good idea adding the new section | 0 |
46,126 | 24,374,396,035 | IssuesEvent | 2022-10-03 22:44:09 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | storage: performance regression in BenchmarkMVCCPutDelete_Pebble | C-performance A-storage T-storage | **Describe the problem**
Currently, on the 22.2 release branch there is a ~20% performance regression in `BenchmarkMVCCPutDelete_Pebble`, relative to 22.1:
```bash
$ benchstat bench-22.1 bench-22.2
name old time/op new time/op delta
MVCCPutDelete_Pebble-16 8.36µs ± 2% 10.09µs ± 1% +20.71% (p=0.000 n=10+10)
```
**To Reproduce**
```bash
$ ./dev bench ./pkg/storage --filter MVCCPutDelete_Pebble --count 10 -v --stream-output --ignore-cache
``` | True | storage: performance regression in BenchmarkMVCCPutDelete_Pebble - **Describe the problem**
Currently, on the 22.2 release branch there is a ~20% performance regression in `BenchmarkMVCCPutDelete_Pebble`, relative to 22.1:
```bash
$ benchstat bench-22.1 bench-22.2
name old time/op new time/op delta
MVCCPutDelete_Pebble-16 8.36µs ± 2% 10.09µs ± 1% +20.71% (p=0.000 n=10+10)
```
**To Reproduce**
```bash
$ ./dev bench ./pkg/storage --filter MVCCPutDelete_Pebble --count 10 -v --stream-output --ignore-cache
``` | perf | storage performance regression in benchmarkmvccputdelete pebble describe the problem currently on the release branch there is a performance regression in benchmarkmvccputdelete pebble relative to bash benchstat bench bench name old time op new time op delta mvccputdelete pebble ± ± p n to reproduce bash dev bench pkg storage filter mvccputdelete pebble count v stream output ignore cache | 1 |
34,035 | 16,385,360,853 | IssuesEvent | 2021-05-17 09:43:23 | space-wizards/RobustToolbox | https://api.github.com/repos/space-wizards/RobustToolbox | closed | Cache GridId and MapId inside TransformComponent | Area: ECS Project: Shared Type: Performance | These now require recursive lookups via the parent to calculate and it's *extremely* expensive. | True | Cache GridId and MapId inside TransformComponent - These now require recursive lookups via the parent to calculate and it's *extremely* expensive. | perf | cache gridid and mapid inside transformcomponent these now require recursive lookups via the parent to calculate and it s extremely expensive | 1 |
35,422 | 17,079,159,128 | IssuesEvent | 2021-07-08 00:43:31 | hashgraph/hedera-mirror-node | https://api.github.com/repos/hashgraph/hedera-mirror-node | closed | Refactor importer NFT logic to use upsert | P1 enhancement nft parser performance | **Problem**
The importer is currently using old logic for handling NFTs and NFT transfers, as opposed to the new upsert logic added in #2006.
**Solution**
Refactor the importer to use similar upsert logic for NFT and NFT transfer handling.
**Alternatives**
**Additional Context**
There may need to be some rework around how NFT transfers update the accountId of the NFT, primarily during an NFT mint transaction. | True | Refactor importer NFT logic to use upsert - **Problem**
The importer is currently using old logic for handling NFTs and NFT transfers, as opposed to the new upsert logic added in #2006.
**Solution**
Refactor the importer to use similar upsert logic for NFT and NFT transfer handling.
**Alternatives**
**Additional Context**
There may need to be some rework around how NFT transfers update the accountId of the NFT, primarily during an NFT mint transaction. | perf | refactor importer nft logic to use upsert problem the importer is currently using old logic for handling nfts and nft transfers as opposed to the new upsert logic added in solution refactor the importer to use similar upsert logic for nft and nft transfer handling alternatives additional context there may need to be some rework around how nft transfers update the accountid of the nft primarily during an nft mint transaction | 1 |
366,983 | 10,832,760,749 | IssuesEvent | 2019-11-11 11:21:03 | francismaria/MaTheX2Java | https://api.github.com/repos/francismaria/MaTheX2Java | closed | Add error panel | frontend medium priority | Add an error panel to show messages of errors that occurred during the code generation. | 1.0 | Add error panel - Add an error panel to show messages of errors that occurred during the code generation. | non_perf | add error panel add an error panel to show messages of errors that occurred during the code generation | 0 |
308,744 | 23,265,202,973 | IssuesEvent | 2022-08-04 16:41:12 | Qiskit/qiskit_sphinx_theme | https://api.github.com/repos/Qiskit/qiskit_sphinx_theme | opened | Set up COS deployment | documentation docs migration | **As a** Qiskit docs dev,
**I need** to be able to deploy html files from this repository to COS when work is merged into main
**So that** frontend development work can be more efficient (i.e. decoupled from qiskit releases) | 1.0 | Set up COS deployment - **As a** Qiskit docs dev,
**I need** to be able to deploy html files from this repository to COS when work is merged into main
**So that** frontend development work can be more efficient (i.e. decoupled from qiskit releases) | non_perf | set up cos deployment as a qiskit docs dev i need to be able to deploy html files from this repository to cos when work is merged into main so that frontend development work can be more efficient i e decoupled from qiskit releases | 0 |
56,341 | 31,881,639,719 | IssuesEvent | 2023-09-16 12:52:44 | neovim/neovim | https://api.github.com/repos/neovim/neovim | closed | Lag is noticeable within comment while the treesitter highlight is enabled. | performance treesitter | ### Problem
This issue is a bit hard to reproduce without following exactly steps.
When typing `/**` comment in vue files, the input lag is noticeable.
### Steps to reproduce
put `slow-ts.lua` in `/tmp`:
```lua
vim.g.mapleader = " "
vim.g.maplocalleader = ","
vim.opt.number = true
for name, url in pairs({
-- ADD PLUGINS _NECESSARY_ TO REPRODUCE THE ISSUE, e.g:
-- some_plugin = 'https://github.com/author/plugin.nvim'
treesitter = "https://github.com/nvim-treesitter/nvim-treesitter",
}) do
local install_path = vim.fn.fnamemodify("slow_ts/" .. name, ":p")
if vim.fn.isdirectory(install_path) == 0 then
if name == "treesitter" then
vim.fn.system({ "git", "clone", "--branch", "main", "--depth=1", url, install_path })
else
vim.fn.system({ "git", "clone", "--depth=1", url, install_path })
end
end
vim.opt.runtimepath:append(install_path)
end
-- ===============
require("nvim-treesitter").setup({
ensure_install = { "vue", "javascript", "typescript", "comment", "jsdoc" },
auto_install = false,
install_dir = vim.fn.fnamemodify("slow_ts/.site", ":p"),
})
local ftau = vim.api.nvim_create_augroup("option_ft", { clear = true })
vim.api.nvim_create_autocmd("FileType", {
group = ftau,
pattern = { "*" },
callback = function(args)
local buf = args.buf
-- start highlighter.
if not pcall(vim.treesitter.start, buf) then
print("ERROR! highlight not started")
return
end
end,
})
vim.cmd("colorscheme default")
```
Put a `test.vue` file along side with `slow-ts.lua`:
file content: https://gist.github.com/towry/74d6ff3312e18c6fd39367d1b4e477cf#file-test-vue
Steps to repro it:
1. `nvim --clean -u slow-ts.lua`, wait for the nvim-treesitter setup is done.
2. make sure `vue`, `comment` parsers is installed.
3. `nvim --clean -u slow-ts.lua test.vue`
4. (important) press `ctrl-d` multiple times to move cursor down and press `ctrl-u` to make sure line 26 is visible and move cursor to line 26 with `j/k`.
5. start insert comment with content `/***` above line 26.
Note: it seems if I press `:26` to directly move cursor to line 26, this bug will not be triggered.
A lag demo can be seen here: https://github.com/neovim/neovim/issues/25074#issuecomment-1713502035
### Expected behavior
No lag in comment.
### Neovim version (nvim -v)
NVIM v0.10.0-dev-1088+g6b5f44817
### Vim (not Nvim) behaves the same?
no
### Operating system/version
macos
### Terminal name/version
kitty + macos terminal
### $TERM environment variable
xterm-256color
### Installation
build from repo | True | Lag is noticeable within comment while the treesitter highlight is enabled. - ### Problem
This issue is a bit hard to reproduce without following exactly steps.
When typing `/**` comment in vue files, the input lag is noticeable.
### Steps to reproduce
put `slow-ts.lua` in `/tmp`:
```lua
vim.g.mapleader = " "
vim.g.maplocalleader = ","
vim.opt.number = true
for name, url in pairs({
-- ADD PLUGINS _NECESSARY_ TO REPRODUCE THE ISSUE, e.g:
-- some_plugin = 'https://github.com/author/plugin.nvim'
treesitter = "https://github.com/nvim-treesitter/nvim-treesitter",
}) do
local install_path = vim.fn.fnamemodify("slow_ts/" .. name, ":p")
if vim.fn.isdirectory(install_path) == 0 then
if name == "treesitter" then
vim.fn.system({ "git", "clone", "--branch", "main", "--depth=1", url, install_path })
else
vim.fn.system({ "git", "clone", "--depth=1", url, install_path })
end
end
vim.opt.runtimepath:append(install_path)
end
-- ===============
require("nvim-treesitter").setup({
ensure_install = { "vue", "javascript", "typescript", "comment", "jsdoc" },
auto_install = false,
install_dir = vim.fn.fnamemodify("slow_ts/.site", ":p"),
})
local ftau = vim.api.nvim_create_augroup("option_ft", { clear = true })
vim.api.nvim_create_autocmd("FileType", {
group = ftau,
pattern = { "*" },
callback = function(args)
local buf = args.buf
-- start highlighter.
if not pcall(vim.treesitter.start, buf) then
print("ERROR! highlight not started")
return
end
end,
})
vim.cmd("colorscheme default")
```
Put a `test.vue` file along side with `slow-ts.lua`:
file content: https://gist.github.com/towry/74d6ff3312e18c6fd39367d1b4e477cf#file-test-vue
Steps to repro it:
1. `nvim --clean -u slow-ts.lua`, wait for the nvim-treesitter setup is done.
2. make sure `vue`, `comment` parsers is installed.
3. `nvim --clean -u slow-ts.lua test.vue`
4. (important) press `ctrl-d` multiple times to move cursor down and press `ctrl-u` to make sure line 26 is visible and move cursor to line 26 with `j/k`.
5. start insert comment with content `/***` above line 26.
Note: it seems if I press `:26` to directly move cursor to line 26, this bug will not be triggered.
A lag demo can be seen here: https://github.com/neovim/neovim/issues/25074#issuecomment-1713502035
### Expected behavior
No lag in comment.
### Neovim version (nvim -v)
NVIM v0.10.0-dev-1088+g6b5f44817
### Vim (not Nvim) behaves the same?
no
### Operating system/version
macos
### Terminal name/version
kitty + macos terminal
### $TERM environment variable
xterm-256color
### Installation
build from repo | perf | lag is noticeable within comment while the treesitter highlight is enabled problem this issue is a bit hard to reproduce without following exactly steps when typing comment in vue files the input lag is noticeable steps to reproduce put slow ts lua in tmp lua vim g mapleader vim g maplocalleader vim opt number true for name url in pairs add plugins necessary to reproduce the issue e g some plugin treesitter do local install path vim fn fnamemodify slow ts name p if vim fn isdirectory install path then if name treesitter then vim fn system git clone branch main depth url install path else vim fn system git clone depth url install path end end vim opt runtimepath append install path end require nvim treesitter setup ensure install vue javascript typescript comment jsdoc auto install false install dir vim fn fnamemodify slow ts site p local ftau vim api nvim create augroup option ft clear true vim api nvim create autocmd filetype group ftau pattern callback function args local buf args buf start highlighter if not pcall vim treesitter start buf then print error highlight not started return end end vim cmd colorscheme default put a test vue file along side with slow ts lua file content steps to repro it nvim clean u slow ts lua wait for the nvim treesitter setup is done make sure vue comment parsers is installed nvim clean u slow ts lua test vue important press ctrl d multiple times to move cursor down and press ctrl u to make sure line is visible and move cursor to line with j k start insert comment with content above line note it seems if i press to directly move cursor to line this bug will not be triggered a lag demo can be seen here expected behavior no lag in comment neovim version nvim v nvim dev vim not nvim behaves the same no operating system version macos terminal name version kitty macos terminal term environment variable xterm installation build from repo | 1 |
17,334 | 9,703,041,157 | IssuesEvent | 2019-05-27 10:15:30 | Dart-Code/Dart-Code | https://api.github.com/repos/Dart-Code/Dart-Code | opened | Improve performance of completions | important is performance | As well as probably being the cause of #1721, the performance of completions with unimported symbols is poor.
Testing on Windows PC with the following test (inside `test\flutter_only\providers\completion_item_provider.test.ts`):
```
describe("with SuggestionSet support", () => {
beforeEach("ensure SuggestionSets are supported", function () {
if (!extApi.analyzerCapabilities.supportsAvailableSuggestions)
this.skip();
});
it.only("includes unimported symbols", async () => {
await setTestContent(`
main() {
ProcessInf
}
`);
const count = 50;
const start = Date.now();
for (let i = 0; i < 50; i++) {
const completions = await getCompletionsAt("ProcessInf^");
ensureCompletion(completions, vs.CompletionItemKind.Class, "ProcessInfo", "ProcessInfo");
}
const end = Date.now();
console.log(`Took ${end - start}ms to do ${count} completion requests`);
});
});
```
Output from a few runs:
```
Took 56426ms to do 50 completion requests
Took 57034ms to do 50 completion requests
Took 57607ms to do 50 completion requests
``` | True | Improve performance of completions - As well as probably being the cause of #1721, the performance of completions with unimported symbols is poor.
Testing on Windows PC with the following test (inside `test\flutter_only\providers\completion_item_provider.test.ts`):
```
describe("with SuggestionSet support", () => {
beforeEach("ensure SuggestionSets are supported", function () {
if (!extApi.analyzerCapabilities.supportsAvailableSuggestions)
this.skip();
});
it.only("includes unimported symbols", async () => {
await setTestContent(`
main() {
ProcessInf
}
`);
const count = 50;
const start = Date.now();
for (let i = 0; i < 50; i++) {
const completions = await getCompletionsAt("ProcessInf^");
ensureCompletion(completions, vs.CompletionItemKind.Class, "ProcessInfo", "ProcessInfo");
}
const end = Date.now();
console.log(`Took ${end - start}ms to do ${count} completion requests`);
});
});
```
Output from a few runs:
```
Took 56426ms to do 50 completion requests
Took 57034ms to do 50 completion requests
Took 57607ms to do 50 completion requests
``` | perf | improve performance of completions as well as probably being the cause of the performance of completions with unimported symbols is poor testing on windows pc with the following test inside test flutter only providers completion item provider test ts describe with suggestionset support beforeeach ensure suggestionsets are supported function if extapi analyzercapabilities supportsavailablesuggestions this skip it only includes unimported symbols async await settestcontent main processinf const count const start date now for let i i i const completions await getcompletionsat processinf ensurecompletion completions vs completionitemkind class processinfo processinfo const end date now console log took end start ms to do count completion requests output from a few runs took to do completion requests took to do completion requests took to do completion requests | 1 |
49,950 | 26,403,482,641 | IssuesEvent | 2023-01-13 04:56:34 | mmtk/mmtk-julia | https://api.github.com/repos/mmtk/mmtk-julia | opened | Improving the performance when querying object size | performance | Julia's stock GC uses page metadata information to ask for the size of smaller objects (pool-allocated). Since MMTk does not have that information available, we perform a [check](https://github.com/mmtk/mmtk-julia/blob/970fc5c49c2f9edfefbffdcff9ec753cf0791fa1/julia/mmtk_julia.c#L318) on the object type to calculate its size, which is likely to cause performance problems. | True | Improving the performance when querying object size - Julia's stock GC uses page metadata information to ask for the size of smaller objects (pool-allocated). Since MMTk does not have that information available, we perform a [check](https://github.com/mmtk/mmtk-julia/blob/970fc5c49c2f9edfefbffdcff9ec753cf0791fa1/julia/mmtk_julia.c#L318) on the object type to calculate its size, which is likely to cause performance problems. | perf | improving the performance when querying object size julia s stock gc uses page metadata information to ask for the size of smaller objects pool allocated since mmtk does not have that information available we perform a on the object type to calculate its size which is likely to cause performance problems | 1 |
34,526 | 16,594,188,279 | IssuesEvent | 2021-06-01 11:30:03 | ethereumjs/ethereumjs-monorepo | https://api.github.com/repos/ethereumjs/ethereumjs-monorepo | closed | VM: Analyze debug functionality performance behavior | eval: needs discussion eval: needs validation package: vm prio: P3 important type: enhancement type: performance | Debug functionality with the `debug` package introduced in #1080 and follow-up PRs might have been coming along with some performance penalty, as raised by @alcuadrado in the chat. This is still unconfirmed though and we should validate before we act upon this.
Performance influencing code parts are mainly (solely?) the preparations for the `debug` strings, posting a random example here from `runBlock.ts`:
```typescript
debug(`Invalid gasUsed received=${result.gasUsed} expected=${block.header.gasUsed}`)
```
A structural solution might be to only run/create the `debug` messages if the `VM` has been passed a flag `debug`. This also needs some evaluation/proof if it really solves the problem or if strings are build nonetheless.
An alternative/additional first somewhat more targeted step would be to have a look at the different calls and eliminate/refactor all calls with computation. I tried to be really careful with this on implementation, but there are e.g. some `hash()` calls added (or at least one?).
Here is an example from `runTx.ts`:
```typescript
`debug(`New tx run hash=${opts.tx.hash().toString('hex')} sender=${caller.toString()}`)`
```
(I would personally be a fan of this second approach before here and then give this again some more time or eventually close directly. Respectively we can do this targeted optimization anyhow since this *will* have *some* effect on performance) | True | VM: Analyze debug functionality performance behavior - Debug functionality with the `debug` package introduced in #1080 and follow-up PRs might have been coming along with some performance penalty, as raised by @alcuadrado in the chat. This is still unconfirmed though and we should validate before we act upon this.
Performance influencing code parts are mainly (solely?) the preparations for the `debug` strings, posting a random example here from `runBlock.ts`:
```typescript
debug(`Invalid gasUsed received=${result.gasUsed} expected=${block.header.gasUsed}`)
```
A structural solution might be to only run/create the `debug` messages if the `VM` has been passed a flag `debug`. This also needs some evaluation/proof if it really solves the problem or if strings are build nonetheless.
An alternative/additional first somewhat more targeted step would be to have a look at the different calls and eliminate/refactor all calls with computation. I tried to be really careful with this on implementation, but there are e.g. some `hash()` calls added (or at least one?).
Here is an example from `runTx.ts`:
```typescript
`debug(`New tx run hash=${opts.tx.hash().toString('hex')} sender=${caller.toString()}`)`
```
(I would personally be a fan of this second approach before here and then give this again some more time or eventually close directly. Respectively we can do this targeted optimization anyhow since this *will* have *some* effect on performance) | perf | vm analyze debug functionality performance behavior debug functionality with the debug package introduced in and follow up prs might have been coming along with some performance penalty as raised by alcuadrado in the chat this is still unconfirmed though and we should validate before we act upon this performance influencing code parts are mainly solely the preparations for the debug strings posting a random example here from runblock ts typescript debug invalid gasused received result gasused expected block header gasused a structural solution might be to only run create the debug messages if the vm has been passed a flag debug this also needs some evaluation proof if it really solves the problem or if strings are build nonetheless an alternative additional first somewhat more targeted step would be to have a look at the different calls and eliminate refactor all calls with computation i tried to be really careful with this on implementation but there are e g some hash calls added or at least one here is an example from runtx ts typescript debug new tx run hash opts tx hash tostring hex sender caller tostring i would personally be a fan of this second approach before here and then give this again some more time or eventually close directly respectively we can do this targeted optimization anyhow since this will have some effect on performance | 1 |
193,625 | 15,382,659,100 | IssuesEvent | 2021-03-03 01:05:52 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | Documentation for ApplyNegMuCorrection Algorithm | Documentation Muon Stale | Create documentation for ApplyNegMuCorrection Algorithm #13201
| 1.0 | Documentation for ApplyNegMuCorrection Algorithm - Create documentation for ApplyNegMuCorrection Algorithm #13201
| non_perf | documentation for applynegmucorrection algorithm create documentation for applynegmucorrection algorithm | 0 |
382,955 | 26,525,950,286 | IssuesEvent | 2023-01-19 08:44:50 | Tinkoff/taiga-ui | https://api.github.com/repos/Tinkoff/taiga-ui | closed | 📚 - Palette in dropdown code example is wrong | documentation P3 | ### What is the affected URL?
https://taiga-ui.dev/next/components/color-picker#palette
### Description
The **Palette in dropdown** sample code was published just like its predecessor, **ColorSelector in dropdown**
### Which browsers have you used?
- [X] Chrome
- [X] Firefox
- [ ] Safari
- [ ] Edge
### Which operating systems have you used?
- [ ] macOS
- [X] Windows
- [ ] Linux
- [ ] iOS
- [ ] Android | 1.0 | 📚 - Palette in dropdown code example is wrong - ### What is the affected URL?
https://taiga-ui.dev/next/components/color-picker#palette
### Description
The **Palette in dropdown** sample code was published just like its predecessor, **ColorSelector in dropdown**
### Which browsers have you used?
- [X] Chrome
- [X] Firefox
- [ ] Safari
- [ ] Edge
### Which operating systems have you used?
- [ ] macOS
- [X] Windows
- [ ] Linux
- [ ] iOS
- [ ] Android | non_perf | 📚 palette in dropdown code example is wrong what is the affected url description the palette in dropdown sample code was published just like its predecessor colorselector in dropdown which browsers have you used chrome firefox safari edge which operating systems have you used macos windows linux ios android | 0 |
43,607 | 23,304,860,903 | IssuesEvent | 2022-08-07 21:37:35 | sourcefrog/cargo-mutants | https://api.github.com/repos/sourcefrog/cargo-mutants | closed | Maybe use `--release` by default | enhancement good first issue performance | `cargo mutants` does many incremental builds and test runs. In most trees, most of the time is spent running the tests.
Tests will typically run faster with `--release`; in some crates it will be much faster.
So that should at least be an option (#59) but according to the principle of "make it easy" perhaps `--release` should be the default.
There are at least two potential downsides:
1. Conceivably in some trees the performance hit from doing the initial and incremental builds in release mode will outweigh the benefit of the tests being faster. However, even if that is true sometimes it's probably a win across most trees.
2. Some projects might have test suites that are less rigorous, or simply don't work properly, in release mode. For example they might rely on `debug_assert` to catch some bugs. I think this would be an unidiomatic use of Rust, but it might well happen.
I think those could both reasonably be handled by having an option to go back to debug builds. | True | Maybe use `--release` by default - `cargo mutants` does many incremental builds and test runs. In most trees, most of the time is spent running the tests.
Tests will typically run faster with `--release`; in some crates it will be much faster.
So that should at least be an option (#59) but according to the principle of "make it easy" perhaps `--release` should be the default.
There are at least two potential downsides:
1. Conceivably in some trees the performance hit from doing the initial and incremental builds in release mode will outweigh the benefit of the tests being faster. However, even if that is true sometimes it's probably a win across most trees.
2. Some projects might have test suites that are less rigorous, or simply don't work properly, in release mode. For example they might rely on `debug_assert` to catch some bugs. I think this would be an unidiomatic use of Rust, but it might well happen.
I think those could both reasonably be handled by having an option to go back to debug builds. | perf | maybe use release by default cargo mutants does many incremental builds and test runs in most trees most of the time is spent running the tests tests will typically run faster with release in some crates it will be much faster so that should at least be an option but according to the principle of make it easy perhaps release should be the default there are at least two potential downsides conceivably in some trees the performance hit from doing the initial and incremental builds in release mode will outweigh the benefit of the tests being faster however even if that is true sometimes it s probably a win across most trees some projects might have test suites that are less rigorous or simply don t work properly in release mode for example they might rely on debug assert to catch some bugs i think this would be an unidiomatic use of rust but it might well happen i think those could both reasonably be handled by having an option to go back to debug builds | 1 |
321,655 | 27,546,266,451 | IssuesEvent | 2023-03-07 12:01:36 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | opened | [Flaky Test] can be created without any text selected | [Type] Flaky Test | <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
can be created without any text selected
## Test path
`specs/editor/various/links.test.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2023-03-07T12:01:35.760Z"><code>[2023-03-07T12:01:35.760Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4353563197"><code>try/dedicated-block-controls-filter</code></a>.
</summary>
```
● Links › can be created without any text selected
expect(jest.fn()).not.toHaveErrored(expected)
Expected mock function not to be called but it was called with:
["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."]
at Object.assertExpectedCalls (../jest-console/build/@wordpress/jest-console/src/index.js:32:4)
at runMicrotasks (<anonymous>)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| 1.0 | [Flaky Test] can be created without any text selected - <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
can be created without any text selected
## Test path
`specs/editor/various/links.test.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2023-03-07T12:01:35.760Z"><code>[2023-03-07T12:01:35.760Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4353563197"><code>try/dedicated-block-controls-filter</code></a>.
</summary>
```
● Links › can be created without any text selected
expect(jest.fn()).not.toHaveErrored(expected)
Expected mock function not to be called but it was called with:
["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."],["The hook name can only contain numbers, letters, dashes, periods and underscores."]
at Object.assertExpectedCalls (../jest-console/build/@wordpress/jest-console/src/index.js:32:4)
at runMicrotasks (<anonymous>)
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
| non_perf | can be created without any text selected flaky test detected this is an auto generated issue by github actions please do not edit this manually test title can be created without any text selected test path specs editor various links test js errors test passed after failed attempt on a href ● links › can be created without any text selected expect jest fn not tohaveerrored expected expected mock function not to be called but it was called with at object assertexpectedcalls jest console build wordpress jest console src index js at runmicrotasks | 0 |
2,798 | 3,012,603,091 | IssuesEvent | 2015-07-29 00:41:56 | orbeon/orbeon-forms | https://api.github.com/repos/orbeon/orbeon-forms | closed | Embedding must support Form Builder | Form Builder | Right now, server-side embedding doesn't work well with Form Builder. We need to make this work properly. | 1.0 | Embedding must support Form Builder - Right now, server-side embedding doesn't work well with Form Builder. We need to make this work properly. | non_perf | embedding must support form builder right now server side embedding doesn t work well with form builder we need to make this work properly | 0 |
10,328 | 7,154,935,519 | IssuesEvent | 2018-01-26 10:35:42 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | closed | Run developer-uploaded images through pngcrush/optipng/jpegoptim | component: performance triaged | We load a lot of images we don't control, and we should make sure they are as optimized as they can. This means running them through lossless compression tools like pngcrush, optipng, and jpegoptim. We can save a lot by doing this.
We did this in Marketplace a while ago. The main caveat was that cachebusting needed to take this into account, because it had to happen on a separate task, so we'd resize the image, save the file, then optimize them, overwriting the previous copy. If somehow a client accessed the resized image before the optimize step, we needed to load the newer version. It can work with the naive way we do cachebusting right now, but it would be better with https://github.com/mozilla/addons-server/issues/2659 instead. | True | Run developer-uploaded images through pngcrush/optipng/jpegoptim - We load a lot of images we don't control, and we should make sure they are as optimized as they can. This means running them through lossless compression tools like pngcrush, optipng, and jpegoptim. We can save a lot by doing this.
We did this in Marketplace a while ago. The main caveat was that cachebusting needed to take this into account, because it had to happen on a separate task, so we'd resize the image, save the file, then optimize them, overwriting the previous copy. If somehow a client accessed the resized image before the optimize step, we needed to load the newer version. It can work with the naive way we do cachebusting right now, but it would be better with https://github.com/mozilla/addons-server/issues/2659 instead. | perf | run developer uploaded images through pngcrush optipng jpegoptim we load a lot of images we don t control and we should make sure they are as optimized as they can this means running them through lossless compression tools like pngcrush optipng and jpegoptim we can save a lot by doing this we did this in marketplace a while ago the main caveat was that cachebusting needed to take this into account because it had to happen on a separate task so we d resize the image save the file then optimize them overwriting the previous copy if somehow a client accessed the resized image before the optimize step we needed to load the newer version it can work with the naive way we do cachebusting right now but it would be better with instead | 1 |
3,192 | 3,833,125,140 | IssuesEvent | 2016-04-01 01:00:03 | piwik/piwik | https://api.github.com/repos/piwik/piwik | closed | Improve all websites dashboard performance | c: Performance | All websites dashboard loads without any hassle when fetched for single day or week. It is problematic when you fetch 2 days for example, merging the data lasts for very long and usually does not finish. This occurrence was observed in instance with around 2k websites, this definitely is important for testing purposes. | True | Improve all websites dashboard performance - All websites dashboard loads without any hassle when fetched for single day or week. It is problematic when you fetch 2 days for example, merging the data lasts for very long and usually does not finish. This occurrence was observed in instance with around 2k websites, this definitely is important for testing purposes. | perf | improve all websites dashboard performance all websites dashboard loads without any hassle when fetched for single day or week it is problematic when you fetch days for example merging the data lasts for very long and usually does not finish this occurrence was observed in instance with around websites this definitely is important for testing purposes | 1 |
24,084 | 12,213,898,885 | IssuesEvent | 2020-05-01 08:26:23 | zeebe-io/zeebe | https://api.github.com/repos/zeebe-io/zeebe | closed | FileChannelJournalSegmentReader throws half buffers away | Impact: Performance Scope: broker Type: Bug Type: Maintenance | **Description**
During our investigation of the engine latency we found out that the atomix journal is a big hotspot performance wise. We had a deeper look at the code and found the `readNext` method.
We have to say before that per Zeebe default the `maxEntrySize` is 4 MB. The reader will allocate a buffer which is `8MB` big. Our assumption is that it wants to do a read a head, but actually it doesn't because it clears the buffer positions etc.
```java
/** Reads the next entry in the segment. */
@SuppressWarnings("unchecked")
private void readNext() {
// Compute the index of the next entry in the segment.
final long index = getNextIndex();
try {
// Read more bytes from the segment if necessary.
if (memory.remaining() < maxEntrySize) {
final long position = channel.position() + memory.position();
channel.position(position);
memory.clear(); // <=== this resets to position 0, limit to capacity
channel.read(memory); // <=== we read now again 8 MB
channel.position(position);
memory.flip();
}
// Mark the buffer so it can be reset if necessary.
memory.mark();
try {
// Read the length of the entry.
final int length = memory.getInt();
// If the buffer length is zero then return.
if (length <= 0 || length > maxEntrySize) {
memory.reset().limit(memory.position());
nextEntry = null;
return;
}
// Read the checksum of the entry.
final long checksum = memory.getInt() & 0xFFFFFFFFL;
// Compute the checksum for the entry bytes.
final Checksum crc32 = new CRC32();
crc32.update(memory.array(), memory.position(), length);
// If the stored checksum equals the computed checksum, return the entry.
if (checksum == crc32.getValue()) {
final int limit = memory.limit();
memory.limit(memory.position() + length);
final E entry = namespace.deserialize(memory);
memory.limit(limit);
nextEntry = new Indexed<>(index, entry, length);
} else {
memory.reset().limit(memory.position());
nextEntry = null;
}
} catch (final BufferUnderflowException e) {
memory.reset().limit(memory.position());
nextEntry = null;
}
} catch (final IOException e) {
throw new StorageException(e);
}
}
```
This means we will read `8 MB` process from these `~4 MB`until our `remaining` is less then `maxEntrySize`, so after we reached half of the buffer. Then we will read again `8MB` where the other remaining is thrown away and will end at the beginning of the buffer, because it is re-read.
This problem is more an issue when the `maxEntrySize` and `realEntrySize` is very different, which is our case. We using `4MB` to support big deployments but actually most of our records are less then 1 kb. If the sizes would be nearer then we would probably also read most of the buffer, but this is not our use case so we should fix this!
| True | FileChannelJournalSegmentReader throws half buffers away - **Description**
During our investigation of the engine latency we found out that the atomix journal is a big hotspot performance wise. We had a deeper look at the code and found the `readNext` method.
We have to say before that per Zeebe default the `maxEntrySize` is 4 MB. The reader will allocate a buffer which is `8MB` big. Our assumption is that it wants to do a read a head, but actually it doesn't because it clears the buffer positions etc.
```java
/** Reads the next entry in the segment. */
@SuppressWarnings("unchecked")
private void readNext() {
// Compute the index of the next entry in the segment.
final long index = getNextIndex();
try {
// Read more bytes from the segment if necessary.
if (memory.remaining() < maxEntrySize) {
final long position = channel.position() + memory.position();
channel.position(position);
memory.clear(); // <=== this resets to position 0, limit to capacity
channel.read(memory); // <=== we read now again 8 MB
channel.position(position);
memory.flip();
}
// Mark the buffer so it can be reset if necessary.
memory.mark();
try {
// Read the length of the entry.
final int length = memory.getInt();
// If the buffer length is zero then return.
if (length <= 0 || length > maxEntrySize) {
memory.reset().limit(memory.position());
nextEntry = null;
return;
}
// Read the checksum of the entry.
final long checksum = memory.getInt() & 0xFFFFFFFFL;
// Compute the checksum for the entry bytes.
final Checksum crc32 = new CRC32();
crc32.update(memory.array(), memory.position(), length);
// If the stored checksum equals the computed checksum, return the entry.
if (checksum == crc32.getValue()) {
final int limit = memory.limit();
memory.limit(memory.position() + length);
final E entry = namespace.deserialize(memory);
memory.limit(limit);
nextEntry = new Indexed<>(index, entry, length);
} else {
memory.reset().limit(memory.position());
nextEntry = null;
}
} catch (final BufferUnderflowException e) {
memory.reset().limit(memory.position());
nextEntry = null;
}
} catch (final IOException e) {
throw new StorageException(e);
}
}
```
This means we will read `8 MB` process from these `~4 MB`until our `remaining` is less then `maxEntrySize`, so after we reached half of the buffer. Then we will read again `8MB` where the other remaining is thrown away and will end at the beginning of the buffer, because it is re-read.
This problem is more an issue when the `maxEntrySize` and `realEntrySize` is very different, which is our case. We using `4MB` to support big deployments but actually most of our records are less then 1 kb. If the sizes would be nearer then we would probably also read most of the buffer, but this is not our use case so we should fix this!
| perf | filechanneljournalsegmentreader throws half buffers away description during our investigation of the engine latency we found out that the atomix journal is a big hotspot performance wise we had a deeper look at the code and found the readnext method we have to say before that per zeebe default the maxentrysize is mb the reader will allocate a buffer which is big our assumption is that it wants to do a read a head but actually it doesn t because it clears the buffer positions etc java reads the next entry in the segment suppresswarnings unchecked private void readnext compute the index of the next entry in the segment final long index getnextindex try read more bytes from the segment if necessary if memory remaining maxentrysize final long position channel position memory position channel position position memory clear this resets to position limit to capacity channel read memory we read now again mb channel position position memory flip mark the buffer so it can be reset if necessary memory mark try read the length of the entry final int length memory getint if the buffer length is zero then return if length maxentrysize memory reset limit memory position nextentry null return read the checksum of the entry final long checksum memory getint compute the checksum for the entry bytes final checksum new update memory array memory position length if the stored checksum equals the computed checksum return the entry if checksum getvalue final int limit memory limit memory limit memory position length final e entry namespace deserialize memory memory limit limit nextentry new indexed index entry length else memory reset limit memory position nextentry null catch final bufferunderflowexception e memory reset limit memory position nextentry null catch final ioexception e throw new storageexception e this means we will read mb process from these mb until our remaining is less then maxentrysize so after we reached half of the buffer then we will read again where the other remaining is thrown away and will end at the beginning of the buffer because it is re read this problem is more an issue when the maxentrysize and realentrysize is very different which is our case we using to support big deployments but actually most of our records are less then kb if the sizes would be nearer then we would probably also read most of the buffer but this is not our use case so we should fix this | 1 |
16,617 | 9,475,902,622 | IssuesEvent | 2019-04-19 13:16:05 | JuliaOpt/JuMP.jl | https://api.github.com/repos/JuliaOpt/JuMP.jl | closed | Extreme performance degradation from v0.18.5 -> v0.19 | bug performance | Heya JuMP team!
I've currently noticed a large performance regression (which has been cited in a few cases, notably #1403 and #1905). I was planning on releasing some code for a paper, which was originally written with v0.18.5, but, with the upgrade, decided to do the necessary edits for v0.19. The code is essentially (sadly :( ) unusable in its current form, due to the speed degradation.
Two MWEs showing a large difference (I will note that I'm using two slightly different environments, with Julia 1.0 and Julia 1.1)
[Julia 1.0, JuMP v0.18.5](https://gist.github.com/angeris/8ac92dee79081c946979bcce8160f656). Time to complete: <2s. Roughly .0001s/constraint.
[Julia 1.1, JuMP v0.19](https://gist.github.com/angeris/2f3ae08642c001d8f0ccf7aad03a6a8f). ETA for completing this (as reported by `ProgressMeter`): 18 minutes. Roughly 1.3s/constraint.
The problem in question solved was much, much larger and would take several days to formulate at this pace. I'd also be happy to contribute to a fix, but I'd need to learn a bit more about profiling and performance in Julia (and definitely more about the current JuMP implementation) in order to help out!
Anyways, thank you so much for this project, it's been super useful :) | True | Extreme performance degradation from v0.18.5 -> v0.19 - Heya JuMP team!
I've currently noticed a large performance regression (which has been cited in a few cases, notably #1403 and #1905). I was planning on releasing some code for a paper, which was originally written with v0.18.5, but, with the upgrade, decided to do the necessary edits for v0.19. The code is essentially (sadly :( ) unusable in its current form, due to the speed degradation.
Two MWEs showing a large difference (I will note that I'm using two slightly different environments, with Julia 1.0 and Julia 1.1)
[Julia 1.0, JuMP v0.18.5](https://gist.github.com/angeris/8ac92dee79081c946979bcce8160f656). Time to complete: <2s. Roughly .0001s/constraint.
[Julia 1.1, JuMP v0.19](https://gist.github.com/angeris/2f3ae08642c001d8f0ccf7aad03a6a8f). ETA for completing this (as reported by `ProgressMeter`): 18 minutes. Roughly 1.3s/constraint.
The problem in question solved was much, much larger and would take several days to formulate at this pace. I'd also be happy to contribute to a fix, but I'd need to learn a bit more about profiling and performance in Julia (and definitely more about the current JuMP implementation) in order to help out!
Anyways, thank you so much for this project, it's been super useful :) | perf | extreme performance degradation from heya jump team i ve currently noticed a large performance regression which has been cited in a few cases notably and i was planning on releasing some code for a paper which was originally written with but with the upgrade decided to do the necessary edits for the code is essentially sadly unusable in its current form due to the speed degradation two mwes showing a large difference i will note that i m using two slightly different environments with julia and julia time to complete roughly constraint eta for completing this as reported by progressmeter minutes roughly constraint the problem in question solved was much much larger and would take several days to formulate at this pace i d also be happy to contribute to a fix but i d need to learn a bit more about profiling and performance in julia and definitely more about the current jump implementation in order to help out anyways thank you so much for this project it s been super useful | 1 |
186,056 | 6,733,153,079 | IssuesEvent | 2017-10-18 14:00:28 | dwyl/library | https://api.github.com/repos/dwyl/library | opened | Create button component to do different things depending on state | enhancement priority-2 | We have a button on the book component, this button should do different things depending on the state of the book/application:
+ If the book is in the library and is not on loan then it should say "Borrow book"
+ If it is in the library but is on loan it should say "Join queue"
+ If it is in the library and is on loan to the user viewing it it should say "Return"
+ If it is not in the library and the user is not an admin it should say "Request book"
+ If it is not in the library and the user is an admin it should say "Add book"
The buttons should of course have relevant actions which fulfil the actions represented on them in text. | 1.0 | Create button component to do different things depending on state - We have a button on the book component, this button should do different things depending on the state of the book/application:
+ If the book is in the library and is not on loan then it should say "Borrow book"
+ If it is in the library but is on loan it should say "Join queue"
+ If it is in the library and is on loan to the user viewing it it should say "Return"
+ If it is not in the library and the user is not an admin it should say "Request book"
+ If it is not in the library and the user is an admin it should say "Add book"
The buttons should of course have relevant actions which fulfil the actions represented on them in text. | non_perf | create button component to do different things depending on state we have a button on the book component this button should do different things depending on the state of the book application if the book is in the library and is not on loan then it should say borrow book if it is in the library but is on loan it should say join queue if it is in the library and is on loan to the user viewing it it should say return if it is not in the library and the user is not an admin it should say request book if it is not in the library and the user is an admin it should say add book the buttons should of course have relevant actions which fulfil the actions represented on them in text | 0 |
6,377 | 3,796,715,523 | IssuesEvent | 2016-03-23 02:20:58 | mapbox/mapbox-gl-native | https://api.github.com/repos/mapbox/mapbox-gl-native | closed | remove build-time dependency on llvm.org | build | This host has poor reliability, and when it goes down our builds fail. | 1.0 | remove build-time dependency on llvm.org - This host has poor reliability, and when it goes down our builds fail. | non_perf | remove build time dependency on llvm org this host has poor reliability and when it goes down our builds fail | 0 |
300,756 | 25,993,454,863 | IssuesEvent | 2022-12-20 09:38:06 | finos/waltz | https://api.github.com/repos/finos/waltz | closed | Prep 1.46 | fixed (test & close) task | ### Description
- [x] mvn versions
- [x] pacakge.json version
- [x] prep liquibase version
### Resourcing
We intend to implement this task | 1.0 | Prep 1.46 - ### Description
- [x] mvn versions
- [x] pacakge.json version
- [x] prep liquibase version
### Resourcing
We intend to implement this task | non_perf | prep description mvn versions pacakge json version prep liquibase version resourcing we intend to implement this task | 0 |
382,010 | 26,481,759,612 | IssuesEvent | 2023-01-17 15:09:37 | StaticJsCMS/static-cms | https://api.github.com/repos/StaticJsCMS/static-cms | closed | Create NetlifyCMS to StaticCMS Migration Guide | documentation | Document all steps necessary to migrate from NetlifyCMS to StaticCMS, covering as many scenarios as possible.
- [ ] All deprecated features were removed
- [ ] Some widget properties/options have changed
- [ ] New markdown editor comes with new plugin interface, old editor components no longer work
- [ ] Widget control properties have changed
- [ ] List widget - Dropped support for the `field` property. A single field in the `fields` property does the same behavior.
- [ ] Dropping Moment for date-fns
- [ ] Passing a config to `.init()` will now completely override `config.yml` (they are not exclusive), instead of being merged with `config.yml`
- [ ] CMS must be explicitly initiated now.
- [ ] Dropped the following beta features from NetlifyCMS.
- GraphQL support for GitHub and GitLab
- Remark plugins (new markdown editor has its own plugin system)
- [ ] Dropped support for AssetStore integration
- [ ] Dropped support for Azure backend
- [ ] `registerWidget` props have changed
- [ ] `isValid` in widget control component replaced by `validator` function in `registerWidget` props
- [ ] `createClass` is deprecated, functional components should be used instead. Basic react hooks are provided as globals now.
- [ ] React 18.2.0 | 1.0 | Create NetlifyCMS to StaticCMS Migration Guide - Document all steps necessary to migrate from NetlifyCMS to StaticCMS, covering as many scenarios as possible.
- [ ] All deprecated features were removed
- [ ] Some widget properties/options have changed
- [ ] New markdown editor comes with new plugin interface, old editor components no longer work
- [ ] Widget control properties have changed
- [ ] List widget - Dropped support for the `field` property. A single field in the `fields` property does the same behavior.
- [ ] Dropping Moment for date-fns
- [ ] Passing a config to `.init()` will now completely override `config.yml` (they are not exclusive), instead of being merged with `config.yml`
- [ ] CMS must be explicitly initiated now.
- [ ] Dropped the following beta features from NetlifyCMS.
- GraphQL support for GitHub and GitLab
- Remark plugins (new markdown editor has its own plugin system)
- [ ] Dropped support for AssetStore integration
- [ ] Dropped support for Azure backend
- [ ] `registerWidget` props have changed
- [ ] `isValid` in widget control component replaced by `validator` function in `registerWidget` props
- [ ] `createClass` is deprecated, functional components should be used instead. Basic react hooks are provided as globals now.
- [ ] React 18.2.0 | non_perf | create netlifycms to staticcms migration guide document all steps necessary to migrate from netlifycms to staticcms covering as many scenarios as possible all deprecated features were removed some widget properties options have changed new markdown editor comes with new plugin interface old editor components no longer work widget control properties have changed list widget dropped support for the field property a single field in the fields property does the same behavior dropping moment for date fns passing a config to init will now completely override config yml they are not exclusive instead of being merged with config yml cms must be explicitly initiated now dropped the following beta features from netlifycms graphql support for github and gitlab remark plugins new markdown editor has its own plugin system dropped support for assetstore integration dropped support for azure backend registerwidget props have changed isvalid in widget control component replaced by validator function in registerwidget props createclass is deprecated functional components should be used instead basic react hooks are provided as globals now react | 0 |
142,042 | 19,024,488,951 | IssuesEvent | 2021-11-24 00:35:28 | udsm-dhis2-project/dim-api | https://api.github.com/repos/udsm-dhis2-project/dim-api | closed | CVE-2021-3795 (Medium) detected in semver-regex-2.0.0.tgz | security vulnerability no-issue-activity | ## CVE-2021-3795 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>semver-regex-2.0.0.tgz</b></p></summary>
<p>Regular expression for matching semver versions</p>
<p>Library home page: <a href="https://registry.npmjs.org/semver-regex/-/semver-regex-2.0.0.tgz">https://registry.npmjs.org/semver-regex/-/semver-regex-2.0.0.tgz</a></p>
<p>Path to dependency file: dim-api/package.json</p>
<p>Path to vulnerable library: dim-api/node_modules/semver-regex/package.json</p>
<p>
Dependency Hierarchy:
- husky-4.2.5.tgz (Root Library)
- find-versions-3.2.0.tgz
- :x: **semver-regex-2.0.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
semver-regex is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3795>CVE-2021-3795</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sindresorhus/semver-regex/releases/tag/v4.0.1">https://github.com/sindresorhus/semver-regex/releases/tag/v4.0.1</a></p>
<p>Release Date: 2021-09-15</p>
<p>Fix Resolution: semver-regex - 3.1.3,4.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3795 (Medium) detected in semver-regex-2.0.0.tgz - ## CVE-2021-3795 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>semver-regex-2.0.0.tgz</b></p></summary>
<p>Regular expression for matching semver versions</p>
<p>Library home page: <a href="https://registry.npmjs.org/semver-regex/-/semver-regex-2.0.0.tgz">https://registry.npmjs.org/semver-regex/-/semver-regex-2.0.0.tgz</a></p>
<p>Path to dependency file: dim-api/package.json</p>
<p>Path to vulnerable library: dim-api/node_modules/semver-regex/package.json</p>
<p>
Dependency Hierarchy:
- husky-4.2.5.tgz (Root Library)
- find-versions-3.2.0.tgz
- :x: **semver-regex-2.0.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
semver-regex is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3795>CVE-2021-3795</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sindresorhus/semver-regex/releases/tag/v4.0.1">https://github.com/sindresorhus/semver-regex/releases/tag/v4.0.1</a></p>
<p>Release Date: 2021-09-15</p>
<p>Fix Resolution: semver-regex - 3.1.3,4.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_perf | cve medium detected in semver regex tgz cve medium severity vulnerability vulnerable library semver regex tgz regular expression for matching semver versions library home page a href path to dependency file dim api package json path to vulnerable library dim api node modules semver regex package json dependency hierarchy husky tgz root library find versions tgz x semver regex tgz vulnerable library found in base branch master vulnerability details semver regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution semver regex step up your open source security game with whitesource | 0 |
28,902 | 13,884,128,745 | IssuesEvent | 2020-10-18 14:56:33 | pandas-dev/pandas | https://api.github.com/repos/pandas-dev/pandas | closed | BUG: GroupBy().fillna() performance regression | Groupby Missing-data Performance | - [x] I have checked that this issue has not already been reported.
- [x] I have confirmed this bug exists on the latest version of pandas.
- [ ] (optional) I have confirmed this bug exists on the master branch of pandas.
---
```python
import pandas as pd
import numpy as np
N = 2000
df = pd.DataFrame({"A": [1] * N, "B": [np.nan, 1.0] * (N // 2)})
df = df.sort_values("A").set_index("A")
df["B"] = df.groupby("A")["B"].fillna(method="ffill")
```
#### Problem description
The groupby + fillna gets extremely slow increasing the N.
This is a regression from 1.0.5->1.1.0.
Note: if I remove the `.set_index("A")` it's fast again.
#### Expected Output
Same output, just faster.
#### Output of ``pd.show_versions()``
<details>
INSTALLED VERSIONS
------------------
commit : d9fff2792bf16178d4e450fe7384244e50635733
python : 3.7.8.final.0
python-bits : 64
OS : Linux
OS-release : 4.4.110-1.el7.elrepo.x86_64
Version : #1 SMP Fri Jan 5 11:35:48 EST 2018
machine : x86_64
processor : x86_64
byteorder : little
LC_ALL : None
LANG : en_US.UTF-8
LOCALE : en_US.UTF-8
pandas : 1.1.0
numpy : 1.19.1
pytz : 2020.1
dateutil : 2.8.1
pip : 20.2.3
setuptools : 49.6.0.post20200917
Cython : None
pytest : None
hypothesis : None
sphinx : None
blosc : None
feather : None
xlsxwriter : None
lxml.etree : None
html5lib : None
pymysql : None
psycopg2 : None
jinja2 : None
IPython : None
pandas_datareader: None
bs4 : None
bottleneck : None
fsspec : None
fastparquet : None
gcsfs : None
matplotlib : None
numexpr : None
odfpy : None
openpyxl : None
pandas_gbq : None
pyarrow : None
pytables : None
pyxlsb : None
s3fs : None
scipy : None
sqlalchemy : None
tables : None
tabulate : None
xarray : None
xlrd : None
xlwt : None
numba : None
</details>
| True | BUG: GroupBy().fillna() performance regression - - [x] I have checked that this issue has not already been reported.
- [x] I have confirmed this bug exists on the latest version of pandas.
- [ ] (optional) I have confirmed this bug exists on the master branch of pandas.
---
```python
import pandas as pd
import numpy as np
N = 2000
df = pd.DataFrame({"A": [1] * N, "B": [np.nan, 1.0] * (N // 2)})
df = df.sort_values("A").set_index("A")
df["B"] = df.groupby("A")["B"].fillna(method="ffill")
```
#### Problem description
The groupby + fillna gets extremely slow increasing the N.
This is a regression from 1.0.5->1.1.0.
Note: if I remove the `.set_index("A")` it's fast again.
#### Expected Output
Same output, just faster.
#### Output of ``pd.show_versions()``
<details>
INSTALLED VERSIONS
------------------
commit : d9fff2792bf16178d4e450fe7384244e50635733
python : 3.7.8.final.0
python-bits : 64
OS : Linux
OS-release : 4.4.110-1.el7.elrepo.x86_64
Version : #1 SMP Fri Jan 5 11:35:48 EST 2018
machine : x86_64
processor : x86_64
byteorder : little
LC_ALL : None
LANG : en_US.UTF-8
LOCALE : en_US.UTF-8
pandas : 1.1.0
numpy : 1.19.1
pytz : 2020.1
dateutil : 2.8.1
pip : 20.2.3
setuptools : 49.6.0.post20200917
Cython : None
pytest : None
hypothesis : None
sphinx : None
blosc : None
feather : None
xlsxwriter : None
lxml.etree : None
html5lib : None
pymysql : None
psycopg2 : None
jinja2 : None
IPython : None
pandas_datareader: None
bs4 : None
bottleneck : None
fsspec : None
fastparquet : None
gcsfs : None
matplotlib : None
numexpr : None
odfpy : None
openpyxl : None
pandas_gbq : None
pyarrow : None
pytables : None
pyxlsb : None
s3fs : None
scipy : None
sqlalchemy : None
tables : None
tabulate : None
xarray : None
xlrd : None
xlwt : None
numba : None
</details>
| perf | bug groupby fillna performance regression i have checked that this issue has not already been reported i have confirmed this bug exists on the latest version of pandas optional i have confirmed this bug exists on the master branch of pandas python import pandas as pd import numpy as np n df pd dataframe a n b n df df sort values a set index a df df groupby a fillna method ffill problem description the groupby fillna gets extremely slow increasing the n this is a regression from note if i remove the set index a it s fast again expected output same output just faster output of pd show versions installed versions commit python final python bits os linux os release elrepo version smp fri jan est machine processor byteorder little lc all none lang en us utf locale en us utf pandas numpy pytz dateutil pip setuptools cython none pytest none hypothesis none sphinx none blosc none feather none xlsxwriter none lxml etree none none pymysql none none none ipython none pandas datareader none none bottleneck none fsspec none fastparquet none gcsfs none matplotlib none numexpr none odfpy none openpyxl none pandas gbq none pyarrow none pytables none pyxlsb none none scipy none sqlalchemy none tables none tabulate none xarray none xlrd none xlwt none numba none | 1 |
19,040 | 10,300,096,313 | IssuesEvent | 2019-08-28 12:09:07 | ArkEcosystem/desktop-wallet | https://api.github.com/repos/ArkEcosystem/desktop-wallet | closed | Change Peer function slow compared to Mobile Wallet | Status: Stale Type: Performance | When using the desktop wallet, I have noticed that change peer (AKA "connect to best") is slow or doesn't do anything when clicked. My peer latencies on desktop wallet are also often in the red. The mobile wallet always does changes peers for me lightning fast and without an issue (and 99% of the time, the mobile wallet stays connected as intended without the need to manually change).
Also, specifying a custom peer, like when I actually type in the peer I see in the mobile wallet into the desktop wallet, returns an error of not being able to connect on the desktop wallet. So I don't understand why my desktop wallet is having so much trouble with peers compared to the mobile wallet | True | Change Peer function slow compared to Mobile Wallet - When using the desktop wallet, I have noticed that change peer (AKA "connect to best") is slow or doesn't do anything when clicked. My peer latencies on desktop wallet are also often in the red. The mobile wallet always does changes peers for me lightning fast and without an issue (and 99% of the time, the mobile wallet stays connected as intended without the need to manually change).
Also, specifying a custom peer, like when I actually type in the peer I see in the mobile wallet into the desktop wallet, returns an error of not being able to connect on the desktop wallet. So I don't understand why my desktop wallet is having so much trouble with peers compared to the mobile wallet | perf | change peer function slow compared to mobile wallet when using the desktop wallet i have noticed that change peer aka connect to best is slow or doesn t do anything when clicked my peer latencies on desktop wallet are also often in the red the mobile wallet always does changes peers for me lightning fast and without an issue and of the time the mobile wallet stays connected as intended without the need to manually change also specifying a custom peer like when i actually type in the peer i see in the mobile wallet into the desktop wallet returns an error of not being able to connect on the desktop wallet so i don t understand why my desktop wallet is having so much trouble with peers compared to the mobile wallet | 1 |
102,849 | 16,590,754,951 | IssuesEvent | 2021-06-01 07:25:01 | Yoavmartin/vulnerable-node | https://api.github.com/repos/Yoavmartin/vulnerable-node | closed | CVE-2020-8116 (High) detected in opennmsopennms-source-24.1.2-1 - autoclosed | security vulnerability | ## CVE-2020-8116 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-24.1.2-1</b></p></summary>
<p>
<p>A Java based fault and performance management system</p>
<p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Yoavmartin/vulnerable-node/commit/c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c">c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in dot-prop npm package versions before 4.2.1 and versions 5.x before 5.1.1 allows an attacker to add arbitrary properties to JavaScript language constructs such as objects.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: dot-prop - 5.1.1</p>
</p>
</details>
<p></p>
| True | CVE-2020-8116 (High) detected in opennmsopennms-source-24.1.2-1 - autoclosed - ## CVE-2020-8116 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-24.1.2-1</b></p></summary>
<p>
<p>A Java based fault and performance management system</p>
<p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Yoavmartin/vulnerable-node/commit/c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c">c4e9b3e5b51e8ecff2ba523f2ee47e2f3143734c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in dot-prop npm package versions before 4.2.1 and versions 5.x before 5.1.1 allows an attacker to add arbitrary properties to JavaScript language constructs such as objects.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: dot-prop - 5.1.1</p>
</p>
</details>
<p></p>
| non_perf | cve high detected in opennmsopennms source autoclosed cve high severity vulnerability vulnerable library opennmsopennms source a java based fault and performance management system library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details prototype pollution vulnerability in dot prop npm package versions before and versions x before allows an attacker to add arbitrary properties to javascript language constructs such as objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution dot prop | 0 |
308,637 | 9,441,338,708 | IssuesEvent | 2019-04-15 00:55:45 | kubeflow/kubeflow | https://api.github.com/repos/kubeflow/kubeflow | closed | Solution for building Docker images from Jupyter | area/jupyter help wanted priority/p2 | We need a solution for building docker images from Jupyter. I think a common use case will be developing/prototyping code in JupyterHub and then wanting to package that code up in a Docker image and either run a job (e.g. a TF training job) or deploy a model (e.g. via Seldon).
There are lots of ways to build Docker images
- bazel
- docker build
- GCB
- [Repo2Docker](https://github.com/jupyter/repo2docker) part of JupyterHub's binder project
- Argo
What's the difference between Repo2Docker and using Argo to pull the code from Git and then run a build? | 1.0 | Solution for building Docker images from Jupyter - We need a solution for building docker images from Jupyter. I think a common use case will be developing/prototyping code in JupyterHub and then wanting to package that code up in a Docker image and either run a job (e.g. a TF training job) or deploy a model (e.g. via Seldon).
There are lots of ways to build Docker images
- bazel
- docker build
- GCB
- [Repo2Docker](https://github.com/jupyter/repo2docker) part of JupyterHub's binder project
- Argo
What's the difference between Repo2Docker and using Argo to pull the code from Git and then run a build? | non_perf | solution for building docker images from jupyter we need a solution for building docker images from jupyter i think a common use case will be developing prototyping code in jupyterhub and then wanting to package that code up in a docker image and either run a job e g a tf training job or deploy a model e g via seldon there are lots of ways to build docker images bazel docker build gcb part of jupyterhub s binder project argo what s the difference between and using argo to pull the code from git and then run a build | 0 |
1,737 | 3,089,681,432 | IssuesEvent | 2015-08-25 22:58:38 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Why does the speed of kubelet running pause container have slowed down? | area/performance kind/support team/node | Today I raise a 2 minions cluster and create a 50 replica redis-controller-master rc, the rc take about 10 minutes to launch, that's much longer than one week ago ( about 50 seconds).
I see the kubelet's log and found it take 10 seconds to pulled a pause images and another 10 seconds to start it:

And after all the pause containers has started, it only takes several seconds to started all the redis containers:

I thought the behavior of starting redis containers is reasonable and that about pause container is unreasonable, why it? | True | Why does the speed of kubelet running pause container have slowed down? - Today I raise a 2 minions cluster and create a 50 replica redis-controller-master rc, the rc take about 10 minutes to launch, that's much longer than one week ago ( about 50 seconds).
I see the kubelet's log and found it take 10 seconds to pulled a pause images and another 10 seconds to start it:

And after all the pause containers has started, it only takes several seconds to started all the redis containers:

I thought the behavior of starting redis containers is reasonable and that about pause container is unreasonable, why it? | perf | why does the speed of kubelet running pause container have slowed down today i raise a minions cluster and create a replica redis controller master rc the rc take about minutes to launch that s much longer than one week ago about seconds i see the kubelet s log and found it take seconds to pulled a pause images and another seconds to start it and after all the pause containers has started it only takes several seconds to started all the redis containers i thought the behavior of starting redis containers is reasonable and that about pause container is unreasonable why it | 1 |
7,835 | 6,260,378,918 | IssuesEvent | 2017-07-14 20:30:19 | KazDragon/terminalpp | https://api.github.com/repos/KazDragon/terminalpp | closed | Add forwarding references to lambda visitor | Improvement in progress Performance | Althought it's supposed to take lambdas, it might also take moveable function objects, so adding forwarding references makes sense. | True | Add forwarding references to lambda visitor - Althought it's supposed to take lambdas, it might also take moveable function objects, so adding forwarding references makes sense. | perf | add forwarding references to lambda visitor althought it s supposed to take lambdas it might also take moveable function objects so adding forwarding references makes sense | 1 |
279,727 | 30,733,971,743 | IssuesEvent | 2023-07-28 05:49:46 | DevSeconsTL/WebGoat-WhiteSource-Bolt | https://api.github.com/repos/DevSeconsTL/WebGoat-WhiteSource-Bolt | opened | commons-io-1.3.2.jar: 1 vulnerabilities (highest severity is: 4.8) | Mend: dependency security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-io-1.3.2.jar</b></p></summary>
<p>Commons-IO contains utility classes, stream implementations, file filters, and endian classes.</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-io/commons-io/1.3.2/commons-io-1.3.2.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/DevSeconsTL/WebGoat-WhiteSource-Bolt/commit/a4b4e21bf0133636e16df71363955c6ce17b0791">a4b4e21bf0133636e16df71363955c6ce17b0791</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (commons-io version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-29425](https://www.mend.io/vulnerability-database/CVE-2021-29425) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 4.8 | commons-io-1.3.2.jar | Direct | 2.7 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-29425</summary>
### Vulnerable Library - <b>commons-io-1.3.2.jar</b></p>
<p>Commons-IO contains utility classes, stream implementations, file filters, and endian classes.</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-io/commons-io/1.3.2/commons-io-1.3.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **commons-io-1.3.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DevSeconsTL/WebGoat-WhiteSource-Bolt/commit/a4b4e21bf0133636e16df71363955c6ce17b0791">a4b4e21bf0133636e16df71363955c6ce17b0791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Apache Commons IO before 2.7, When invoking the method FileNameUtils.normalize with an improper input string, like "//../foo", or "\\..\foo", the result would be the same value, thus possibly providing access to files in the parent directory, but not further above (thus "limited" path traversal), if the calling code would use the result to construct a path value.
<p>Publish Date: 2021-04-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29425>CVE-2021-29425</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-29425">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-29425</a></p>
<p>Release Date: 2021-04-13</p>
<p>Fix Resolution: 2.7</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | commons-io-1.3.2.jar: 1 vulnerabilities (highest severity is: 4.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-io-1.3.2.jar</b></p></summary>
<p>Commons-IO contains utility classes, stream implementations, file filters, and endian classes.</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-io/commons-io/1.3.2/commons-io-1.3.2.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/DevSeconsTL/WebGoat-WhiteSource-Bolt/commit/a4b4e21bf0133636e16df71363955c6ce17b0791">a4b4e21bf0133636e16df71363955c6ce17b0791</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (commons-io version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-29425](https://www.mend.io/vulnerability-database/CVE-2021-29425) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 4.8 | commons-io-1.3.2.jar | Direct | 2.7 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-29425</summary>
### Vulnerable Library - <b>commons-io-1.3.2.jar</b></p>
<p>Commons-IO contains utility classes, stream implementations, file filters, and endian classes.</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-io/commons-io/1.3.2/commons-io-1.3.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **commons-io-1.3.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DevSeconsTL/WebGoat-WhiteSource-Bolt/commit/a4b4e21bf0133636e16df71363955c6ce17b0791">a4b4e21bf0133636e16df71363955c6ce17b0791</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Apache Commons IO before 2.7, When invoking the method FileNameUtils.normalize with an improper input string, like "//../foo", or "\\..\foo", the result would be the same value, thus possibly providing access to files in the parent directory, but not further above (thus "limited" path traversal), if the calling code would use the result to construct a path value.
<p>Publish Date: 2021-04-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29425>CVE-2021-29425</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-29425">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-29425</a></p>
<p>Release Date: 2021-04-13</p>
<p>Fix Resolution: 2.7</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_perf | commons io jar vulnerabilities highest severity is vulnerable library commons io jar commons io contains utility classes stream implementations file filters and endian classes path to dependency file pom xml path to vulnerable library home wss scanner repository commons io commons io commons io jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in commons io version remediation available medium commons io jar direct details cve vulnerable library commons io jar commons io contains utility classes stream implementations file filters and endian classes path to dependency file pom xml path to vulnerable library home wss scanner repository commons io commons io commons io jar dependency hierarchy x commons io jar vulnerable library found in head commit a href found in base branch master vulnerability details in apache commons io before when invoking the method filenameutils normalize with an improper input string like foo or foo the result would be the same value thus possibly providing access to files in the parent directory but not further above thus limited path traversal if the calling code would use the result to construct a path value publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
25,951 | 7,763,198,148 | IssuesEvent | 2018-06-01 15:46:20 | apache/incubator-openwhisk-release | https://api.github.com/repos/apache/incubator-openwhisk-release | opened | Remove wskdeploy specification PDFs (archives) and diagram from source release | Release build priority-low | The following folders and their contents for the wskdeploy repo. should be excluded from the TAR file produced for the source code release:
- specification/archive
- specification/diagrams
This would mean (binary) files of the following format/extensions would be excluded:
- .pdf
- .pptx
- .jpg
| 1.0 | Remove wskdeploy specification PDFs (archives) and diagram from source release - The following folders and their contents for the wskdeploy repo. should be excluded from the TAR file produced for the source code release:
- specification/archive
- specification/diagrams
This would mean (binary) files of the following format/extensions would be excluded:
- .pdf
- .pptx
- .jpg
| non_perf | remove wskdeploy specification pdfs archives and diagram from source release the following folders and their contents for the wskdeploy repo should be excluded from the tar file produced for the source code release specification archive specification diagrams this would mean binary files of the following format extensions would be excluded pdf pptx jpg | 0 |
11,185 | 7,460,943,233 | IssuesEvent | 2018-03-30 22:14:10 | zfsonlinux/zfs | https://api.github.com/repos/zfsonlinux/zfs | closed | Large synchronous writes are slow when a slog is present | Performance | Note that this issue seems to impact all ZFS implementations, not just ZFS On Linux.
ZFS uses a complicated process when it comes to deciding whether a write should be logged in indirect mode (written once by the DMU, the log records store a pointer) or in immediate mode (written in the log record, rewritten later by the DMU). Basically, it goes like this:
- Write in indirect mode to the data vdevs if:
- `logbias=throughput`, or
- There is no slog and the write is larger than `zfs_immediate_write_sz`.
- Write in immediate mode to the data vdevs if `logbias=latency` and:
- There is no slog and the write is smaller than `zfs_immediate_write_sz`, or
- There is a slog and the total commit size if larger than `zil_slog_limit`.
- Write in immediate mode to the slog vdevs if `logbias=latency`, there is a slog, and the total commit size is smaller than `zil_slog_limit`.
The decision to use indirect or immediate mode is implemented in `zfs_log_write()` and `zvol_log_write()`. The decision to use the slog or the normal vdevs is implemented in the `USE_SLOG()` macro used by `zil_lwb_write_start`.
The issue is, this decision process makes sense except for one particularly painful edge case, when these conditions are all true:
- `logbias=latency`, and
- There is a slog, and
- There are large writes in the ZIL to be commited (e.g. > 100 MB).
In this situation, the optimal choice would be to write to the normal pool in indirect mode, which should give us the minimum latency considering this is a large sequential write. Indeed, for very large writes, you don't want to use immediate mode because it means writing the data twice. Even if you write the log records to the slog, this will be slower with most pool configurations with e.g. lots of spindles and one SSD slog because the aggregate sequential write throughput of all the spindles is usually greater than the SSD's.
Instead, the algorithm makes the worst decision possible: it writes the data in immediate mode to the main data disks. This means that all the (large) data will be commited as ZIL log records on the data disks first, then immediately after, it will get written _again_ by the DMU. This means the overall throughput is halved, and if this is a sustained load, the ZIL commit latency will be doubled compared to indirect mode.
It is shockingly easy to reproduce this issue. In pseudo-code:
```
open(file)
write(file, lots of data) // e.g. 2 GB
fsync(file)
```
Watch the `zil_stats` kstat page when that runs.
If you don't have a slog in your pool, then the `fsync()` call will complete in roughly the time it takes to write 2 GB sequentially to your main disks. This is optimal.
If you have a slog in your pool, then the `fsync()` call will generate twice as much write activity, and will write up to 4 GB to your main disks. Ironically, the slog won't be used at all when that happens.
The solution would be to modify the algorithm `zfs_log_write()` and `zvol_log_write()` so that, in the conditions mentioned above, it switches to indirect writes when the commit size reaches a certain threshold (e.g. 32 MB).
I would gladly write a patch, but I won't have the time to do it, so I'm just leaving the result of my research here in case anyone's interested. If anyone wants to write the patch, it should be very simple to implement it.
| True | Large synchronous writes are slow when a slog is present - Note that this issue seems to impact all ZFS implementations, not just ZFS On Linux.
ZFS uses a complicated process when it comes to deciding whether a write should be logged in indirect mode (written once by the DMU, the log records store a pointer) or in immediate mode (written in the log record, rewritten later by the DMU). Basically, it goes like this:
- Write in indirect mode to the data vdevs if:
- `logbias=throughput`, or
- There is no slog and the write is larger than `zfs_immediate_write_sz`.
- Write in immediate mode to the data vdevs if `logbias=latency` and:
- There is no slog and the write is smaller than `zfs_immediate_write_sz`, or
- There is a slog and the total commit size if larger than `zil_slog_limit`.
- Write in immediate mode to the slog vdevs if `logbias=latency`, there is a slog, and the total commit size is smaller than `zil_slog_limit`.
The decision to use indirect or immediate mode is implemented in `zfs_log_write()` and `zvol_log_write()`. The decision to use the slog or the normal vdevs is implemented in the `USE_SLOG()` macro used by `zil_lwb_write_start`.
The issue is, this decision process makes sense except for one particularly painful edge case, when these conditions are all true:
- `logbias=latency`, and
- There is a slog, and
- There are large writes in the ZIL to be commited (e.g. > 100 MB).
In this situation, the optimal choice would be to write to the normal pool in indirect mode, which should give us the minimum latency considering this is a large sequential write. Indeed, for very large writes, you don't want to use immediate mode because it means writing the data twice. Even if you write the log records to the slog, this will be slower with most pool configurations with e.g. lots of spindles and one SSD slog because the aggregate sequential write throughput of all the spindles is usually greater than the SSD's.
Instead, the algorithm makes the worst decision possible: it writes the data in immediate mode to the main data disks. This means that all the (large) data will be commited as ZIL log records on the data disks first, then immediately after, it will get written _again_ by the DMU. This means the overall throughput is halved, and if this is a sustained load, the ZIL commit latency will be doubled compared to indirect mode.
It is shockingly easy to reproduce this issue. In pseudo-code:
```
open(file)
write(file, lots of data) // e.g. 2 GB
fsync(file)
```
Watch the `zil_stats` kstat page when that runs.
If you don't have a slog in your pool, then the `fsync()` call will complete in roughly the time it takes to write 2 GB sequentially to your main disks. This is optimal.
If you have a slog in your pool, then the `fsync()` call will generate twice as much write activity, and will write up to 4 GB to your main disks. Ironically, the slog won't be used at all when that happens.
The solution would be to modify the algorithm `zfs_log_write()` and `zvol_log_write()` so that, in the conditions mentioned above, it switches to indirect writes when the commit size reaches a certain threshold (e.g. 32 MB).
I would gladly write a patch, but I won't have the time to do it, so I'm just leaving the result of my research here in case anyone's interested. If anyone wants to write the patch, it should be very simple to implement it.
| perf | large synchronous writes are slow when a slog is present note that this issue seems to impact all zfs implementations not just zfs on linux zfs uses a complicated process when it comes to deciding whether a write should be logged in indirect mode written once by the dmu the log records store a pointer or in immediate mode written in the log record rewritten later by the dmu basically it goes like this write in indirect mode to the data vdevs if logbias throughput or there is no slog and the write is larger than zfs immediate write sz write in immediate mode to the data vdevs if logbias latency and there is no slog and the write is smaller than zfs immediate write sz or there is a slog and the total commit size if larger than zil slog limit write in immediate mode to the slog vdevs if logbias latency there is a slog and the total commit size is smaller than zil slog limit the decision to use indirect or immediate mode is implemented in zfs log write and zvol log write the decision to use the slog or the normal vdevs is implemented in the use slog macro used by zil lwb write start the issue is this decision process makes sense except for one particularly painful edge case when these conditions are all true logbias latency and there is a slog and there are large writes in the zil to be commited e g mb in this situation the optimal choice would be to write to the normal pool in indirect mode which should give us the minimum latency considering this is a large sequential write indeed for very large writes you don t want to use immediate mode because it means writing the data twice even if you write the log records to the slog this will be slower with most pool configurations with e g lots of spindles and one ssd slog because the aggregate sequential write throughput of all the spindles is usually greater than the ssd s instead the algorithm makes the worst decision possible it writes the data in immediate mode to the main data disks this means that all the large data will be commited as zil log records on the data disks first then immediately after it will get written again by the dmu this means the overall throughput is halved and if this is a sustained load the zil commit latency will be doubled compared to indirect mode it is shockingly easy to reproduce this issue in pseudo code open file write file lots of data e g gb fsync file watch the zil stats kstat page when that runs if you don t have a slog in your pool then the fsync call will complete in roughly the time it takes to write gb sequentially to your main disks this is optimal if you have a slog in your pool then the fsync call will generate twice as much write activity and will write up to gb to your main disks ironically the slog won t be used at all when that happens the solution would be to modify the algorithm zfs log write and zvol log write so that in the conditions mentioned above it switches to indirect writes when the commit size reaches a certain threshold e g mb i would gladly write a patch but i won t have the time to do it so i m just leaving the result of my research here in case anyone s interested if anyone wants to write the patch it should be very simple to implement it | 1 |
222,818 | 24,711,335,768 | IssuesEvent | 2022-10-20 01:14:34 | Lob2018/YannLobjois_6_25032021 | https://api.github.com/repos/Lob2018/YannLobjois_6_25032021 | opened | CVE-2022-37603 (High) detected in loader-utils-2.0.0.tgz | security vulnerability | ## CVE-2022-37603 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-2.0.0.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/css-loader/node_modules/loader-utils/package.json,/node_modules/style-loader/node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- style-loader-2.0.0.tgz (Root Library)
- :x: **loader-utils-2.0.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js.
<p>Publish Date: 2022-10-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37603>CVE-2022-37603</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-14</p>
<p>Fix Resolution (loader-utils): 2.0.1</p>
<p>Direct dependency fix Resolution (style-loader): 3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-37603 (High) detected in loader-utils-2.0.0.tgz - ## CVE-2022-37603 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-2.0.0.tgz</b></p></summary>
<p>utils for webpack loaders</p>
<p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/css-loader/node_modules/loader-utils/package.json,/node_modules/style-loader/node_modules/loader-utils/package.json</p>
<p>
Dependency Hierarchy:
- style-loader-2.0.0.tgz (Root Library)
- :x: **loader-utils-2.0.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js.
<p>Publish Date: 2022-10-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37603>CVE-2022-37603</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-14</p>
<p>Fix Resolution (loader-utils): 2.0.1</p>
<p>Direct dependency fix Resolution (style-loader): 3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_perf | cve high detected in loader utils tgz cve high severity vulnerability vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules css loader node modules loader utils package json node modules style loader node modules loader utils package json dependency hierarchy style loader tgz root library x loader utils tgz vulnerable library found in base branch main vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the url variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution loader utils direct dependency fix resolution style loader step up your open source security game with mend | 0 |
1,297 | 2,931,188,117 | IssuesEvent | 2015-06-29 10:42:00 | mapbox/mapbox-gl-native | https://api.github.com/repos/mapbox/mapbox-gl-native | closed | All polylines flicker when adding adding/removing a polyline (iOS) | bug iOS performance | I see an old issue (#1712) referencing flickering of annotations that's marked as closed for 0.4, but the issue still exists.
Steps to reproduce:
- Add a number of polylines to a map when it's first made
- Add a new polyline or remove an existing one triggered by a user action
http://f.cl.ly/items/0w1e2K2w3N2q3Y2K0V1x/Flicker.m4v | True | All polylines flicker when adding adding/removing a polyline (iOS) - I see an old issue (#1712) referencing flickering of annotations that's marked as closed for 0.4, but the issue still exists.
Steps to reproduce:
- Add a number of polylines to a map when it's first made
- Add a new polyline or remove an existing one triggered by a user action
http://f.cl.ly/items/0w1e2K2w3N2q3Y2K0V1x/Flicker.m4v | perf | all polylines flicker when adding adding removing a polyline ios i see an old issue referencing flickering of annotations that s marked as closed for but the issue still exists steps to reproduce add a number of polylines to a map when it s first made add a new polyline or remove an existing one triggered by a user action | 1 |
386,387 | 11,437,879,599 | IssuesEvent | 2020-02-05 01:25:37 | zulip/zulip | https://api.github.com/repos/zulip/zulip | opened | Follow-ups to addition of new stream_post_policy feature | area: stream settings priority: blocker | Copied from https://github.com/zulip/zulip/pull/13768#issuecomment-582196283:
There's some follow-ups we should definitely do:
* [ ] Once this is deployed to chat.zulip.org (very soon), test to make sure the mobile app can still create streams and still provides a good error message posting to streams with posting restrictions.
* [ ] I think it'd be helpful for `stream_types.hbs` to have a link to the /help/ documentation for what a "full member" is, since it's not clear from context. This can use the same styling we use in the "settings" pages that define these users.
* [ ] The real-time synchronization for changes to this field does not propagate to other browsers' display. @pragatiagrawal31 can you take this part since you're actively working on that component?
* [ ] The `test_messages` testing approach cleanups I suggested to use `do_change_` functions rather than raw `stream.save()` etc. (did you see them? I'm concerned GitHub may have lost my comment).
* [ ] Investigate further improving our migrations as described here: https://github.com/zulip/zulip/pull/13768/files/3bec321bf9e32751ddd62c18763c4b6cb00f107e#diff-52fee2c5eed6104b21eefefd0d2ef76e
* [ ] Reworking the `create_stream` and `ajaxSubscribeForCreation` interface as we discussed above to not involve N positional arguments?
| 1.0 | Follow-ups to addition of new stream_post_policy feature - Copied from https://github.com/zulip/zulip/pull/13768#issuecomment-582196283:
There's some follow-ups we should definitely do:
* [ ] Once this is deployed to chat.zulip.org (very soon), test to make sure the mobile app can still create streams and still provides a good error message posting to streams with posting restrictions.
* [ ] I think it'd be helpful for `stream_types.hbs` to have a link to the /help/ documentation for what a "full member" is, since it's not clear from context. This can use the same styling we use in the "settings" pages that define these users.
* [ ] The real-time synchronization for changes to this field does not propagate to other browsers' display. @pragatiagrawal31 can you take this part since you're actively working on that component?
* [ ] The `test_messages` testing approach cleanups I suggested to use `do_change_` functions rather than raw `stream.save()` etc. (did you see them? I'm concerned GitHub may have lost my comment).
* [ ] Investigate further improving our migrations as described here: https://github.com/zulip/zulip/pull/13768/files/3bec321bf9e32751ddd62c18763c4b6cb00f107e#diff-52fee2c5eed6104b21eefefd0d2ef76e
* [ ] Reworking the `create_stream` and `ajaxSubscribeForCreation` interface as we discussed above to not involve N positional arguments?
| non_perf | follow ups to addition of new stream post policy feature copied from there s some follow ups we should definitely do once this is deployed to chat zulip org very soon test to make sure the mobile app can still create streams and still provides a good error message posting to streams with posting restrictions i think it d be helpful for stream types hbs to have a link to the help documentation for what a full member is since it s not clear from context this can use the same styling we use in the settings pages that define these users the real time synchronization for changes to this field does not propagate to other browsers display can you take this part since you re actively working on that component the test messages testing approach cleanups i suggested to use do change functions rather than raw stream save etc did you see them i m concerned github may have lost my comment investigate further improving our migrations as described here reworking the create stream and ajaxsubscribeforcreation interface as we discussed above to not involve n positional arguments | 0 |
45,507 | 24,084,716,302 | IssuesEvent | 2022-09-19 09:52:14 | iterative/dvc | https://api.github.com/repos/iterative/dvc | closed | regression: dvc init is slow | p2-medium performance regression | The 3 major issue seems to be:
- [x] `tpi` (see https://github.com/iterative/tpi/issues/14) (~300ms)
- [ ] `fsspec` callback (~200ms)
- [x] prompts import on experiments from `rich` (for `dvc exp init`) (~50ms) #6738
- [ ] [`dvc.fs` importing fsspec by default](https://github.com/iterative/dvc/blob/612696cf4ae01885a89269b5d26b4631380614b8/dvc/fs/azure.py#L5) (~100ms)
(There may not be much point in making fsspec lazy, as we'll need local filesystem anyway).
The following is a startup log for `dvc init`.
<img width="1107" alt="Screen Shot 2021-09-29 at 17 52 16" src="https://user-images.githubusercontent.com/18718008/135265276-7f9279b5-8610-497a-8f59-654a7a859870.png">
Following is the benchmark for `dvc init`.
```console
$ hyperfine "dvc init" --warmup 3 -p "rm -rf .dvc"
Benchmark #1: dvc init
Time (mean ± σ): 1.311 s ± 0.065 s [User: 1.066 s, System: 0.203 s]
Range (min … max): 1.230 s … 1.435 s 10 runs
```
`dvc init` crossing 1 sec is really bad, this was the result a month ago (before tpi and prompts, and with fsspec.Callback made lazy):
```console
$ hyperfine "dvc init" --warmup 3 -p "rm -rf .dvc"
Benchmark #1: dvc init
Time (mean ± σ): 506.8 ms ± 31.4 ms [User: 369.2 ms, System: 103.7 ms]
Range (min … max): 477.4 ms … 566.6 ms 10 runs
```
| True | regression: dvc init is slow - The 3 major issue seems to be:
- [x] `tpi` (see https://github.com/iterative/tpi/issues/14) (~300ms)
- [ ] `fsspec` callback (~200ms)
- [x] prompts import on experiments from `rich` (for `dvc exp init`) (~50ms) #6738
- [ ] [`dvc.fs` importing fsspec by default](https://github.com/iterative/dvc/blob/612696cf4ae01885a89269b5d26b4631380614b8/dvc/fs/azure.py#L5) (~100ms)
(There may not be much point in making fsspec lazy, as we'll need local filesystem anyway).
The following is a startup log for `dvc init`.
<img width="1107" alt="Screen Shot 2021-09-29 at 17 52 16" src="https://user-images.githubusercontent.com/18718008/135265276-7f9279b5-8610-497a-8f59-654a7a859870.png">
Following is the benchmark for `dvc init`.
```console
$ hyperfine "dvc init" --warmup 3 -p "rm -rf .dvc"
Benchmark #1: dvc init
Time (mean ± σ): 1.311 s ± 0.065 s [User: 1.066 s, System: 0.203 s]
Range (min … max): 1.230 s … 1.435 s 10 runs
```
`dvc init` crossing 1 sec is really bad, this was the result a month ago (before tpi and prompts, and with fsspec.Callback made lazy):
```console
$ hyperfine "dvc init" --warmup 3 -p "rm -rf .dvc"
Benchmark #1: dvc init
Time (mean ± σ): 506.8 ms ± 31.4 ms [User: 369.2 ms, System: 103.7 ms]
Range (min … max): 477.4 ms … 566.6 ms 10 runs
```
| perf | regression dvc init is slow the major issue seems to be tpi see fsspec callback prompts import on experiments from rich for dvc exp init there may not be much point in making fsspec lazy as we ll need local filesystem anyway the following is a startup log for dvc init img width alt screen shot at src following is the benchmark for dvc init console hyperfine dvc init warmup p rm rf dvc benchmark dvc init time mean ± σ s ± s range min … max s … s runs dvc init crossing sec is really bad this was the result a month ago before tpi and prompts and with fsspec callback made lazy console hyperfine dvc init warmup p rm rf dvc benchmark dvc init time mean ± σ ms ± ms range min … max ms … ms runs | 1 |
254,912 | 19,277,021,646 | IssuesEvent | 2021-12-10 13:05:51 | Djordje261/github-slideshow | https://api.github.com/repos/Djordje261/github-slideshow | closed | Getting Started with GitHub | documentation | # :wave: Bienvenido a "Introducción a GitHub" de GitHub Learning Lab
Para empezar, te guiaré a través de unos primeros pasos que son importantes para programar y colaborar en GitHub.
:point_down: _Esta flecha indica que puedes expandir la ventana! Haz clic en ellas durante el curso para más información._
<details><summary>¿Qué es GitHub?</summary>
<hr>
## ¿Qué es GitHub?
¡Me alegra que lo hayas preguntado! Muchas personas usan GitHub porque quieren contribuir a proyectos de código abierto <sup>[:book:](https://help.github.com/articles/github-glossary/#open-source)</sup>, o porque son invitados por compañeros de trabajo o de clase que usan GitHub en sus proyectos. ¿Por qué la gente usa GitHub en sus proyectos?
**En esencia, GitHub es una plataforma de colaboración.**
Desde software hasta documentos legales, puedes contar con GitHub para ayudarte a hacer tu mejor trabajo con las herramientas de colaboración y seguridad que tu equipo necesita. Con GitHub, puedes mantener proyectos completamente privados, invitar al mundo a colaborar y optimizar cada paso de tu proyecto.
**GitHub también es una poderosa herramienta de versión de control.**
GitHub usa Git <sup>[:book:](https://help.github.com/articles/github-glossary/#git)</sup>, el software de control de versión más popular para rastrear cada contribución y colaborador <sup>[:book:](https://help.github.com/articles/github-glossary/#contributor)</sup> de tu proyecto--de esta manera sabrás exactamente de dónde vino cada línea de código de tu proyecto.
**GitHub ayuda a las personas a hacer mucho más.**
La gente usa GitHub para construir varias de las tecnologías más avanzadas del mundo. Ya sea que estes visualizando datos o construyendo un nuevo juego, hay una comunidad entera y un set de herramientas en GitHub que te puedan llevar al siguiente nivel. Este curso empieza con lo básico pero ¡profundizaremos el resto después!
:tv: [Video: ¿Qué es GitHub?](https://www.youtube.com/watch?v=w3jLJU7DT5E)
<hr>
</details><br>
<details><summary>Explorando un repositorio de GitHub</summary>
<hr>
## Explorando un repositorio de GitHub
:tv: [Video: Explorando un repositorio](https://www.youtube.com/watch?v=R8OAwrcMlRw)
### Más características
El video cubre algunas de las características usadas con mayor frecuencia. Aquí hay otras cosas que puedes encontrar en los repositorios de GitHub:
- Tableros para el proyecto: Crear un tablero estilo Kanban para seguir tus tareas dentro de GitHub
- Wiki: Crea y guarda documentación relevante acerca de tu proyecto.
- Estadísticas: Ve un menú desplegable que contiene enlaces a herramientas de análisis para tu repositorio, que incluyen:
- Pulso: Encuentra información sobre el trabajo que ha sido completado y el trabajo que está en progreso en este panel del proyecto.
- Gráficas: Las gráficas proporcionan una vista más granular de la actividad del repositorio, incluyendo quién contribuyó al repositorio, quién lo bifurcó y cuándo completaron el trabajo.
### Archivos especiales
En el video aprendiste sobre un archivo especial llamado README.md. Aquí hay una lista de algunos otros archivos especiales que puedes añadir a tus repositorios:
- CONTRIBUTING.md: El archivo `CONTRIBUTING.md` es usado para describir el proceso para contribuir al repositorio. Un link al archivo `CONTRIBUTING.md` es mostrado cada vez que alguien crea una nueva propuesta o una solicitud de extracción.
- ISSUE_TEMPLATE.md: El archivo `ISSUE_TEMPLATE.md` lo puedes usar para rellenar el cuerpo de una propuesta. Por ejemplo, si siempre necesitas el mismo tipo de información para reportar errores puedes incluirlo en la plantilla de tu propuesta y cada nueva propuesta será abierta con el texto que has recomendado para empezar.
<hr>
</details>
### Usando las propuestas (o _issues_)
Esto es una propuesta (o _issue_) <sup>[:book:](https://help.github.com/articles/github-glossary/#issue)</sup>: un lugar en el que puedes tener conversaciones sobre los errores en tu código, revisiones de código y casi cualquier otra cosas que te puedas imaginar.
Los títulos de tus propuestas son como la línea de asunto de un email. Avisan de un vistazo a tus colaboradores de qué trata tu propuesta. Por ejemplo, el título de esta propuesta es Getting Started with GitHub.
<details><summary>Usando GitHub Issues (propuestas)</summary>
## Usando propuestas de GitHub
Las propuestas se usan para discutir ideas, mejoras, tareas y errores. Facilitan la colaboración al:
- Proveer a todos (incluso miembros futuros) la historia completa en un solo lugar
- Permitirte entrelazar tu propuesta con otras propuestas o una solicitud de extracción (o _pull request_) <sup>[:book:](https://help.github.com/articles/github-glossary/#pull-request)</sup>
- Crear un registro único y completo de cómo y por qué tomaste ciertas decisiones
- Permitirte atraer fácilmente a las personas y los equipos adecuados a una conversación con @mencion
:tv: [Video: Usando propuestas](https://www.youtube.com/watch?v=Zhj46r5D0nQ)
<hr>
</details>
<details><summary>Administrando notificaciones</summary>
<hr>
## Administrando notificaciones
:tv: [Video: Observando, notificaciones, estrellas y explorar](https://www.youtube.com/watch?v=ocQldxF7fMY)
Una vez que hayas comentado en una propuesta o solicitud de extracción, empezarás a recibir notificaciones por email cada vez que haya actividad en tu hilo.
## Cómo silenciar o dejar de silenciar conversaciones en específico.
1. Ve a la propuesta o solicitud de extracción
2. Bajo _"Notifications"_, haz clic en el botón de **Unsubscribe** a la derecha para silenciar notificaciones o **Subscribe** para dejar de silenciarlas.
Verás una descripción corta que explica tu estado actual de notificaciones.
### Cómo personalizar notificaciones en Settings
1. Haz clic en el icono de tu perfil
2. Haz clic en **Settings**
3. Haz clic en **Notifications** desde el menú a tu izquierda y [ajusta tus preferencias de notificaciones](https://help.github.com/articles/managing-notification-delivery-methods/)
### Opciones de notificación de un repositorio
* **Watch**: Recibirás una notificación cuando una nueva propuesta, solicitud de extracción o comentario ha sido publicado y también cuando una propuesta ha sido cerrado o una solicitud de extracción ha sido fusionada.
* **Not watching**: No vas a recibir notificaciones al menos que te hayan mencionado usando @
* **Ignore**: No vas a recibir ningún tipo de notificación sobre este respoitorio
## Cómo revisar las notificaciones de los repositorios que estás observando
1. Haz clic en tu icono de perfil
2. Haz clic en **Settings**
3. Haz clic en **Notification** desde el menú a tu izquierda
4. Haz clic en el link de [repositorios que estás observando](https://github.com/watching)
5. Selecciona la pestaña **Watching**
6. Haz clic en el botón de **Unwatch** para inhabilitar las notificaciones o en **Watch** para habilitarlas.
<hr>
</details>
<hr>
<h3 align="center">Sigue leyendo a continuación para encontrar tu primera tarea</h3>
| 1.0 | Getting Started with GitHub - # :wave: Bienvenido a "Introducción a GitHub" de GitHub Learning Lab
Para empezar, te guiaré a través de unos primeros pasos que son importantes para programar y colaborar en GitHub.
:point_down: _Esta flecha indica que puedes expandir la ventana! Haz clic en ellas durante el curso para más información._
<details><summary>¿Qué es GitHub?</summary>
<hr>
## ¿Qué es GitHub?
¡Me alegra que lo hayas preguntado! Muchas personas usan GitHub porque quieren contribuir a proyectos de código abierto <sup>[:book:](https://help.github.com/articles/github-glossary/#open-source)</sup>, o porque son invitados por compañeros de trabajo o de clase que usan GitHub en sus proyectos. ¿Por qué la gente usa GitHub en sus proyectos?
**En esencia, GitHub es una plataforma de colaboración.**
Desde software hasta documentos legales, puedes contar con GitHub para ayudarte a hacer tu mejor trabajo con las herramientas de colaboración y seguridad que tu equipo necesita. Con GitHub, puedes mantener proyectos completamente privados, invitar al mundo a colaborar y optimizar cada paso de tu proyecto.
**GitHub también es una poderosa herramienta de versión de control.**
GitHub usa Git <sup>[:book:](https://help.github.com/articles/github-glossary/#git)</sup>, el software de control de versión más popular para rastrear cada contribución y colaborador <sup>[:book:](https://help.github.com/articles/github-glossary/#contributor)</sup> de tu proyecto--de esta manera sabrás exactamente de dónde vino cada línea de código de tu proyecto.
**GitHub ayuda a las personas a hacer mucho más.**
La gente usa GitHub para construir varias de las tecnologías más avanzadas del mundo. Ya sea que estes visualizando datos o construyendo un nuevo juego, hay una comunidad entera y un set de herramientas en GitHub que te puedan llevar al siguiente nivel. Este curso empieza con lo básico pero ¡profundizaremos el resto después!
:tv: [Video: ¿Qué es GitHub?](https://www.youtube.com/watch?v=w3jLJU7DT5E)
<hr>
</details><br>
<details><summary>Explorando un repositorio de GitHub</summary>
<hr>
## Explorando un repositorio de GitHub
:tv: [Video: Explorando un repositorio](https://www.youtube.com/watch?v=R8OAwrcMlRw)
### Más características
El video cubre algunas de las características usadas con mayor frecuencia. Aquí hay otras cosas que puedes encontrar en los repositorios de GitHub:
- Tableros para el proyecto: Crear un tablero estilo Kanban para seguir tus tareas dentro de GitHub
- Wiki: Crea y guarda documentación relevante acerca de tu proyecto.
- Estadísticas: Ve un menú desplegable que contiene enlaces a herramientas de análisis para tu repositorio, que incluyen:
- Pulso: Encuentra información sobre el trabajo que ha sido completado y el trabajo que está en progreso en este panel del proyecto.
- Gráficas: Las gráficas proporcionan una vista más granular de la actividad del repositorio, incluyendo quién contribuyó al repositorio, quién lo bifurcó y cuándo completaron el trabajo.
### Archivos especiales
En el video aprendiste sobre un archivo especial llamado README.md. Aquí hay una lista de algunos otros archivos especiales que puedes añadir a tus repositorios:
- CONTRIBUTING.md: El archivo `CONTRIBUTING.md` es usado para describir el proceso para contribuir al repositorio. Un link al archivo `CONTRIBUTING.md` es mostrado cada vez que alguien crea una nueva propuesta o una solicitud de extracción.
- ISSUE_TEMPLATE.md: El archivo `ISSUE_TEMPLATE.md` lo puedes usar para rellenar el cuerpo de una propuesta. Por ejemplo, si siempre necesitas el mismo tipo de información para reportar errores puedes incluirlo en la plantilla de tu propuesta y cada nueva propuesta será abierta con el texto que has recomendado para empezar.
<hr>
</details>
### Usando las propuestas (o _issues_)
Esto es una propuesta (o _issue_) <sup>[:book:](https://help.github.com/articles/github-glossary/#issue)</sup>: un lugar en el que puedes tener conversaciones sobre los errores en tu código, revisiones de código y casi cualquier otra cosas que te puedas imaginar.
Los títulos de tus propuestas son como la línea de asunto de un email. Avisan de un vistazo a tus colaboradores de qué trata tu propuesta. Por ejemplo, el título de esta propuesta es Getting Started with GitHub.
<details><summary>Usando GitHub Issues (propuestas)</summary>
## Usando propuestas de GitHub
Las propuestas se usan para discutir ideas, mejoras, tareas y errores. Facilitan la colaboración al:
- Proveer a todos (incluso miembros futuros) la historia completa en un solo lugar
- Permitirte entrelazar tu propuesta con otras propuestas o una solicitud de extracción (o _pull request_) <sup>[:book:](https://help.github.com/articles/github-glossary/#pull-request)</sup>
- Crear un registro único y completo de cómo y por qué tomaste ciertas decisiones
- Permitirte atraer fácilmente a las personas y los equipos adecuados a una conversación con @mencion
:tv: [Video: Usando propuestas](https://www.youtube.com/watch?v=Zhj46r5D0nQ)
<hr>
</details>
<details><summary>Administrando notificaciones</summary>
<hr>
## Administrando notificaciones
:tv: [Video: Observando, notificaciones, estrellas y explorar](https://www.youtube.com/watch?v=ocQldxF7fMY)
Una vez que hayas comentado en una propuesta o solicitud de extracción, empezarás a recibir notificaciones por email cada vez que haya actividad en tu hilo.
## Cómo silenciar o dejar de silenciar conversaciones en específico.
1. Ve a la propuesta o solicitud de extracción
2. Bajo _"Notifications"_, haz clic en el botón de **Unsubscribe** a la derecha para silenciar notificaciones o **Subscribe** para dejar de silenciarlas.
Verás una descripción corta que explica tu estado actual de notificaciones.
### Cómo personalizar notificaciones en Settings
1. Haz clic en el icono de tu perfil
2. Haz clic en **Settings**
3. Haz clic en **Notifications** desde el menú a tu izquierda y [ajusta tus preferencias de notificaciones](https://help.github.com/articles/managing-notification-delivery-methods/)
### Opciones de notificación de un repositorio
* **Watch**: Recibirás una notificación cuando una nueva propuesta, solicitud de extracción o comentario ha sido publicado y también cuando una propuesta ha sido cerrado o una solicitud de extracción ha sido fusionada.
* **Not watching**: No vas a recibir notificaciones al menos que te hayan mencionado usando @
* **Ignore**: No vas a recibir ningún tipo de notificación sobre este respoitorio
## Cómo revisar las notificaciones de los repositorios que estás observando
1. Haz clic en tu icono de perfil
2. Haz clic en **Settings**
3. Haz clic en **Notification** desde el menú a tu izquierda
4. Haz clic en el link de [repositorios que estás observando](https://github.com/watching)
5. Selecciona la pestaña **Watching**
6. Haz clic en el botón de **Unwatch** para inhabilitar las notificaciones o en **Watch** para habilitarlas.
<hr>
</details>
<hr>
<h3 align="center">Sigue leyendo a continuación para encontrar tu primera tarea</h3>
| non_perf | getting started with github wave bienvenido a introducción a github de github learning lab para empezar te guiaré a través de unos primeros pasos que son importantes para programar y colaborar en github point down esta flecha indica que puedes expandir la ventana haz clic en ellas durante el curso para más información ¿qué es github ¿qué es github ¡me alegra que lo hayas preguntado muchas personas usan github porque quieren contribuir a proyectos de código abierto o porque son invitados por compañeros de trabajo o de clase que usan github en sus proyectos ¿por qué la gente usa github en sus proyectos en esencia github es una plataforma de colaboración desde software hasta documentos legales puedes contar con github para ayudarte a hacer tu mejor trabajo con las herramientas de colaboración y seguridad que tu equipo necesita con github puedes mantener proyectos completamente privados invitar al mundo a colaborar y optimizar cada paso de tu proyecto github también es una poderosa herramienta de versión de control github usa git el software de control de versión más popular para rastrear cada contribución y colaborador de tu proyecto de esta manera sabrás exactamente de dónde vino cada línea de código de tu proyecto github ayuda a las personas a hacer mucho más la gente usa github para construir varias de las tecnologías más avanzadas del mundo ya sea que estes visualizando datos o construyendo un nuevo juego hay una comunidad entera y un set de herramientas en github que te puedan llevar al siguiente nivel este curso empieza con lo básico pero ¡profundizaremos el resto después tv explorando un repositorio de github explorando un repositorio de github tv más características el video cubre algunas de las características usadas con mayor frecuencia aquí hay otras cosas que puedes encontrar en los repositorios de github tableros para el proyecto crear un tablero estilo kanban para seguir tus tareas dentro de github wiki crea y guarda documentación relevante acerca de tu proyecto estadísticas ve un menú desplegable que contiene enlaces a herramientas de análisis para tu repositorio que incluyen pulso encuentra información sobre el trabajo que ha sido completado y el trabajo que está en progreso en este panel del proyecto gráficas las gráficas proporcionan una vista más granular de la actividad del repositorio incluyendo quién contribuyó al repositorio quién lo bifurcó y cuándo completaron el trabajo archivos especiales en el video aprendiste sobre un archivo especial llamado readme md aquí hay una lista de algunos otros archivos especiales que puedes añadir a tus repositorios contributing md el archivo contributing md es usado para describir el proceso para contribuir al repositorio un link al archivo contributing md es mostrado cada vez que alguien crea una nueva propuesta o una solicitud de extracción issue template md el archivo issue template md lo puedes usar para rellenar el cuerpo de una propuesta por ejemplo si siempre necesitas el mismo tipo de información para reportar errores puedes incluirlo en la plantilla de tu propuesta y cada nueva propuesta será abierta con el texto que has recomendado para empezar usando las propuestas o issues esto es una propuesta o issue un lugar en el que puedes tener conversaciones sobre los errores en tu código revisiones de código y casi cualquier otra cosas que te puedas imaginar los títulos de tus propuestas son como la línea de asunto de un email avisan de un vistazo a tus colaboradores de qué trata tu propuesta por ejemplo el título de esta propuesta es getting started with github usando github issues propuestas usando propuestas de github las propuestas se usan para discutir ideas mejoras tareas y errores facilitan la colaboración al proveer a todos incluso miembros futuros la historia completa en un solo lugar permitirte entrelazar tu propuesta con otras propuestas o una solicitud de extracción o pull request crear un registro único y completo de cómo y por qué tomaste ciertas decisiones permitirte atraer fácilmente a las personas y los equipos adecuados a una conversación con mencion tv administrando notificaciones administrando notificaciones tv una vez que hayas comentado en una propuesta o solicitud de extracción empezarás a recibir notificaciones por email cada vez que haya actividad en tu hilo cómo silenciar o dejar de silenciar conversaciones en específico ve a la propuesta o solicitud de extracción bajo notifications haz clic en el botón de unsubscribe a la derecha para silenciar notificaciones o subscribe para dejar de silenciarlas verás una descripción corta que explica tu estado actual de notificaciones cómo personalizar notificaciones en settings haz clic en el icono de tu perfil haz clic en settings haz clic en notifications desde el menú a tu izquierda y opciones de notificación de un repositorio watch recibirás una notificación cuando una nueva propuesta solicitud de extracción o comentario ha sido publicado y también cuando una propuesta ha sido cerrado o una solicitud de extracción ha sido fusionada not watching no vas a recibir notificaciones al menos que te hayan mencionado usando ignore no vas a recibir ningún tipo de notificación sobre este respoitorio cómo revisar las notificaciones de los repositorios que estás observando haz clic en tu icono de perfil haz clic en settings haz clic en notification desde el menú a tu izquierda haz clic en el link de selecciona la pestaña watching haz clic en el botón de unwatch para inhabilitar las notificaciones o en watch para habilitarlas sigue leyendo a continuación para encontrar tu primera tarea | 0 |
22,179 | 11,502,662,263 | IssuesEvent | 2020-02-12 19:32:13 | influxdata/influxdb | https://api.github.com/repos/influxdata/influxdb | closed | investigate why cli is so slow | area/cli performance | with the influxd locally.
compare the cli time and api time. have 2 orgs
curl http://localhost:9999/api/v2/orgs took 0.02 secs.
run
```
influx org find|gnomon
```
got
```
2.5534s ID Name
0.0006s 60b45ab5d3826209 org2
0.0011s b3ef3f1a0e3cebc0 org1
0.0002s
Total 2.5562s
```
almost all command took at least 2.5 sec.
exp: `influx -h|gnomon`
```
2.7557s Influx Client
0.0014s
0.0001s Usage:
0.0001s influx [flags]
0.0001s influx [command]
0.0001s
0.0001s Available Commands:
0.0001s auth Authorization management commands
0.0001s bucket Bucket management commands
0.0001s help Help about any command
0.0001s org Organization management commands
0.0001s ping Check the InfluxDB /health endpoint
0.0001s query Execute a Flux query
0.0001s repl Interactive REPL (read-eval-print-loop)
0.0001s setup Setup instance with initial user, org, bucket
0.0001s task Task management commands
0.0001s user User management commands
0.0001s write Write points to InfluxDB
0.0000s
0.0000s Flags:
0.0000s -h, --help Help for the influx command
0.0000s --host string HTTP address of Influx (default "http://loca
lhost:9999")
0.0000s -t, --token string API token to be used throughout client calls
0.0010s Use "influx [command] --help" for more information about a command.
0.0001s Use "influx [command] --help" for more information about a command.
Total 2.7628s
``` | True | investigate why cli is so slow - with the influxd locally.
compare the cli time and api time. have 2 orgs
curl http://localhost:9999/api/v2/orgs took 0.02 secs.
run
```
influx org find|gnomon
```
got
```
2.5534s ID Name
0.0006s 60b45ab5d3826209 org2
0.0011s b3ef3f1a0e3cebc0 org1
0.0002s
Total 2.5562s
```
almost all command took at least 2.5 sec.
exp: `influx -h|gnomon`
```
2.7557s Influx Client
0.0014s
0.0001s Usage:
0.0001s influx [flags]
0.0001s influx [command]
0.0001s
0.0001s Available Commands:
0.0001s auth Authorization management commands
0.0001s bucket Bucket management commands
0.0001s help Help about any command
0.0001s org Organization management commands
0.0001s ping Check the InfluxDB /health endpoint
0.0001s query Execute a Flux query
0.0001s repl Interactive REPL (read-eval-print-loop)
0.0001s setup Setup instance with initial user, org, bucket
0.0001s task Task management commands
0.0001s user User management commands
0.0001s write Write points to InfluxDB
0.0000s
0.0000s Flags:
0.0000s -h, --help Help for the influx command
0.0000s --host string HTTP address of Influx (default "http://loca
lhost:9999")
0.0000s -t, --token string API token to be used throughout client calls
0.0010s Use "influx [command] --help" for more information about a command.
0.0001s Use "influx [command] --help" for more information about a command.
Total 2.7628s
``` | perf | investigate why cli is so slow with the influxd locally compare the cli time and api time have orgs curl took secs run influx org find gnomon got id name total almost all command took at least sec exp influx h gnomon influx client usage influx influx available commands auth authorization management commands bucket bucket management commands help help about any command org organization management commands ping check the influxdb health endpoint query execute a flux query repl interactive repl read eval print loop setup setup instance with initial user org bucket task task management commands user user management commands write write points to influxdb flags h help help for the influx command host string http address of influx default lhost t token string api token to be used throughout client calls use influx help for more information about a command use influx help for more information about a command total | 1 |
448,375 | 31,791,092,011 | IssuesEvent | 2023-09-13 03:25:40 | networkx/networkx | https://api.github.com/repos/networkx/networkx | opened | Some examples in tournament.py don't use tournament graphs | type: Documentation | Most `tournament` functions assume the input graph is a tournament. And the docs say the user must check whether it is a tournament. But some of the example don't use tournament graphs. The docs say that the function's output is undefined when the graph is not a tournament graph. So, while the examples do produce the output they claim to produce, it is a meaningless result.
For example: `tournament.is_strongly_connected(G)` returns False even though the graph in the example is strongly connected.
```python
>>> G = nx.DiGraph([(0, 1), (0, 2), (0, 3), (1, 2), (1, 3), (2, 3), (3, 0)])
>>> nx.tournament.is_strongly_connected(G)
True
>>> G.remove_edge(1, 3)
>>> nx.tournament.is_strongly_connected(G) # this graph is strongly connected.
False
```
The function `is_strongly_connected` returns `False` even though the graph G is strongly connected. This is because both examples used are not tournament graphs (exactly one directed edge between each pair of nodes).
I think every example in this module should add a line checking if the graph is a tournament before running the function. There may be others that are not valid example uses of the function. And those examples will reinforce the good practice of checking each time a graph is changed. | 1.0 | Some examples in tournament.py don't use tournament graphs - Most `tournament` functions assume the input graph is a tournament. And the docs say the user must check whether it is a tournament. But some of the example don't use tournament graphs. The docs say that the function's output is undefined when the graph is not a tournament graph. So, while the examples do produce the output they claim to produce, it is a meaningless result.
For example: `tournament.is_strongly_connected(G)` returns False even though the graph in the example is strongly connected.
```python
>>> G = nx.DiGraph([(0, 1), (0, 2), (0, 3), (1, 2), (1, 3), (2, 3), (3, 0)])
>>> nx.tournament.is_strongly_connected(G)
True
>>> G.remove_edge(1, 3)
>>> nx.tournament.is_strongly_connected(G) # this graph is strongly connected.
False
```
The function `is_strongly_connected` returns `False` even though the graph G is strongly connected. This is because both examples used are not tournament graphs (exactly one directed edge between each pair of nodes).
I think every example in this module should add a line checking if the graph is a tournament before running the function. There may be others that are not valid example uses of the function. And those examples will reinforce the good practice of checking each time a graph is changed. | non_perf | some examples in tournament py don t use tournament graphs most tournament functions assume the input graph is a tournament and the docs say the user must check whether it is a tournament but some of the example don t use tournament graphs the docs say that the function s output is undefined when the graph is not a tournament graph so while the examples do produce the output they claim to produce it is a meaningless result for example tournament is strongly connected g returns false even though the graph in the example is strongly connected python g nx digraph nx tournament is strongly connected g true g remove edge nx tournament is strongly connected g this graph is strongly connected false the function is strongly connected returns false even though the graph g is strongly connected this is because both examples used are not tournament graphs exactly one directed edge between each pair of nodes i think every example in this module should add a line checking if the graph is a tournament before running the function there may be others that are not valid example uses of the function and those examples will reinforce the good practice of checking each time a graph is changed | 0 |
7,919 | 6,301,749,950 | IssuesEvent | 2017-07-21 09:00:50 | akvo/akvo-flow | https://api.github.com/repos/akvo/akvo-flow | reopened | Improve loading time of device list | Improving performance Ready for release | I noticed on a partner dashboard that the Device list is loading all the devices ever connected to the dashboard (with some being last connected in 2013)
<img width="1184" alt="screen shot 2017-04-11 at 16 20 56" src="https://cloud.githubusercontent.com/assets/12456965/24913898/e50c0ab6-1ed2-11e7-878d-5e172c466dca.png">
Considering the user can organise the Device list based on all the header items, I think we need to reconsider how many entries we show in the list. Secondly, are we showing only the last time device connected to the instance or are we showing every time a device connected to the instance? | True | Improve loading time of device list - I noticed on a partner dashboard that the Device list is loading all the devices ever connected to the dashboard (with some being last connected in 2013)
<img width="1184" alt="screen shot 2017-04-11 at 16 20 56" src="https://cloud.githubusercontent.com/assets/12456965/24913898/e50c0ab6-1ed2-11e7-878d-5e172c466dca.png">
Considering the user can organise the Device list based on all the header items, I think we need to reconsider how many entries we show in the list. Secondly, are we showing only the last time device connected to the instance or are we showing every time a device connected to the instance? | perf | improve loading time of device list i noticed on a partner dashboard that the device list is loading all the devices ever connected to the dashboard with some being last connected in img width alt screen shot at src considering the user can organise the device list based on all the header items i think we need to reconsider how many entries we show in the list secondly are we showing only the last time device connected to the instance or are we showing every time a device connected to the instance | 1 |
46,931 | 7,295,757,005 | IssuesEvent | 2018-02-26 08:23:43 | laravel-enso/VueDatatable | https://api.github.com/repos/laravel-enso/VueDatatable | closed | Missing dependencies | documentation | Hi there and thank you for the component.
Just a heads-up that using this component independant of laravel/enso project needs a lot more
setup than described in the readme.
Don't know if it's in scope, but I'll try to resume here what needs to be done:
1. In the documentation, on step 4 there is a Toastr import that is not in the assets for this package. I assume that this package should have laravel-enso/vuecomponents as a dependency, since there is where the Toastr component lives.
2. There are a few npm dependencies missing for the missing icons if using the example provided:
- @ fortawesome/fontawesome
- @ fortawesome/fontawesome-free-solid
- @ fortawesome/fontawesome-free-brands
This icons need to be imported at some point during the app like this:
import { faUpload } from '@fortawesome/fontawesome-free-solid/shakable.es';
fontawesome.library.add([faUpload]);
This has to be done for any icons that will be used in the datatables (or referenced by table json that gets loaded into the controller).
3. @ fortawesome/vue-fontawesome needs to be installed with npm/yarn, imported in the bootstrap file and the component registered globally:
```
import FontAwesomeIcon from '@fortawesome/vue-fontawesome';
Vue.component('fa', FontAwesomeIcon);
```
Hope it helps someone. | 1.0 | Missing dependencies - Hi there and thank you for the component.
Just a heads-up that using this component independant of laravel/enso project needs a lot more
setup than described in the readme.
Don't know if it's in scope, but I'll try to resume here what needs to be done:
1. In the documentation, on step 4 there is a Toastr import that is not in the assets for this package. I assume that this package should have laravel-enso/vuecomponents as a dependency, since there is where the Toastr component lives.
2. There are a few npm dependencies missing for the missing icons if using the example provided:
- @ fortawesome/fontawesome
- @ fortawesome/fontawesome-free-solid
- @ fortawesome/fontawesome-free-brands
This icons need to be imported at some point during the app like this:
import { faUpload } from '@fortawesome/fontawesome-free-solid/shakable.es';
fontawesome.library.add([faUpload]);
This has to be done for any icons that will be used in the datatables (or referenced by table json that gets loaded into the controller).
3. @ fortawesome/vue-fontawesome needs to be installed with npm/yarn, imported in the bootstrap file and the component registered globally:
```
import FontAwesomeIcon from '@fortawesome/vue-fontawesome';
Vue.component('fa', FontAwesomeIcon);
```
Hope it helps someone. | non_perf | missing dependencies hi there and thank you for the component just a heads up that using this component independant of laravel enso project needs a lot more setup than described in the readme don t know if it s in scope but i ll try to resume here what needs to be done in the documentation on step there is a toastr import that is not in the assets for this package i assume that this package should have laravel enso vuecomponents as a dependency since there is where the toastr component lives there are a few npm dependencies missing for the missing icons if using the example provided fortawesome fontawesome fortawesome fontawesome free solid fortawesome fontawesome free brands this icons need to be imported at some point during the app like this import faupload from fortawesome fontawesome free solid shakable es fontawesome library add this has to be done for any icons that will be used in the datatables or referenced by table json that gets loaded into the controller fortawesome vue fontawesome needs to be installed with npm yarn imported in the bootstrap file and the component registered globally import fontawesomeicon from fortawesome vue fontawesome vue component fa fontawesomeicon hope it helps someone | 0 |
109,086 | 16,828,286,102 | IssuesEvent | 2021-06-17 22:03:37 | jtimberlake/rl-bakery | https://api.github.com/repos/jtimberlake/rl-bakery | opened | CVE-2021-33503 (High) detected in urllib3-1.25.11-py2.py3-none-any.whl | security vulnerability | ## CVE-2021-33503 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>urllib3-1.25.11-py2.py3-none-any.whl</b></p></summary>
<p>HTTP library with thread-safe connection pooling, file post, and more.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: rl-bakery/requirements.txt</p>
<p>Path to vulnerable library: rl-bakery/requirements.txt,rl-bakery</p>
<p>
Dependency Hierarchy:
- moto-1.3.14-py2.py3-none-any.whl (Root Library)
- boto3-1.12.18-py2.py3-none-any.whl
- botocore-1.15.49-py2.py3-none-any.whl
- :x: **urllib3-1.25.11-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/rl-bakery/commit/d91469bc8923d4e9b3605580b1f374632d029ad0">d91469bc8923d4e9b3605580b1f374632d029ad0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security issue has been found in python-urllib3 before version 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL was passed as a parameter or redirected to via an HTTP redirect.
<p>Publish Date: 2021-05-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503>CVE-2021-33503</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg">https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg</a></p>
<p>Release Date: 2021-05-22</p>
<p>Fix Resolution: urllib3 - 1.26.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"urllib3","packageVersion":"1.25.11","packageFilePaths":["/requirements.txt","rl-bakery"],"isTransitiveDependency":true,"dependencyTree":"moto:1.3.14;boto3:1.12.18;botocore:1.15.49;urllib3:1.25.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"urllib3 - 1.26.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33503","vulnerabilityDetails":"A security issue has been found in python-urllib3 before version 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL was passed as a parameter or redirected to via an HTTP redirect.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-33503 (High) detected in urllib3-1.25.11-py2.py3-none-any.whl - ## CVE-2021-33503 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>urllib3-1.25.11-py2.py3-none-any.whl</b></p></summary>
<p>HTTP library with thread-safe connection pooling, file post, and more.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: rl-bakery/requirements.txt</p>
<p>Path to vulnerable library: rl-bakery/requirements.txt,rl-bakery</p>
<p>
Dependency Hierarchy:
- moto-1.3.14-py2.py3-none-any.whl (Root Library)
- boto3-1.12.18-py2.py3-none-any.whl
- botocore-1.15.49-py2.py3-none-any.whl
- :x: **urllib3-1.25.11-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/rl-bakery/commit/d91469bc8923d4e9b3605580b1f374632d029ad0">d91469bc8923d4e9b3605580b1f374632d029ad0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A security issue has been found in python-urllib3 before version 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL was passed as a parameter or redirected to via an HTTP redirect.
<p>Publish Date: 2021-05-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503>CVE-2021-33503</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg">https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg</a></p>
<p>Release Date: 2021-05-22</p>
<p>Fix Resolution: urllib3 - 1.26.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"urllib3","packageVersion":"1.25.11","packageFilePaths":["/requirements.txt","rl-bakery"],"isTransitiveDependency":true,"dependencyTree":"moto:1.3.14;boto3:1.12.18;botocore:1.15.49;urllib3:1.25.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"urllib3 - 1.26.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33503","vulnerabilityDetails":"A security issue has been found in python-urllib3 before version 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL was passed as a parameter or redirected to via an HTTP redirect.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_perf | cve high detected in none any whl cve high severity vulnerability vulnerable library none any whl http library with thread safe connection pooling file post and more library home page a href path to dependency file rl bakery requirements txt path to vulnerable library rl bakery requirements txt rl bakery dependency hierarchy moto none any whl root library none any whl botocore none any whl x none any whl vulnerable library found in head commit a href found in base branch master vulnerability details a security issue has been found in python before version when provided with a url containing many characters in the authority component the authority regular expression exhibits catastrophic backtracking causing a denial of service if a url was passed as a parameter or redirected to via an http redirect publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree moto botocore isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails a security issue has been found in python before version when provided with a url containing many characters in the authority component the authority regular expression exhibits catastrophic backtracking causing a denial of service if a url was passed as a parameter or redirected to via an http redirect vulnerabilityurl | 0 |
51,535 | 27,125,630,965 | IssuesEvent | 2023-02-16 04:54:42 | keras-team/keras | https://api.github.com/repos/keras-team/keras | closed | tf.keras.layers.Normalization(invert=True) does not restore from save | type:bug/performance stat:awaiting response from contributor |
**System information**.
- Have I written custom code: yes
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 22.04
- TensorFlow installed from (source or binary): tf-nightly from pip
- TensorFlow version (use command below): 2.13.0-dev20230212
- Python version: 3.10.6
- Bazel version (if compiling from source):
- GPU model and memory: CPU/N/A
- Exact command to reproduce: run the code below:
**Describe the problem**.
`keras.models.load_model()` does not load `tf.keras.layers.Normalization` layer with `invert=True`.
**Describe the current behavior**.
It seems to omit `invert=True` flag when loading (or saving) the model.
**Describe the expected behavior**.
original trained and saved/loaded model output the same when using `Normalization` with `invert=True`.
**Standalone code to reproduce the issue**.
```python
import numpy as np
import pandas as pd
import sys
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import layers
print(tf.__version__)
dataset = pd.read_csv('https://raw.githubusercontent.com/mzakharo/micropython/tubby/ports/esp32/orp2.csv')
train_dataset = dataset.sample(frac=0.9, random_state=0)
train_features = train_dataset.copy()
train_labels = train_features.pop('orp')
inorm = tf.keras.layers.Normalization(axis=-1, input_shape=[2, ])
inorm.adapt(np.array(train_features))
onorm = tf.keras.layers.Normalization(axis=-1, invert=True)
onorm.adapt(train_labels)
def build_and_compile_model(inorm, onorm):
model = keras.Sequential([
inorm,
layers.Dense(10, activation='relu'),
layers.Dense(1),
onorm,
])
model.compile(loss='mean_squared_error', optimizer='adam')
return model
model = build_and_compile_model(inorm, onorm)
history = model.fit(
train_features,
train_labels,
validation_split=0.1,
verbose=0, epochs=10)
fname = 'model.h5'
model.save(fname)
p0 = 7.6
p1 = 0.85
v = model.predict([(p0, p1)], verbose=0)[0]
print('original output: ', v)
lmodel = keras.models.load_model(fname)
v = lmodel.predict([(p0, p1)], verbose=0)[0]
print('restored output: ', v)
```
**logs**.
```
original output: [667.36383]
restored output: [-8.889455]
```
| True | tf.keras.layers.Normalization(invert=True) does not restore from save -
**System information**.
- Have I written custom code: yes
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 22.04
- TensorFlow installed from (source or binary): tf-nightly from pip
- TensorFlow version (use command below): 2.13.0-dev20230212
- Python version: 3.10.6
- Bazel version (if compiling from source):
- GPU model and memory: CPU/N/A
- Exact command to reproduce: run the code below:
**Describe the problem**.
`keras.models.load_model()` does not load `tf.keras.layers.Normalization` layer with `invert=True`.
**Describe the current behavior**.
It seems to omit `invert=True` flag when loading (or saving) the model.
**Describe the expected behavior**.
original trained and saved/loaded model output the same when using `Normalization` with `invert=True`.
**Standalone code to reproduce the issue**.
```python
import numpy as np
import pandas as pd
import sys
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import layers
print(tf.__version__)
dataset = pd.read_csv('https://raw.githubusercontent.com/mzakharo/micropython/tubby/ports/esp32/orp2.csv')
train_dataset = dataset.sample(frac=0.9, random_state=0)
train_features = train_dataset.copy()
train_labels = train_features.pop('orp')
inorm = tf.keras.layers.Normalization(axis=-1, input_shape=[2, ])
inorm.adapt(np.array(train_features))
onorm = tf.keras.layers.Normalization(axis=-1, invert=True)
onorm.adapt(train_labels)
def build_and_compile_model(inorm, onorm):
model = keras.Sequential([
inorm,
layers.Dense(10, activation='relu'),
layers.Dense(1),
onorm,
])
model.compile(loss='mean_squared_error', optimizer='adam')
return model
model = build_and_compile_model(inorm, onorm)
history = model.fit(
train_features,
train_labels,
validation_split=0.1,
verbose=0, epochs=10)
fname = 'model.h5'
model.save(fname)
p0 = 7.6
p1 = 0.85
v = model.predict([(p0, p1)], verbose=0)[0]
print('original output: ', v)
lmodel = keras.models.load_model(fname)
v = lmodel.predict([(p0, p1)], verbose=0)[0]
print('restored output: ', v)
```
**logs**.
```
original output: [667.36383]
restored output: [-8.889455]
```
| perf | tf keras layers normalization invert true does not restore from save system information have i written custom code yes os platform and distribution e g linux ubuntu ubuntu tensorflow installed from source or binary tf nightly from pip tensorflow version use command below python version bazel version if compiling from source gpu model and memory cpu n a exact command to reproduce run the code below describe the problem keras models load model does not load tf keras layers normalization layer with invert true describe the current behavior it seems to omit invert true flag when loading or saving the model describe the expected behavior original trained and saved loaded model output the same when using normalization with invert true standalone code to reproduce the issue python import numpy as np import pandas as pd import sys import tensorflow as tf from tensorflow import keras from tensorflow keras import layers print tf version dataset pd read csv train dataset dataset sample frac random state train features train dataset copy train labels train features pop orp inorm tf keras layers normalization axis input shape inorm adapt np array train features onorm tf keras layers normalization axis invert true onorm adapt train labels def build and compile model inorm onorm model keras sequential inorm layers dense activation relu layers dense onorm model compile loss mean squared error optimizer adam return model model build and compile model inorm onorm history model fit train features train labels validation split verbose epochs fname model model save fname v model predict verbose print original output v lmodel keras models load model fname v lmodel predict verbose print restored output v logs original output restored output | 1 |
6,758 | 5,626,997,676 | IssuesEvent | 2017-04-04 23:41:47 | flutter/flutter | https://api.github.com/repos/flutter/flutter | opened | Engine roll caused regression to Android stock_layout_iteration times | performance regression ▣ platform-android | This PR: https://github.com/flutter/flutter/pull/9137
...which rolled the engine from: https://github.com/flutter/engine/commit/0005285781cd8eab7a43571bccdaad8354c6faa4
...to: https://github.com/flutter/engine/commit/4c05830aafb75c2893ea98bfd932ac95e883bce2
...may have caused a minor (1.05x) regression to the stock_layout_iteration benchmark on Android. | True | Engine roll caused regression to Android stock_layout_iteration times - This PR: https://github.com/flutter/flutter/pull/9137
...which rolled the engine from: https://github.com/flutter/engine/commit/0005285781cd8eab7a43571bccdaad8354c6faa4
...to: https://github.com/flutter/engine/commit/4c05830aafb75c2893ea98bfd932ac95e883bce2
...may have caused a minor (1.05x) regression to the stock_layout_iteration benchmark on Android. | perf | engine roll caused regression to android stock layout iteration times this pr which rolled the engine from to may have caused a minor regression to the stock layout iteration benchmark on android | 1 |
37,575 | 18,536,617,108 | IssuesEvent | 2021-10-21 12:14:30 | getsentry/sentry-javascript | https://api.github.com/repos/getsentry/sentry-javascript | closed | Question: Can we customize the end-time of Page Load transactions? Plus, capture paint time | Type: Discussion Feature: Performance Status: Stale | Hey folks! I'm coming up to speed on Sentry and trying to integrate it with my organization's frontend. One thing we'd like to do is understand better how Sentry starts / ends Page Load (and React Router) transactions. We've got some pages that we _believe_ are taking longer to "load" than Sentry is reporting, and we'd like a way to customize when those transactions are completed. Basically, I think our TTI is later than Sentry seems to believe it is.
For reference, here's a sample transaction and a few spans from it:

We're also using the React Router v5 integration to track load times for internal page switches, and we'd like to include some React metrics in there as well. It looks like, currently, only Navigation time and HTTP requests are being tracked. This would also be handleable if we knew how these transactions are being completed, and if we could attach more data to them.

Finally, we've got a component that's taking a LONG time to paint that we'd like to capture some metrics about. We can create a transaction inside React to handle the `render` bits, but it's a Canvas component and I believe the paint is taking upwards of 1s. We'd really like to capture that time (along with some metadata like number of painted elements) so we can measure improvement. Does Sentry have a way to capture this paint time?
Thanks again! | True | Question: Can we customize the end-time of Page Load transactions? Plus, capture paint time - Hey folks! I'm coming up to speed on Sentry and trying to integrate it with my organization's frontend. One thing we'd like to do is understand better how Sentry starts / ends Page Load (and React Router) transactions. We've got some pages that we _believe_ are taking longer to "load" than Sentry is reporting, and we'd like a way to customize when those transactions are completed. Basically, I think our TTI is later than Sentry seems to believe it is.
For reference, here's a sample transaction and a few spans from it:

We're also using the React Router v5 integration to track load times for internal page switches, and we'd like to include some React metrics in there as well. It looks like, currently, only Navigation time and HTTP requests are being tracked. This would also be handleable if we knew how these transactions are being completed, and if we could attach more data to them.

Finally, we've got a component that's taking a LONG time to paint that we'd like to capture some metrics about. We can create a transaction inside React to handle the `render` bits, but it's a Canvas component and I believe the paint is taking upwards of 1s. We'd really like to capture that time (along with some metadata like number of painted elements) so we can measure improvement. Does Sentry have a way to capture this paint time?
Thanks again! | perf | question can we customize the end time of page load transactions plus capture paint time hey folks i m coming up to speed on sentry and trying to integrate it with my organization s frontend one thing we d like to do is understand better how sentry starts ends page load and react router transactions we ve got some pages that we believe are taking longer to load than sentry is reporting and we d like a way to customize when those transactions are completed basically i think our tti is later than sentry seems to believe it is for reference here s a sample transaction and a few spans from it we re also using the react router integration to track load times for internal page switches and we d like to include some react metrics in there as well it looks like currently only navigation time and http requests are being tracked this would also be handleable if we knew how these transactions are being completed and if we could attach more data to them finally we ve got a component that s taking a long time to paint that we d like to capture some metrics about we can create a transaction inside react to handle the render bits but it s a canvas component and i believe the paint is taking upwards of we d really like to capture that time along with some metadata like number of painted elements so we can measure improvement does sentry have a way to capture this paint time thanks again | 1 |
32,738 | 15,605,943,770 | IssuesEvent | 2021-03-19 07:12:05 | ropensci/targets | https://api.github.com/repos/ropensci/targets | closed | tar_make slow when callr_function not NULL | depends: reprex topic: performance | ## Prework
* [x] Read and agree to the [code of conduct](https://ropensci.org/code-of-conduct/) and [contributing guidelines](https://github.com/ropensci/targets/blob/main/CONTRIBUTING.md).
* [x] If there is [already a relevant issue](https://github.com/ropensci/targets/issues), whether open or closed, comment on the existing thread instead of posting a new issue.
* [ ] Post a [minimal reproducible example](https://www.tidyverse.org/help/) like [this one](https://github.com/ropensci/targets/issues/256#issuecomment-754229683) so the maintainer can troubleshoot the problems you identify. A reproducible example is:
* [ ] **Runnable**: post enough R code and data so any onlooker can create the error on their own computer.
* [ ] **Minimal**: reduce runtime wherever possible and remove complicated details that are irrelevant to the issue at hand.
* [ ] **Readable**: format your code according to the [tidyverse style guide](https://style.tidyverse.org/).
## Description
`tar_make()` takes a lot of time to start up (several minutes). In comparison, `tar_make(callr_function = NULL)` starts almost without delay. See the benchmarks.
## Reproducible example
Since I'm working with confidential data, creating a reprex is quite some work. Therefore, I hope the benchmarks already help. If they don't I will take the time to create a reprex.
## Benchmarks
I tried to use `proffer`, but I got some errors. Also, as far as I understand, `proffer` should only work with `callr_function = NULL`.
<details>
```r
> proffer::pprof(tar_make())
* run target mifid_files
v skip target ids
v skip branch mifid_e5ccf3fc
v skip branch mifid_276b0dbc
v skip branch mifid_85ba955c
v skip branch mifid_32c58c89
v skip branch ids_list_d984c7ca
v skip branch ids_list_afadd32b
v skip branch ids_list_1622817d
v skip branch ids_list_5ff49e59
v skip branch id_counts_07b60a7a
v skip branch id_counts_4ffd5edf
v skip branch id_counts_6d3fac2a
v skip branch id_counts_f10827a2
v skip branch id_counts_259d208a
v skip branch id_counts_5bced0f6
v skip branch sdc_07b60a7a
v skip branch sdc_4ffd5edf
v skip branch sdc_6d3fac2a
v skip branch sdc_f10827a2
v skip branch sdc_259d208a
v skip branch sdc_5bced0f6
* end pipeline
Error in rethrow_call(c_processx_exec, command, c(command, args), stdin, :
Command '' not found @win/processx.c:994 (processx_exec)
Type .Last.error.trace to see where the error occured
> proffer::pprof(tar_make(callr_function = NULL))
* run target mifid_files
v skip target ids
v skip branch mifid_e5ccf3fc
v skip branch mifid_276b0dbc
v skip branch mifid_85ba955c
v skip branch mifid_32c58c89
v skip branch ids_list_d984c7ca
v skip branch ids_list_afadd32b
v skip branch ids_list_1622817d
v skip branch ids_list_5ff49e59
v skip branch id_counts_07b60a7a
v skip branch id_counts_4ffd5edf
v skip branch id_counts_6d3fac2a
v skip branch id_counts_f10827a2
v skip branch id_counts_259d208a
v skip branch id_counts_5bced0f6
v skip branch sdc_07b60a7a
v skip branch sdc_4ffd5edf
v skip branch sdc_6d3fac2a
v skip branch sdc_f10827a2
v skip branch sdc_259d208a
v skip branch sdc_5bced0f6
* end pipeline
Error in rethrow_call(c_processx_exec, command, c(command, args), stdin, :
Command '' not found @win/processx.c:994 (processx_exec)
Type .Last.error.trace to see where the error occured
> .Last.error.trace
Stack trace:
1. proffer::pprof(tar_make(callr_function = NULL))
2. proffer:::serve_pprof(pprof = pprof, host = host, port = port, ...
3. proffer:::serve_pprof_impl(args)
4. proffer:::with_safe_path(Sys.getenv("PROFFER_GRAPHVIZ_BIN"), ...
5. processx::process$new(command = pprof_path(), args = args, stdout = "|", ...
6. .subset2(public_bind_env, "initialize")(...)
7. processx:::process_initialize(self, private, command, args, stdin, ...
8. rethrow_call(c_processx_exec, command, c(command, args), stdin, ...
x Command '' not found @win/processx.c:994 (processx_exec)
```
</details>
Thus, I measured the time for `tar_make()` with `system.time()`:
```r
> system.time(tar_make(callr_function = NULL))
* run target mifid_files
v skip target ids
v skip branch mifid_e5ccf3fc
v skip branch mifid_276b0dbc
v skip branch mifid_85ba955c
v skip branch mifid_32c58c89
v skip branch ids_list_d984c7ca
v skip branch ids_list_afadd32b
v skip branch ids_list_1622817d
v skip branch ids_list_5ff49e59
v skip branch id_counts_07b60a7a
v skip branch id_counts_4ffd5edf
v skip branch id_counts_6d3fac2a
v skip branch id_counts_f10827a2
v skip branch id_counts_259d208a
v skip branch id_counts_5bced0f6
v skip branch sdc_07b60a7a
v skip branch sdc_4ffd5edf
v skip branch sdc_6d3fac2a
v skip branch sdc_f10827a2
v skip branch sdc_259d208a
v skip branch sdc_5bced0f6
* end pipeline
user system elapsed
0.20 0.39 10.23
> system.time(tar_make())
* run target mifid_files
v skip target ids
v skip branch mifid_e5ccf3fc
v skip branch mifid_276b0dbc
v skip branch mifid_85ba955c
v skip branch mifid_32c58c89
v skip branch ids_list_d984c7ca
v skip branch ids_list_afadd32b
v skip branch ids_list_1622817d
v skip branch ids_list_5ff49e59
v skip branch id_counts_07b60a7a
v skip branch id_counts_4ffd5edf
v skip branch id_counts_6d3fac2a
v skip branch id_counts_f10827a2
v skip branch id_counts_259d208a
v skip branch id_counts_5bced0f6
v skip branch sdc_07b60a7a
v skip branch sdc_4ffd5edf
v skip branch sdc_6d3fac2a
v skip branch sdc_f10827a2
v skip branch sdc_259d208a
v skip branch sdc_5bced0f6
* end pipeline
user system elapsed
2.39 0.59 372.45
```
I also profiled using `profvis`. Not sure if this helps, but see for yourself.
For `profvis::profvis(tar_make(callr_function = NULL))`:

And for `profvis::profvis(tar_make())`:

So it seems as if starting a new R session for `tar_make()` takes a lot of time, which I don't understand.
Am I missing something obvious here?
Also, as written before, please say if you need a real reprex and I'll try to create one.
| True | tar_make slow when callr_function not NULL - ## Prework
* [x] Read and agree to the [code of conduct](https://ropensci.org/code-of-conduct/) and [contributing guidelines](https://github.com/ropensci/targets/blob/main/CONTRIBUTING.md).
* [x] If there is [already a relevant issue](https://github.com/ropensci/targets/issues), whether open or closed, comment on the existing thread instead of posting a new issue.
* [ ] Post a [minimal reproducible example](https://www.tidyverse.org/help/) like [this one](https://github.com/ropensci/targets/issues/256#issuecomment-754229683) so the maintainer can troubleshoot the problems you identify. A reproducible example is:
* [ ] **Runnable**: post enough R code and data so any onlooker can create the error on their own computer.
* [ ] **Minimal**: reduce runtime wherever possible and remove complicated details that are irrelevant to the issue at hand.
* [ ] **Readable**: format your code according to the [tidyverse style guide](https://style.tidyverse.org/).
## Description
`tar_make()` takes a lot of time to start up (several minutes). In comparison, `tar_make(callr_function = NULL)` starts almost without delay. See the benchmarks.
## Reproducible example
Since I'm working with confidential data, creating a reprex is quite some work. Therefore, I hope the benchmarks already help. If they don't I will take the time to create a reprex.
## Benchmarks
I tried to use `proffer`, but I got some errors. Also, as far as I understand, `proffer` should only work with `callr_function = NULL`.
<details>
```r
> proffer::pprof(tar_make())
* run target mifid_files
v skip target ids
v skip branch mifid_e5ccf3fc
v skip branch mifid_276b0dbc
v skip branch mifid_85ba955c
v skip branch mifid_32c58c89
v skip branch ids_list_d984c7ca
v skip branch ids_list_afadd32b
v skip branch ids_list_1622817d
v skip branch ids_list_5ff49e59
v skip branch id_counts_07b60a7a
v skip branch id_counts_4ffd5edf
v skip branch id_counts_6d3fac2a
v skip branch id_counts_f10827a2
v skip branch id_counts_259d208a
v skip branch id_counts_5bced0f6
v skip branch sdc_07b60a7a
v skip branch sdc_4ffd5edf
v skip branch sdc_6d3fac2a
v skip branch sdc_f10827a2
v skip branch sdc_259d208a
v skip branch sdc_5bced0f6
* end pipeline
Error in rethrow_call(c_processx_exec, command, c(command, args), stdin, :
Command '' not found @win/processx.c:994 (processx_exec)
Type .Last.error.trace to see where the error occured
> proffer::pprof(tar_make(callr_function = NULL))
* run target mifid_files
v skip target ids
v skip branch mifid_e5ccf3fc
v skip branch mifid_276b0dbc
v skip branch mifid_85ba955c
v skip branch mifid_32c58c89
v skip branch ids_list_d984c7ca
v skip branch ids_list_afadd32b
v skip branch ids_list_1622817d
v skip branch ids_list_5ff49e59
v skip branch id_counts_07b60a7a
v skip branch id_counts_4ffd5edf
v skip branch id_counts_6d3fac2a
v skip branch id_counts_f10827a2
v skip branch id_counts_259d208a
v skip branch id_counts_5bced0f6
v skip branch sdc_07b60a7a
v skip branch sdc_4ffd5edf
v skip branch sdc_6d3fac2a
v skip branch sdc_f10827a2
v skip branch sdc_259d208a
v skip branch sdc_5bced0f6
* end pipeline
Error in rethrow_call(c_processx_exec, command, c(command, args), stdin, :
Command '' not found @win/processx.c:994 (processx_exec)
Type .Last.error.trace to see where the error occured
> .Last.error.trace
Stack trace:
1. proffer::pprof(tar_make(callr_function = NULL))
2. proffer:::serve_pprof(pprof = pprof, host = host, port = port, ...
3. proffer:::serve_pprof_impl(args)
4. proffer:::with_safe_path(Sys.getenv("PROFFER_GRAPHVIZ_BIN"), ...
5. processx::process$new(command = pprof_path(), args = args, stdout = "|", ...
6. .subset2(public_bind_env, "initialize")(...)
7. processx:::process_initialize(self, private, command, args, stdin, ...
8. rethrow_call(c_processx_exec, command, c(command, args), stdin, ...
x Command '' not found @win/processx.c:994 (processx_exec)
```
</details>
Thus, I measured the time for `tar_make()` with `system.time()`:
```r
> system.time(tar_make(callr_function = NULL))
* run target mifid_files
v skip target ids
v skip branch mifid_e5ccf3fc
v skip branch mifid_276b0dbc
v skip branch mifid_85ba955c
v skip branch mifid_32c58c89
v skip branch ids_list_d984c7ca
v skip branch ids_list_afadd32b
v skip branch ids_list_1622817d
v skip branch ids_list_5ff49e59
v skip branch id_counts_07b60a7a
v skip branch id_counts_4ffd5edf
v skip branch id_counts_6d3fac2a
v skip branch id_counts_f10827a2
v skip branch id_counts_259d208a
v skip branch id_counts_5bced0f6
v skip branch sdc_07b60a7a
v skip branch sdc_4ffd5edf
v skip branch sdc_6d3fac2a
v skip branch sdc_f10827a2
v skip branch sdc_259d208a
v skip branch sdc_5bced0f6
* end pipeline
user system elapsed
0.20 0.39 10.23
> system.time(tar_make())
* run target mifid_files
v skip target ids
v skip branch mifid_e5ccf3fc
v skip branch mifid_276b0dbc
v skip branch mifid_85ba955c
v skip branch mifid_32c58c89
v skip branch ids_list_d984c7ca
v skip branch ids_list_afadd32b
v skip branch ids_list_1622817d
v skip branch ids_list_5ff49e59
v skip branch id_counts_07b60a7a
v skip branch id_counts_4ffd5edf
v skip branch id_counts_6d3fac2a
v skip branch id_counts_f10827a2
v skip branch id_counts_259d208a
v skip branch id_counts_5bced0f6
v skip branch sdc_07b60a7a
v skip branch sdc_4ffd5edf
v skip branch sdc_6d3fac2a
v skip branch sdc_f10827a2
v skip branch sdc_259d208a
v skip branch sdc_5bced0f6
* end pipeline
user system elapsed
2.39 0.59 372.45
```
I also profiled using `profvis`. Not sure if this helps, but see for yourself.
For `profvis::profvis(tar_make(callr_function = NULL))`:

And for `profvis::profvis(tar_make())`:

So it seems as if starting a new R session for `tar_make()` takes a lot of time, which I don't understand.
Am I missing something obvious here?
Also, as written before, please say if you need a real reprex and I'll try to create one.
| perf | tar make slow when callr function not null prework read and agree to the and if there is whether open or closed comment on the existing thread instead of posting a new issue post a like so the maintainer can troubleshoot the problems you identify a reproducible example is runnable post enough r code and data so any onlooker can create the error on their own computer minimal reduce runtime wherever possible and remove complicated details that are irrelevant to the issue at hand readable format your code according to the description tar make takes a lot of time to start up several minutes in comparison tar make callr function null starts almost without delay see the benchmarks reproducible example since i m working with confidential data creating a reprex is quite some work therefore i hope the benchmarks already help if they don t i will take the time to create a reprex benchmarks i tried to use proffer but i got some errors also as far as i understand proffer should only work with callr function null r proffer pprof tar make run target mifid files v skip target ids v skip branch mifid v skip branch mifid v skip branch mifid v skip branch mifid v skip branch ids list v skip branch ids list v skip branch ids list v skip branch ids list v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc end pipeline error in rethrow call c processx exec command c command args stdin command not found win processx c processx exec type last error trace to see where the error occured proffer pprof tar make callr function null run target mifid files v skip target ids v skip branch mifid v skip branch mifid v skip branch mifid v skip branch mifid v skip branch ids list v skip branch ids list v skip branch ids list v skip branch ids list v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc end pipeline error in rethrow call c processx exec command c command args stdin command not found win processx c processx exec type last error trace to see where the error occured last error trace stack trace proffer pprof tar make callr function null proffer serve pprof pprof pprof host host port port proffer serve pprof impl args proffer with safe path sys getenv proffer graphviz bin processx process new command pprof path args args stdout public bind env initialize processx process initialize self private command args stdin rethrow call c processx exec command c command args stdin x command not found win processx c processx exec thus i measured the time for tar make with system time r system time tar make callr function null run target mifid files v skip target ids v skip branch mifid v skip branch mifid v skip branch mifid v skip branch mifid v skip branch ids list v skip branch ids list v skip branch ids list v skip branch ids list v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc end pipeline user system elapsed system time tar make run target mifid files v skip target ids v skip branch mifid v skip branch mifid v skip branch mifid v skip branch mifid v skip branch ids list v skip branch ids list v skip branch ids list v skip branch ids list v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch id counts v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc v skip branch sdc end pipeline user system elapsed i also profiled using profvis not sure if this helps but see for yourself for profvis profvis tar make callr function null and for profvis profvis tar make so it seems as if starting a new r session for tar make takes a lot of time which i don t understand am i missing something obvious here also as written before please say if you need a real reprex and i ll try to create one | 1 |
15,840 | 6,044,823,684 | IssuesEvent | 2017-06-12 07:27:46 | LuaJIT/LuaJIT | https://api.github.com/repos/LuaJIT/LuaJIT | closed | Static library doesn't build for LuaJIT v2.0.5 on Windows | 2.0 2.1 bug build system Windows | `AR` doesn't work correctly on `cmd.exe` on Windows.
The problem is that `AR` is `ar rcus 2>/dev/null`.
Since Windows doesn't have `/dev/null`, `AR` doesn't work.
Excerpt from [neovim build log](https://ci.appveyor.com/project/neovim/neovim/build/13809/job/sljudgsg3ykaw4uj):
```
cd /d C:\projects\neovim\.deps\build\src\luajit && C:\msys64\mingw64\bin\mingw32-make.exe CC=C:/msys64/mingw64/bin/gcc.exe PREFIX=C:/projects/neovim/.deps/usr CFLAGS+=-DLUAJIT_DISABLE_JIT CFLAGS+=-DLUA_USE_APICHECK CFLAGS+=-DLUA_USE_ASSERT CCDEBUG+=-g BUILDMODE=static
mingw32-make[3]: Entering directory 'C:/projects/neovim/.deps/build/src/luajit'
"==== Building LuaJIT 2.0.5 ===="
C:/msys64/mingw64/bin/mingw32-make -C src
mingw32-make[4]: Entering directory 'C:/projects/neovim/.deps/build/src/luajit/src'
"HOSTCC host/minilua.o"
"HOSTLINK host/minilua.exe"
"DYNASM host/buildvm_arch.h"
"HOSTCC host/buildvm.o"
"HOSTCC host/buildvm_asm.o"
"HOSTCC host/buildvm_peobj.o"
"HOSTCC host/buildvm_lib.o"
"HOSTCC host/buildvm_fold.o"
"HOSTLINK host/buildvm.exe"
"BUILDVM lj_vm.o"
"CC lj_gc.o"
"BUILDVM lj_ffdef.h"
"CC lj_err.o"
"CC lj_char.o"
"BUILDVM lj_bcdef.h"
"CC lj_bc.o"
"CC lj_obj.o"
"CC lj_str.o"
"CC lj_tab.o"
"CC lj_func.o"
"CC lj_udata.o"
"CC lj_meta.o"
"CC lj_debug.o"
"CC lj_state.o"
"CC lj_dispatch.o"
"CC lj_vmevent.o"
"CC lj_vmmath.o"
"CC lj_strscan.o"
"CC lj_api.o"
"CC lj_lex.o"
"CC lj_parse.o"
"CC lj_bcread.o"
"CC lj_bcwrite.o"
"CC lj_load.o"
"CC lj_ir.o"
"CC lj_opt_mem.o"
"BUILDVM lj_folddef.h"
"CC lj_opt_fold.o"
"CC lj_opt_narrow.o"
"CC lj_opt_dce.o"
"CC lj_opt_loop.o"
"CC lj_opt_split.o"
"CC lj_opt_sink.o"
"CC lj_mcode.o"
"CC lj_snap.o"
"CC lj_record.o"
"CC lj_crecord.o"
"BUILDVM lj_recdef.h"
"CC lj_ffrecord.o"
"CC lj_asm.o"
"CC lj_trace.o"
"CC lj_gdbjit.o"
"CC lj_ctype.o"
"CC lj_cdata.o"
"CC lj_cconv.o"
"CC lj_ccall.o"
"CC lj_ccallback.o"
"CC lj_carith.o"
"CC lj_clib.o"
"CC lj_cparse.o"
"CC lj_lib.o"
"CC lj_alloc.o"
"CC lib_aux.o"
"BUILDVM lj_libdef.h"
"CC lib_base.o"
"CC lib_math.o"
"CC lib_bit.o"
"CC lib_string.o"
"CC lib_table.o"
"CC lib_io.o"
"CC lib_os.o"
"CC lib_package.o"
"CC lib_debug.o"
"CC lib_jit.o"
"CC lib_ffi.o"
"CC lib_init.o"
"AR libluajit.a"
The system cannot find the path specified.
``` | 1.0 | Static library doesn't build for LuaJIT v2.0.5 on Windows - `AR` doesn't work correctly on `cmd.exe` on Windows.
The problem is that `AR` is `ar rcus 2>/dev/null`.
Since Windows doesn't have `/dev/null`, `AR` doesn't work.
Excerpt from [neovim build log](https://ci.appveyor.com/project/neovim/neovim/build/13809/job/sljudgsg3ykaw4uj):
```
cd /d C:\projects\neovim\.deps\build\src\luajit && C:\msys64\mingw64\bin\mingw32-make.exe CC=C:/msys64/mingw64/bin/gcc.exe PREFIX=C:/projects/neovim/.deps/usr CFLAGS+=-DLUAJIT_DISABLE_JIT CFLAGS+=-DLUA_USE_APICHECK CFLAGS+=-DLUA_USE_ASSERT CCDEBUG+=-g BUILDMODE=static
mingw32-make[3]: Entering directory 'C:/projects/neovim/.deps/build/src/luajit'
"==== Building LuaJIT 2.0.5 ===="
C:/msys64/mingw64/bin/mingw32-make -C src
mingw32-make[4]: Entering directory 'C:/projects/neovim/.deps/build/src/luajit/src'
"HOSTCC host/minilua.o"
"HOSTLINK host/minilua.exe"
"DYNASM host/buildvm_arch.h"
"HOSTCC host/buildvm.o"
"HOSTCC host/buildvm_asm.o"
"HOSTCC host/buildvm_peobj.o"
"HOSTCC host/buildvm_lib.o"
"HOSTCC host/buildvm_fold.o"
"HOSTLINK host/buildvm.exe"
"BUILDVM lj_vm.o"
"CC lj_gc.o"
"BUILDVM lj_ffdef.h"
"CC lj_err.o"
"CC lj_char.o"
"BUILDVM lj_bcdef.h"
"CC lj_bc.o"
"CC lj_obj.o"
"CC lj_str.o"
"CC lj_tab.o"
"CC lj_func.o"
"CC lj_udata.o"
"CC lj_meta.o"
"CC lj_debug.o"
"CC lj_state.o"
"CC lj_dispatch.o"
"CC lj_vmevent.o"
"CC lj_vmmath.o"
"CC lj_strscan.o"
"CC lj_api.o"
"CC lj_lex.o"
"CC lj_parse.o"
"CC lj_bcread.o"
"CC lj_bcwrite.o"
"CC lj_load.o"
"CC lj_ir.o"
"CC lj_opt_mem.o"
"BUILDVM lj_folddef.h"
"CC lj_opt_fold.o"
"CC lj_opt_narrow.o"
"CC lj_opt_dce.o"
"CC lj_opt_loop.o"
"CC lj_opt_split.o"
"CC lj_opt_sink.o"
"CC lj_mcode.o"
"CC lj_snap.o"
"CC lj_record.o"
"CC lj_crecord.o"
"BUILDVM lj_recdef.h"
"CC lj_ffrecord.o"
"CC lj_asm.o"
"CC lj_trace.o"
"CC lj_gdbjit.o"
"CC lj_ctype.o"
"CC lj_cdata.o"
"CC lj_cconv.o"
"CC lj_ccall.o"
"CC lj_ccallback.o"
"CC lj_carith.o"
"CC lj_clib.o"
"CC lj_cparse.o"
"CC lj_lib.o"
"CC lj_alloc.o"
"CC lib_aux.o"
"BUILDVM lj_libdef.h"
"CC lib_base.o"
"CC lib_math.o"
"CC lib_bit.o"
"CC lib_string.o"
"CC lib_table.o"
"CC lib_io.o"
"CC lib_os.o"
"CC lib_package.o"
"CC lib_debug.o"
"CC lib_jit.o"
"CC lib_ffi.o"
"CC lib_init.o"
"AR libluajit.a"
The system cannot find the path specified.
``` | non_perf | static library doesn t build for luajit on windows ar doesn t work correctly on cmd exe on windows the problem is that ar is ar rcus dev null since windows doesn t have dev null ar doesn t work excerpt from cd d c projects neovim deps build src luajit c bin make exe cc c bin gcc exe prefix c projects neovim deps usr cflags dluajit disable jit cflags dlua use apicheck cflags dlua use assert ccdebug g buildmode static make entering directory c projects neovim deps build src luajit building luajit c bin make c src make entering directory c projects neovim deps build src luajit src hostcc host minilua o hostlink host minilua exe dynasm host buildvm arch h hostcc host buildvm o hostcc host buildvm asm o hostcc host buildvm peobj o hostcc host buildvm lib o hostcc host buildvm fold o hostlink host buildvm exe buildvm lj vm o cc lj gc o buildvm lj ffdef h cc lj err o cc lj char o buildvm lj bcdef h cc lj bc o cc lj obj o cc lj str o cc lj tab o cc lj func o cc lj udata o cc lj meta o cc lj debug o cc lj state o cc lj dispatch o cc lj vmevent o cc lj vmmath o cc lj strscan o cc lj api o cc lj lex o cc lj parse o cc lj bcread o cc lj bcwrite o cc lj load o cc lj ir o cc lj opt mem o buildvm lj folddef h cc lj opt fold o cc lj opt narrow o cc lj opt dce o cc lj opt loop o cc lj opt split o cc lj opt sink o cc lj mcode o cc lj snap o cc lj record o cc lj crecord o buildvm lj recdef h cc lj ffrecord o cc lj asm o cc lj trace o cc lj gdbjit o cc lj ctype o cc lj cdata o cc lj cconv o cc lj ccall o cc lj ccallback o cc lj carith o cc lj clib o cc lj cparse o cc lj lib o cc lj alloc o cc lib aux o buildvm lj libdef h cc lib base o cc lib math o cc lib bit o cc lib string o cc lib table o cc lib io o cc lib os o cc lib package o cc lib debug o cc lib jit o cc lib ffi o cc lib init o ar libluajit a the system cannot find the path specified | 0 |
35,805 | 17,267,354,675 | IssuesEvent | 2021-07-22 15:13:07 | getsentry/sentry-cocoa | https://api.github.com/repos/getsentry/sentry-cocoa | closed | Name for SwiftUI auto generated transactions is long | Performance Product bug | ### Environment
iOSSwiftUI Sample app on this PR https://github.com/getsentry/sentry-cocoa/pull/1228
### Steps to Reproduce
Start iOSSwiftUI Sample app
Wait until the transaction is sent to Sentry
### Expected Result
A shorter nicer more descriptive transaction name.
### Actual Result
The name of the transaction is `_TtGC7SwiftUI19UIHostingControllerGVS_15ModifiedContentVS_7AnyViewVS_12RootModifier__`
<img width="724" alt="Screen Shot 2021-07-19 at 11 17 31" src="https://user-images.githubusercontent.com/2443292/126136172-446be568-b64d-4aa1-819a-866d1f02c928.png">
| True | Name for SwiftUI auto generated transactions is long - ### Environment
iOSSwiftUI Sample app on this PR https://github.com/getsentry/sentry-cocoa/pull/1228
### Steps to Reproduce
Start iOSSwiftUI Sample app
Wait until the transaction is sent to Sentry
### Expected Result
A shorter nicer more descriptive transaction name.
### Actual Result
The name of the transaction is `_TtGC7SwiftUI19UIHostingControllerGVS_15ModifiedContentVS_7AnyViewVS_12RootModifier__`
<img width="724" alt="Screen Shot 2021-07-19 at 11 17 31" src="https://user-images.githubusercontent.com/2443292/126136172-446be568-b64d-4aa1-819a-866d1f02c928.png">
| perf | name for swiftui auto generated transactions is long environment iosswiftui sample app on this pr steps to reproduce start iosswiftui sample app wait until the transaction is sent to sentry expected result a shorter nicer more descriptive transaction name actual result the name of the transaction is img width alt screen shot at src | 1 |
331,720 | 10,076,093,095 | IssuesEvent | 2019-07-24 15:31:04 | Kennygunderman/SunDownShowdown | https://api.github.com/repos/Kennygunderman/SunDownShowdown | closed | Clear mobs from arena | enhancement priority | We need a way to clear mobs from the arena either before the showdown starts (to make room for adding the new mobs) or after it ends (this would mean setting a time on when the showdown is 'over').
To do accomplish this we can specify the location and size of the showdown arena.
add cmd `/showdown arena set radius <radius size>` to set the location and size of the arena from where the user is standing.
With the arena location and size set, we can clear mobs from this location by using something like: `world.getNearbyEntitys`to get all entities within the arena, then having logic to remove mob entities. | 1.0 | Clear mobs from arena - We need a way to clear mobs from the arena either before the showdown starts (to make room for adding the new mobs) or after it ends (this would mean setting a time on when the showdown is 'over').
To do accomplish this we can specify the location and size of the showdown arena.
add cmd `/showdown arena set radius <radius size>` to set the location and size of the arena from where the user is standing.
With the arena location and size set, we can clear mobs from this location by using something like: `world.getNearbyEntitys`to get all entities within the arena, then having logic to remove mob entities. | non_perf | clear mobs from arena we need a way to clear mobs from the arena either before the showdown starts to make room for adding the new mobs or after it ends this would mean setting a time on when the showdown is over to do accomplish this we can specify the location and size of the showdown arena add cmd showdown arena set radius to set the location and size of the arena from where the user is standing with the arena location and size set we can clear mobs from this location by using something like world getnearbyentitys to get all entities within the arena then having logic to remove mob entities | 0 |
621,319 | 19,582,951,267 | IssuesEvent | 2022-01-05 00:41:43 | Qiskit-Partners/qiskit-ibm | https://api.github.com/repos/Qiskit-Partners/qiskit-ibm | closed | FAIL: test_provider_backends | type: bug priority: high | <!-- ⚠️ If you do not respect this template, your issue will be closed -->
<!-- ⚠️ Make sure to browse the opened and closed issues -->
### Information
- **Qiskit IBM Provider version**:
- **Python version**:
- **Operating system**:
### What is the current behavior?
3 tests are failing due to a deprecation warning in terra-main runs.
https://github.com/Qiskit/qiskit-ibmq-provider/runs/4377970666?check_suite_focus=true#step:5:4229
```
DeprecationWarning: The `backends` provider attribute is deprecated. Please use `provider.backend` (singular) instead. You can continue to use `provider.backends()` to retrieve all backends.
```
### Steps to reproduce the problem
### What is the expected behavior?
### Suggested solutions
| 1.0 | FAIL: test_provider_backends - <!-- ⚠️ If you do not respect this template, your issue will be closed -->
<!-- ⚠️ Make sure to browse the opened and closed issues -->
### Information
- **Qiskit IBM Provider version**:
- **Python version**:
- **Operating system**:
### What is the current behavior?
3 tests are failing due to a deprecation warning in terra-main runs.
https://github.com/Qiskit/qiskit-ibmq-provider/runs/4377970666?check_suite_focus=true#step:5:4229
```
DeprecationWarning: The `backends` provider attribute is deprecated. Please use `provider.backend` (singular) instead. You can continue to use `provider.backends()` to retrieve all backends.
```
### Steps to reproduce the problem
### What is the expected behavior?
### Suggested solutions
| non_perf | fail test provider backends information qiskit ibm provider version python version operating system what is the current behavior tests are failing due to a deprecation warning in terra main runs deprecationwarning the backends provider attribute is deprecated please use provider backend singular instead you can continue to use provider backends to retrieve all backends steps to reproduce the problem what is the expected behavior suggested solutions | 0 |
33,016 | 15,761,456,446 | IssuesEvent | 2021-03-31 10:00:25 | wazuh/wazuh-qa | https://api.github.com/repos/wazuh/wazuh-qa | opened | Performance tests: improve the deploy of the simulated agents | performance type/enhancement | The current implementation of this deployment adds a huge overhead in the pipeline because it is in charge of deploying the instances where the simulated agents will run, the toolset, and run the simulated agents.
We must move this deployment to Fargate to get rid of the instances deployment and the installation of the toolset. Using Fargate we can define a Docker image that will run the simulated agents, making it completely transparent to the pipeline.
**Tasks**
- [ ] Create the ECS cluster.
- [ ] Create a new Docker image to run the simulated agents.
- [ ] Create a new ECS task to run it:
- [ ] Create the task.
- [ ] Analyze the behavior of the Docker and establish the best hardware resources for it.
- [ ] Refactor the pipeline.
| True | Performance tests: improve the deploy of the simulated agents - The current implementation of this deployment adds a huge overhead in the pipeline because it is in charge of deploying the instances where the simulated agents will run, the toolset, and run the simulated agents.
We must move this deployment to Fargate to get rid of the instances deployment and the installation of the toolset. Using Fargate we can define a Docker image that will run the simulated agents, making it completely transparent to the pipeline.
**Tasks**
- [ ] Create the ECS cluster.
- [ ] Create a new Docker image to run the simulated agents.
- [ ] Create a new ECS task to run it:
- [ ] Create the task.
- [ ] Analyze the behavior of the Docker and establish the best hardware resources for it.
- [ ] Refactor the pipeline.
| perf | performance tests improve the deploy of the simulated agents the current implementation of this deployment adds a huge overhead in the pipeline because it is in charge of deploying the instances where the simulated agents will run the toolset and run the simulated agents we must move this deployment to fargate to get rid of the instances deployment and the installation of the toolset using fargate we can define a docker image that will run the simulated agents making it completely transparent to the pipeline tasks create the ecs cluster create a new docker image to run the simulated agents create a new ecs task to run it create the task analyze the behavior of the docker and establish the best hardware resources for it refactor the pipeline | 1 |
6,091 | 8,951,501,866 | IssuesEvent | 2019-01-25 14:08:53 | jasonblais/mattermost-community | https://api.github.com/repos/jasonblais/mattermost-community | opened | Add "PR Submitted" label for issues that already have a PR submitted | Contributor Journey Process | This could be automated for those PRs that reference a given help wanted issue. | 1.0 | Add "PR Submitted" label for issues that already have a PR submitted - This could be automated for those PRs that reference a given help wanted issue. | non_perf | add pr submitted label for issues that already have a pr submitted this could be automated for those prs that reference a given help wanted issue | 0 |
102,167 | 12,748,735,159 | IssuesEvent | 2020-06-26 20:49:28 | Hubvet8/hubvet-web | https://api.github.com/repos/Hubvet8/hubvet-web | opened | [Tarefa] Novo Breadcrumbs (Layout para novo posicionamento) | Módulo: Workspaces Tarefa UX Design | ## Descrição
Breve resumo do que é essa tarefa, descreva o trabalho a ser realizado
- Deve ser feito a tela de exames conforme layout e épico
Dados | Tipo
----|----
Texto 1 | String
Número 2 | Integer,int
Booleano 3 | boolean
Monetáio 4 | Float
Data 5 | datetime
Hora 6 | Time
Data e Hora | datetime
Lista 7 | Select/combobox
## Ações
- Selecionar
- Deve aparecer o toolbar com a quantidade de itens selecionados e um botão para excluir todos os selecionados
- Editar:
- Deve chamar o modal/dialog para editar o Exame OBS: fazer após tela de cadastro estar concluída
- Excluir:
- Deve excluir o exame da mesma linha
- Pesquisa:
- Filtros sigla ou nome
- [ ] Dados
- [ ] Botões novo Exame e importar
- [ ] Ações/Métodos
- [ ] Selecionar: Deve aparecer o toolbar com a quantidade de itens selecionados e um botão para
- [ ] excluir todos os selecionados
- [ ] Editar: Deve chamar o modal/dialog para editar o Exame OBS: fazer após tela de cadastro estar concluída
- [ ] Excluir: Deve excluir o exame da mesma linha
- [ ] Botão Novo Exame: Deve abrir o modal para cadastrar novo Exame
- [ ] Botão importar : Deve Abrir uma tela para escolher exames para importar exames
- [ ] Pesquisa:
- [ ] Pesquisar por sigla ou nome do exame
Sugestões:
Layout:
### API da tarefa
Link:
| 1.0 | [Tarefa] Novo Breadcrumbs (Layout para novo posicionamento) - ## Descrição
Breve resumo do que é essa tarefa, descreva o trabalho a ser realizado
- Deve ser feito a tela de exames conforme layout e épico
Dados | Tipo
----|----
Texto 1 | String
Número 2 | Integer,int
Booleano 3 | boolean
Monetáio 4 | Float
Data 5 | datetime
Hora 6 | Time
Data e Hora | datetime
Lista 7 | Select/combobox
## Ações
- Selecionar
- Deve aparecer o toolbar com a quantidade de itens selecionados e um botão para excluir todos os selecionados
- Editar:
- Deve chamar o modal/dialog para editar o Exame OBS: fazer após tela de cadastro estar concluída
- Excluir:
- Deve excluir o exame da mesma linha
- Pesquisa:
- Filtros sigla ou nome
- [ ] Dados
- [ ] Botões novo Exame e importar
- [ ] Ações/Métodos
- [ ] Selecionar: Deve aparecer o toolbar com a quantidade de itens selecionados e um botão para
- [ ] excluir todos os selecionados
- [ ] Editar: Deve chamar o modal/dialog para editar o Exame OBS: fazer após tela de cadastro estar concluída
- [ ] Excluir: Deve excluir o exame da mesma linha
- [ ] Botão Novo Exame: Deve abrir o modal para cadastrar novo Exame
- [ ] Botão importar : Deve Abrir uma tela para escolher exames para importar exames
- [ ] Pesquisa:
- [ ] Pesquisar por sigla ou nome do exame
Sugestões:
Layout:
### API da tarefa
Link:
| non_perf | novo breadcrumbs layout para novo posicionamento descrição breve resumo do que é essa tarefa descreva o trabalho a ser realizado deve ser feito a tela de exames conforme layout e épico dados tipo texto string número integer int booleano boolean monetáio float data datetime hora time data e hora datetime lista select combobox ações selecionar deve aparecer o toolbar com a quantidade de itens selecionados e um botão para excluir todos os selecionados editar deve chamar o modal dialog para editar o exame obs fazer após tela de cadastro estar concluída excluir deve excluir o exame da mesma linha pesquisa filtros sigla ou nome dados botões novo exame e importar ações métodos selecionar deve aparecer o toolbar com a quantidade de itens selecionados e um botão para excluir todos os selecionados editar deve chamar o modal dialog para editar o exame obs fazer após tela de cadastro estar concluída excluir deve excluir o exame da mesma linha botão novo exame deve abrir o modal para cadastrar novo exame botão importar deve abrir uma tela para escolher exames para importar exames pesquisa pesquisar por sigla ou nome do exame sugestões layout api da tarefa link | 0 |
70,268 | 23,086,210,479 | IssuesEvent | 2022-07-26 11:39:25 | matrix-org/synapse | https://api.github.com/repos/matrix-org/synapse | closed | Faster joins: check for partial state when handling backfill | A-Federated-Join T-Defect | Suppose:
* We have two chunks of room timeline, each with partial state.
* We attempt to backfill backwards from the later chunk, and fill the gap (so we reach the earlier chunk)
I think there is a bug here whereby we will incorrectly assume we have full state at the start of the backfilled chunk, which sounds messy.
https://github.com/matrix-org/synapse/blob/7c6b2204d143550d81e5bf9612c4e69fe0866b4c/synapse/handlers/federation_event.py#L778-L785
Part of #12646 | 1.0 | Faster joins: check for partial state when handling backfill - Suppose:
* We have two chunks of room timeline, each with partial state.
* We attempt to backfill backwards from the later chunk, and fill the gap (so we reach the earlier chunk)
I think there is a bug here whereby we will incorrectly assume we have full state at the start of the backfilled chunk, which sounds messy.
https://github.com/matrix-org/synapse/blob/7c6b2204d143550d81e5bf9612c4e69fe0866b4c/synapse/handlers/federation_event.py#L778-L785
Part of #12646 | non_perf | faster joins check for partial state when handling backfill suppose we have two chunks of room timeline each with partial state we attempt to backfill backwards from the later chunk and fill the gap so we reach the earlier chunk i think there is a bug here whereby we will incorrectly assume we have full state at the start of the backfilled chunk which sounds messy part of | 0 |
54,604 | 30,271,839,378 | IssuesEvent | 2023-07-07 15:57:53 | playcanvas/engine | https://api.github.com/repos/playcanvas/engine | closed | Low frame rates on Android (Chrome 94) with WebGL 2 | performance area: graphics | ### Description
Considerably lower frame rates when using WebGL 2 as compared to WebGL on Android (Chrome 94).
Reproduced on Samsung Galaxy Tab S7+, Samsung Galaxy S8+
The screenshots below were taken on the Galaxy Tab.
### WebGL 2

### WebGL

Example project: https://playcanvas.com/project/836325/overview/webglwebgl-2-perf-test | True | Low frame rates on Android (Chrome 94) with WebGL 2 - ### Description
Considerably lower frame rates when using WebGL 2 as compared to WebGL on Android (Chrome 94).
Reproduced on Samsung Galaxy Tab S7+, Samsung Galaxy S8+
The screenshots below were taken on the Galaxy Tab.
### WebGL 2

### WebGL

Example project: https://playcanvas.com/project/836325/overview/webglwebgl-2-perf-test | perf | low frame rates on android chrome with webgl description considerably lower frame rates when using webgl as compared to webgl on android chrome reproduced on samsung galaxy tab samsung galaxy the screenshots below were taken on the galaxy tab webgl webgl example project | 1 |
287,062 | 8,798,595,475 | IssuesEvent | 2018-12-24 08:45:44 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | support.microsoft.com - see bug description | browser-firefox priority-critical | <!-- @browser: Firefox 65.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update
**Browser / Version**: Firefox 65.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Access denied, this happens all the time with the Microsoft website searched through Google.com
**Steps to Reproduce**:
If I go to the Microsoft website directly (without searching through Google) there seems to be no problem...
But problem always appears in a case like this:
-> Searched on Google for "windows 10 update download"
-> And clicked on the following link:
De Update voor Windows 10, oktober 2018 downloaden
https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update
31 okt. 2018 - Windows 10 waarschuwt u wanneer de update van oktober 2018 beschikbaar is voor uw apparaat.
-> Then it showed "access denied"
[](https://webcompat.com/uploads/2018/12/e0231ff2-3ff4-4fd4-97a8-45736c9749fb.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181220174318</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Error: "The character encoding of the HTML document was not declared. The document will render with garbled text in some browser configurations if the document contains characters from outside the US-ASCII range. The character encoding of the page must be declared in the document or in the transfer protocol." {file: "https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update" line: 0}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | support.microsoft.com - see bug description - <!-- @browser: Firefox 65.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update
**Browser / Version**: Firefox 65.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Access denied, this happens all the time with the Microsoft website searched through Google.com
**Steps to Reproduce**:
If I go to the Microsoft website directly (without searching through Google) there seems to be no problem...
But problem always appears in a case like this:
-> Searched on Google for "windows 10 update download"
-> And clicked on the following link:
De Update voor Windows 10, oktober 2018 downloaden
https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update
31 okt. 2018 - Windows 10 waarschuwt u wanneer de update van oktober 2018 beschikbaar is voor uw apparaat.
-> Then it showed "access denied"
[](https://webcompat.com/uploads/2018/12/e0231ff2-3ff4-4fd4-97a8-45736c9749fb.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181220174318</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Error: "The character encoding of the HTML document was not declared. The document will render with garbled text in some browser configurations if the document contains characters from outside the US-ASCII range. The character encoding of the page must be declared in the document or in the transfer protocol." {file: "https://support.microsoft.com/nl-nl/help/4028685/windows-10-get-the-update" line: 0}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_perf | support microsoft com see bug description url browser version firefox operating system windows tested another browser yes problem type something else description access denied this happens all the time with the microsoft website searched through google com steps to reproduce if i go to the microsoft website directly without searching through google there seems to be no problem but problem always appears in a case like this searched on google for windows update download and clicked on the following link de update voor windows oktober downloaden okt windows waarschuwt u wanneer de update van oktober beschikbaar is voor uw apparaat then it showed access denied browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages from with ❤️ | 0 |
45,763 | 24,201,757,365 | IssuesEvent | 2022-09-24 17:14:22 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | closed | Question: Does the ZFS kernel modul benefit from compiler option -march=znver2 or znver3? | Type: Performance Type: Question Component: Encryption | I do not know where to ask that question so I put it here as a bug report. Sorry for that.
With the graysky2 kernel compiler patch (https://github.com/graysky2/kernel_compiler_patch) it is possible to compile the linux kernel with compiler optimizations for Ryzen processors for example (-march=znver2 or -march=znver3).
This patch is already included in linux-xanmod and linux-zen. And if the kernel is compiled with -march=znver2 the zfs module is also compiled with that option. I tested that.
Does that make any difference? Is that good or bad for the zfs performance or stability? I was wondering if encryption or compression would benefit from that.
| True | Question: Does the ZFS kernel modul benefit from compiler option -march=znver2 or znver3? - I do not know where to ask that question so I put it here as a bug report. Sorry for that.
With the graysky2 kernel compiler patch (https://github.com/graysky2/kernel_compiler_patch) it is possible to compile the linux kernel with compiler optimizations for Ryzen processors for example (-march=znver2 or -march=znver3).
This patch is already included in linux-xanmod and linux-zen. And if the kernel is compiled with -march=znver2 the zfs module is also compiled with that option. I tested that.
Does that make any difference? Is that good or bad for the zfs performance or stability? I was wondering if encryption or compression would benefit from that.
| perf | question does the zfs kernel modul benefit from compiler option march or i do not know where to ask that question so i put it here as a bug report sorry for that with the kernel compiler patch it is possible to compile the linux kernel with compiler optimizations for ryzen processors for example march or march this patch is already included in linux xanmod and linux zen and if the kernel is compiled with march the zfs module is also compiled with that option i tested that does that make any difference is that good or bad for the zfs performance or stability i was wondering if encryption or compression would benefit from that | 1 |
37,454 | 18,408,526,751 | IssuesEvent | 2021-10-13 00:39:04 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | opened | different output value in pytorch->onnx->tflite(int8 quantization) | type:performance | <em>Please make sure that this is an issue related to performance of TensorFlow.
As per our
[GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md),
we only address code/doc bugs, performance issues, feature requests and
build/installation issues on GitHub. tag:performance_template</em>
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 20.04
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): tensorflow:2.5.0-gpu docker
- TensorFlow version (use command below): 2.5.0
- Python version: 3.6.9
- Bazel version (if compiling from source):
- GCC/Compiler version (if compiling from source):
- CUDA/cuDNN version: 11.2/ 8.1.0
- GPU model and memory: RTX 3090
You can collect some of this information using our environment capture
[script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh)
You can also obtain the TensorFlow version with:
1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"`
2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"`
**Describe the current behavior**
I convert resnet50 pytorch -> onnx -> tflite with int8 quantization.
output value validation between pytorch <-> onnx, pytorch <-> pb, pytorch <-> tflite, pb <-> tflite
input is same image with size 256, check output value "np.testing.assert_allclose(output1, output2, rtol=1e-3, atol=1e-05)"
(using tflite interpreter only when i inference tflite "https://www.tensorflow.org/lite/guide/python?hl=ko")
Max absolute difference: 0.00076199 in pytorch <-> onnx
Max absolute difference: 0.00112534 in pytorch <-> pb
Max absolute difference: 13.387602 in pytorch <-> tflite(quantized)
Max absolute difference: 13.387438 in pb <-> tflite(quantized)
it's same max absolute difference between tflite(no quantized) and something(pytorch, onnx, pb)
ex) 0.0076~ in pytorch <-> tflite(no quant), 0.0011~ in pytorch <-> tflite(no quant)
i don't know why occur this difference
**Standalone code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary to generate
the problem. If possible, please share a link to Colab/Jupyter/any notebook.
**Other info / logs** Include any logs or source code that would be helpful to
diagnose the problem. If including tracebacks, please include the full
traceback. Large logs and files should be attached.
**pb to tflite log**
2021-10-13 09:18:56.162936: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0
2021-10-13 09:18:57.485452: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcuda.so.1
2021-10-13 09:18:57.511230: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.511916: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties:
pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6
coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s
2021-10-13 09:18:57.511955: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0
2021-10-13 09:18:57.513717: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcublas.so.11
2021-10-13 09:18:57.513767: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcublasLt.so.11
2021-10-13 09:18:57.514354: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcufft.so.10
2021-10-13 09:18:57.514537: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcurand.so.10
2021-10-13 09:18:57.515198: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcusolver.so.11
2021-10-13 09:18:57.515720: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcusparse.so.11
2021-10-13 09:18:57.515866: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudnn.so.8
2021-10-13 09:18:57.515918: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.516398: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.516976: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0
2021-10-13 09:18:57.517199: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 FMA
To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.
2021-10-13 09:18:57.517766: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.518224: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties:
pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6
coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s
2021-10-13 09:18:57.518272: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.518814: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.519243: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0
2021-10-13 09:18:57.519268: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0
2021-10-13 09:18:57.810376: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix:
2021-10-13 09:18:57.810410: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 0
2021-10-13 09:18:57.810420: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1277] 0: N
2021-10-13 09:18:57.810591: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.811162: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.811684: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.812186: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1418] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 21512 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 3090, pci bus id: 0000:01:00.0, compute capability: 8.6)
2021-10-13 09:18:58.498192: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:345] Ignored output_format.
2021-10-13 09:18:58.498225: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:348] Ignored drop_control_dependency.
2021-10-13 09:18:58.498234: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:354] Ignored change_concat_input_ranges.
2021-10-13 09:18:58.498881: I tensorflow/cc/saved_model/reader.cc:38] Reading SavedModel from: backbone_saved_model/
2021-10-13 09:18:58.515289: I tensorflow/cc/saved_model/reader.cc:90] Reading meta graph with tags { serve }
2021-10-13 09:18:58.515331: I tensorflow/cc/saved_model/reader.cc:132] Reading SavedModel debug info (if present) from: backbone_saved_model/
2021-10-13 09:18:58.515383: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix:
2021-10-13 09:18:58.515393: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264]
2021-10-13 09:18:58.527926: I tensorflow/cc/saved_model/loader.cc:206] Restoring SavedModel bundle.
2021-10-13 09:18:58.546224: I tensorflow/core/platform/profile_utils/cpu_utils.cc:114] CPU Frequency: 3699850000 Hz
2021-10-13 09:18:58.563849: I tensorflow/cc/saved_model/loader.cc:190] Running initialization op on SavedModel bundle at path: backbone_saved_model/
2021-10-13 09:18:58.577967: I tensorflow/cc/saved_model/loader.cc:277] SavedModel load for tags { serve }; Status: success: OK. Took 79088 microseconds.
2021-10-13 09:18:58.657933: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:210] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2021-10-13 09:18:58.675431: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.675985: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties:
pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6
coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s
2021-10-13 09:18:58.676068: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.676635: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.677107: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0
2021-10-13 09:18:58.677148: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix:
2021-10-13 09:18:58.677157: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 0
2021-10-13 09:18:58.677165: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1277] 0: N
2021-10-13 09:18:58.677253: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.677779: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.678280: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1418] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 21512 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 3090, pci bus id: 0000:01:00.0, compute capability: 8.6)
fully_quantize: 0, inference_type: 6, input_inference_type: 0, output_inference_type: 0
| True | different output value in pytorch->onnx->tflite(int8 quantization) - <em>Please make sure that this is an issue related to performance of TensorFlow.
As per our
[GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md),
we only address code/doc bugs, performance issues, feature requests and
build/installation issues on GitHub. tag:performance_template</em>
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 20.04
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): tensorflow:2.5.0-gpu docker
- TensorFlow version (use command below): 2.5.0
- Python version: 3.6.9
- Bazel version (if compiling from source):
- GCC/Compiler version (if compiling from source):
- CUDA/cuDNN version: 11.2/ 8.1.0
- GPU model and memory: RTX 3090
You can collect some of this information using our environment capture
[script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh)
You can also obtain the TensorFlow version with:
1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"`
2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"`
**Describe the current behavior**
I convert resnet50 pytorch -> onnx -> tflite with int8 quantization.
output value validation between pytorch <-> onnx, pytorch <-> pb, pytorch <-> tflite, pb <-> tflite
input is same image with size 256, check output value "np.testing.assert_allclose(output1, output2, rtol=1e-3, atol=1e-05)"
(using tflite interpreter only when i inference tflite "https://www.tensorflow.org/lite/guide/python?hl=ko")
Max absolute difference: 0.00076199 in pytorch <-> onnx
Max absolute difference: 0.00112534 in pytorch <-> pb
Max absolute difference: 13.387602 in pytorch <-> tflite(quantized)
Max absolute difference: 13.387438 in pb <-> tflite(quantized)
it's same max absolute difference between tflite(no quantized) and something(pytorch, onnx, pb)
ex) 0.0076~ in pytorch <-> tflite(no quant), 0.0011~ in pytorch <-> tflite(no quant)
i don't know why occur this difference
**Standalone code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary to generate
the problem. If possible, please share a link to Colab/Jupyter/any notebook.
**Other info / logs** Include any logs or source code that would be helpful to
diagnose the problem. If including tracebacks, please include the full
traceback. Large logs and files should be attached.
**pb to tflite log**
2021-10-13 09:18:56.162936: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0
2021-10-13 09:18:57.485452: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcuda.so.1
2021-10-13 09:18:57.511230: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.511916: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties:
pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6
coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s
2021-10-13 09:18:57.511955: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0
2021-10-13 09:18:57.513717: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcublas.so.11
2021-10-13 09:18:57.513767: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcublasLt.so.11
2021-10-13 09:18:57.514354: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcufft.so.10
2021-10-13 09:18:57.514537: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcurand.so.10
2021-10-13 09:18:57.515198: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcusolver.so.11
2021-10-13 09:18:57.515720: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcusparse.so.11
2021-10-13 09:18:57.515866: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudnn.so.8
2021-10-13 09:18:57.515918: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.516398: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.516976: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0
2021-10-13 09:18:57.517199: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 FMA
To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.
2021-10-13 09:18:57.517766: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.518224: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties:
pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6
coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s
2021-10-13 09:18:57.518272: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.518814: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.519243: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0
2021-10-13 09:18:57.519268: I tensorflow/stream_executor/platform/default/dso_loader.cc:53] Successfully opened dynamic library libcudart.so.11.0
2021-10-13 09:18:57.810376: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix:
2021-10-13 09:18:57.810410: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 0
2021-10-13 09:18:57.810420: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1277] 0: N
2021-10-13 09:18:57.810591: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.811162: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.811684: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:57.812186: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1418] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 21512 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 3090, pci bus id: 0000:01:00.0, compute capability: 8.6)
2021-10-13 09:18:58.498192: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:345] Ignored output_format.
2021-10-13 09:18:58.498225: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:348] Ignored drop_control_dependency.
2021-10-13 09:18:58.498234: W tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc:354] Ignored change_concat_input_ranges.
2021-10-13 09:18:58.498881: I tensorflow/cc/saved_model/reader.cc:38] Reading SavedModel from: backbone_saved_model/
2021-10-13 09:18:58.515289: I tensorflow/cc/saved_model/reader.cc:90] Reading meta graph with tags { serve }
2021-10-13 09:18:58.515331: I tensorflow/cc/saved_model/reader.cc:132] Reading SavedModel debug info (if present) from: backbone_saved_model/
2021-10-13 09:18:58.515383: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix:
2021-10-13 09:18:58.515393: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264]
2021-10-13 09:18:58.527926: I tensorflow/cc/saved_model/loader.cc:206] Restoring SavedModel bundle.
2021-10-13 09:18:58.546224: I tensorflow/core/platform/profile_utils/cpu_utils.cc:114] CPU Frequency: 3699850000 Hz
2021-10-13 09:18:58.563849: I tensorflow/cc/saved_model/loader.cc:190] Running initialization op on SavedModel bundle at path: backbone_saved_model/
2021-10-13 09:18:58.577967: I tensorflow/cc/saved_model/loader.cc:277] SavedModel load for tags { serve }; Status: success: OK. Took 79088 microseconds.
2021-10-13 09:18:58.657933: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:210] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2021-10-13 09:18:58.675431: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.675985: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1733] Found device 0 with properties:
pciBusID: 0000:01:00.0 name: NVIDIA GeForce RTX 3090 computeCapability: 8.6
coreClock: 1.86GHz coreCount: 82 deviceMemorySize: 23.68GiB deviceMemoryBandwidth: 871.81GiB/s
2021-10-13 09:18:58.676068: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.676635: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.677107: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1871] Adding visible gpu devices: 0
2021-10-13 09:18:58.677148: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1258] Device interconnect StreamExecutor with strength 1 edge matrix:
2021-10-13 09:18:58.677157: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1264] 0
2021-10-13 09:18:58.677165: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1277] 0: N
2021-10-13 09:18:58.677253: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.677779: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:937] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2021-10-13 09:18:58.678280: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1418] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 21512 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 3090, pci bus id: 0000:01:00.0, compute capability: 8.6)
fully_quantize: 0, inference_type: 6, input_inference_type: 0, output_inference_type: 0
| perf | different output value in pytorch onnx tflite quantization please make sure that this is an issue related to performance of tensorflow as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag performance template system information have i written custom code as opposed to using a stock example script provided in tensorflow no os platform and distribution e g linux ubuntu ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow installed from source or binary tensorflow gpu docker tensorflow version use command below python version bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory rtx you can collect some of this information using our environment capture you can also obtain the tensorflow version with tf python c import tensorflow as tf print tf git version tf version tf python c import tensorflow as tf print tf version git version tf version version describe the current behavior i convert pytorch onnx tflite with quantization output value validation between pytorch onnx pytorch pb pytorch tflite pb tflite input is same image with size check output value np testing assert allclose rtol atol using tflite interpreter only when i inference tflite max absolute difference in pytorch onnx max absolute difference in pytorch pb max absolute difference in pytorch tflite quantized max absolute difference in pb tflite quantized it s same max absolute difference between tflite no quantized and something pytorch onnx pb ex in pytorch tflite no quant in pytorch tflite no quant i don t know why occur this difference standalone code to reproduce the issue provide a reproducible test case that is the bare minimum necessary to generate the problem if possible please share a link to colab jupyter any notebook other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached pb to tflite log i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcuda so i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc found device with properties pcibusid name nvidia geforce rtx computecapability coreclock corecount devicememorysize devicememorybandwidth s i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcublas so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcublaslt so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcufft so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcurand so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcusolver so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcusparse so i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudnn so i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc adding visible gpu devices i tensorflow core platform cpu feature guard cc this tensorflow binary is optimized with oneapi deep neural network library onednn to use the following cpu instructions in performance critical operations fma to enable them in other operations rebuild tensorflow with the appropriate compiler flags i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc found device with properties pcibusid name nvidia geforce rtx computecapability coreclock corecount devicememorysize devicememorybandwidth s i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc adding visible gpu devices i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudart so i tensorflow core common runtime gpu gpu device cc device interconnect streamexecutor with strength edge matrix i tensorflow core common runtime gpu gpu device cc i tensorflow core common runtime gpu gpu device cc n i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc created tensorflow device job localhost replica task device gpu with mb memory physical gpu device name nvidia geforce rtx pci bus id compute capability w tensorflow compiler mlir lite python tf tfl flatbuffer helpers cc ignored output format w tensorflow compiler mlir lite python tf tfl flatbuffer helpers cc ignored drop control dependency w tensorflow compiler mlir lite python tf tfl flatbuffer helpers cc ignored change concat input ranges i tensorflow cc saved model reader cc reading savedmodel from backbone saved model i tensorflow cc saved model reader cc reading meta graph with tags serve i tensorflow cc saved model reader cc reading savedmodel debug info if present from backbone saved model i tensorflow core common runtime gpu gpu device cc device interconnect streamexecutor with strength edge matrix i tensorflow core common runtime gpu gpu device cc i tensorflow cc saved model loader cc restoring savedmodel bundle i tensorflow core platform profile utils cpu utils cc cpu frequency hz i tensorflow cc saved model loader cc running initialization op on savedmodel bundle at path backbone saved model i tensorflow cc saved model loader cc savedmodel load for tags serve status success ok took microseconds i tensorflow compiler mlir tensorflow utils dump mlir util cc disabling mlir crash reproducer set env var mlir crash reproducer directory to enable i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc found device with properties pcibusid name nvidia geforce rtx computecapability coreclock corecount devicememorysize devicememorybandwidth s i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc adding visible gpu devices i tensorflow core common runtime gpu gpu device cc device interconnect streamexecutor with strength edge matrix i tensorflow core common runtime gpu gpu device cc i tensorflow core common runtime gpu gpu device cc n i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow stream executor cuda cuda gpu executor cc successful numa node read from sysfs had negative value but there must be at least one numa node so returning numa node zero i tensorflow core common runtime gpu gpu device cc created tensorflow device job localhost replica task device gpu with mb memory physical gpu device name nvidia geforce rtx pci bus id compute capability fully quantize inference type input inference type output inference type | 1 |
1,389 | 2,984,858,044 | IssuesEvent | 2015-07-18 12:12:17 | akamsteeg/SwissArmyKnife | https://api.github.com/repos/akamsteeg/SwissArmyKnife | closed | string.Truncate() with suffix uses string.FormatWith(), instead of cheaper concatenation | Performance | `string.Truncate(int length, string suffix)` uses `string.FormatWith()` which in turn uses `string.Format()` internally. That's a bit wasteful, because a simple concatenation is far cheaper.
`string.Truncate()` with `string.FormatWith()`: 106 ns.
`string.Truncate()` with string concatenation: 64 ns. | True | string.Truncate() with suffix uses string.FormatWith(), instead of cheaper concatenation - `string.Truncate(int length, string suffix)` uses `string.FormatWith()` which in turn uses `string.Format()` internally. That's a bit wasteful, because a simple concatenation is far cheaper.
`string.Truncate()` with `string.FormatWith()`: 106 ns.
`string.Truncate()` with string concatenation: 64 ns. | perf | string truncate with suffix uses string formatwith instead of cheaper concatenation string truncate int length string suffix uses string formatwith which in turn uses string format internally that s a bit wasteful because a simple concatenation is far cheaper string truncate with string formatwith ns string truncate with string concatenation ns | 1 |
29,754 | 14,256,177,664 | IssuesEvent | 2020-11-20 00:22:06 | davidgarland/circa | https://api.github.com/repos/davidgarland/circa | closed | Differentiate between in-place and to-array higher-order functions. | enhancement memory performance | Right now:
* foldl/foldr are to-value.
* filter is in-place.
* map is in-place.
These are fairly sane defaults, but to-array variants of `filter` and `map`, as well as in-place variatns of `foldl/foldr`, would be nice to have for convenience and performance reasons depending upon situation. A naming convention differentiating between these (perhaps `_to` for "to-value"/"to-array" versions?) would be nice. | True | Differentiate between in-place and to-array higher-order functions. - Right now:
* foldl/foldr are to-value.
* filter is in-place.
* map is in-place.
These are fairly sane defaults, but to-array variants of `filter` and `map`, as well as in-place variatns of `foldl/foldr`, would be nice to have for convenience and performance reasons depending upon situation. A naming convention differentiating between these (perhaps `_to` for "to-value"/"to-array" versions?) would be nice. | perf | differentiate between in place and to array higher order functions right now foldl foldr are to value filter is in place map is in place these are fairly sane defaults but to array variants of filter and map as well as in place variatns of foldl foldr would be nice to have for convenience and performance reasons depending upon situation a naming convention differentiating between these perhaps to for to value to array versions would be nice | 1 |
5,271 | 4,864,245,574 | IssuesEvent | 2016-11-14 17:27:48 | elmsln/elmsln | https://api.github.com/repos/elmsln/elmsln | closed | Disable aggregation for better performance | scale / performance | Weird but true. There's no gain from aggregating resources. | True | Disable aggregation for better performance - Weird but true. There's no gain from aggregating resources. | perf | disable aggregation for better performance weird but true there s no gain from aggregating resources | 1 |
9,099 | 6,760,875,616 | IssuesEvent | 2017-10-24 22:21:07 | Scratchcat1/AATC | https://api.github.com/repos/Scratchcat1/AATC | opened | Monitor viewer render inefficiency | performance | https://github.com/Scratchcat1/AATC/blob/dd43ecaff710c77337c95ffbc70f30e791883068/AATC_Monitor_Viewer.py#L94-L95
These lines need not be calculated if object is not to be drawn. | True | Monitor viewer render inefficiency - https://github.com/Scratchcat1/AATC/blob/dd43ecaff710c77337c95ffbc70f30e791883068/AATC_Monitor_Viewer.py#L94-L95
These lines need not be calculated if object is not to be drawn. | perf | monitor viewer render inefficiency these lines need not be calculated if object is not to be drawn | 1 |
27,070 | 13,171,778,431 | IssuesEvent | 2020-08-11 17:15:43 | Giveth/giveth-dapp | https://api.github.com/repos/Giveth/giveth-dapp | closed | Loading times are extremely high on the dapp | bug performance | Loading times shouldn't be as high specially on the main page | True | Loading times are extremely high on the dapp - Loading times shouldn't be as high specially on the main page | perf | loading times are extremely high on the dapp loading times shouldn t be as high specially on the main page | 1 |
48,151 | 25,381,221,055 | IssuesEvent | 2022-11-21 17:42:15 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | kv: pebbleReadOnly constructed in replicaRaftStorage.Entries even when reading from entry cache | C-performance A-kv-replication T-kv-replication | I noticed in a recent round of benchmarking that we construct a `pebbleReadOnly` in `replicaRaftStorage.Entries` even in the vastly common case where all entries are served out of the raft entry cache:
https://github.com/cockroachdb/cockroach/blob/2a041d61d021f563828597fe9961f54727419c1d/pkg/kv/kvserver/replica_raftstorage.go#L102
This is wasteful.
I would have made the fix of pushing this into `entries` myself, but there's was just enough friction due to `term`'s [use of the `reader`](https://github.com/cockroachdb/cockroach/blob/2a041d61d021f563828597fe9961f54727419c1d/pkg/kv/kvserver/replica_raftstorage.go#L296) that I figured I'd leave it to the experts.
Is it worth asking whether this code needs to operate on a LSM snapshot? There should not be concurrent mutation to the state it is looking at. | True | kv: pebbleReadOnly constructed in replicaRaftStorage.Entries even when reading from entry cache - I noticed in a recent round of benchmarking that we construct a `pebbleReadOnly` in `replicaRaftStorage.Entries` even in the vastly common case where all entries are served out of the raft entry cache:
https://github.com/cockroachdb/cockroach/blob/2a041d61d021f563828597fe9961f54727419c1d/pkg/kv/kvserver/replica_raftstorage.go#L102
This is wasteful.
I would have made the fix of pushing this into `entries` myself, but there's was just enough friction due to `term`'s [use of the `reader`](https://github.com/cockroachdb/cockroach/blob/2a041d61d021f563828597fe9961f54727419c1d/pkg/kv/kvserver/replica_raftstorage.go#L296) that I figured I'd leave it to the experts.
Is it worth asking whether this code needs to operate on a LSM snapshot? There should not be concurrent mutation to the state it is looking at. | perf | kv pebblereadonly constructed in replicaraftstorage entries even when reading from entry cache i noticed in a recent round of benchmarking that we construct a pebblereadonly in replicaraftstorage entries even in the vastly common case where all entries are served out of the raft entry cache this is wasteful i would have made the fix of pushing this into entries myself but there s was just enough friction due to term s that i figured i d leave it to the experts is it worth asking whether this code needs to operate on a lsm snapshot there should not be concurrent mutation to the state it is looking at | 1 |
3,611 | 2,773,523,446 | IssuesEvent | 2015-05-03 18:24:31 | RexOps/Rex | https://api.github.com/repos/RexOps/Rex | closed | PATH environment problem | documentation | Hi ,
in /etc/profile , I export some path to PATH environment ,
so I use source_global_profile(1) command to make sure my run command will work fine.
But this does not work , because rex reset PATH environment after sourcing /etc/profile.
the workaround for me is sourcing again before run command
run(". /etc/profile >/dev/null 2>&1 ; mvn -version");
== part of /etc/profile ==
export M2_HOME=/usr/local/apache-maven
export M2=$M2_HOME/bin
export PATH=$M2:$PATH
== part of log ==
[2015-05-03 16:27:06] DEBUG - SSH/executing: . /etc/profile >/dev/null 2>&1 ; LC_ALL=C ; export LC_ALL; PATH=/bin:/sbin:/usr/bin:/usr/sbin:/usr/local/bin:/usr/local/sbin:/usr/pkg/bin:/usr/pkg/sbin; export PATH; mvn -version | 1.0 | PATH environment problem - Hi ,
in /etc/profile , I export some path to PATH environment ,
so I use source_global_profile(1) command to make sure my run command will work fine.
But this does not work , because rex reset PATH environment after sourcing /etc/profile.
the workaround for me is sourcing again before run command
run(". /etc/profile >/dev/null 2>&1 ; mvn -version");
== part of /etc/profile ==
export M2_HOME=/usr/local/apache-maven
export M2=$M2_HOME/bin
export PATH=$M2:$PATH
== part of log ==
[2015-05-03 16:27:06] DEBUG - SSH/executing: . /etc/profile >/dev/null 2>&1 ; LC_ALL=C ; export LC_ALL; PATH=/bin:/sbin:/usr/bin:/usr/sbin:/usr/local/bin:/usr/local/sbin:/usr/pkg/bin:/usr/pkg/sbin; export PATH; mvn -version | non_perf | path environment problem hi in etc profile i export some path to path environment so i use source global profile command to make sure my run command will work fine but this does not work because rex reset path environment after sourcing etc profile the workaround for me is sourcing again before run command run etc profile dev null mvn version part of etc profile export home usr local apache maven export home bin export path path part of log debug ssh executing etc profile dev null lc all c export lc all path bin sbin usr bin usr sbin usr local bin usr local sbin usr pkg bin usr pkg sbin export path mvn version | 0 |
38,084 | 8,408,510,994 | IssuesEvent | 2018-10-12 02:04:36 | robot-lab/judyst-link-analysis | https://api.github.com/repos/robot-lab/judyst-link-analysis | closed | Выделить webcrawler в компоненту | code maintenance task | Нужно перенести компонент сбора данных в репозиторий judyst-web-crawler | 1.0 | Выделить webcrawler в компоненту - Нужно перенести компонент сбора данных в репозиторий judyst-web-crawler | non_perf | выделить webcrawler в компоненту нужно перенести компонент сбора данных в репозиторий judyst web crawler | 0 |
159,619 | 20,085,863,570 | IssuesEvent | 2022-02-05 01:03:11 | doc-ai/tensorflow-ios-framework | https://api.github.com/repos/doc-ai/tensorflow-ios-framework | opened | CVE-2020-26266 (Medium) detected in tensorflowv2.0.2 | security vulnerability | ## CVE-2020-26266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflowv2.0.2</b></p></summary>
<p>
<p>An Open Source Machine Learning Framework for Everyone</p>
<p>Library home page: <a href=https://github.com/tensorflow/tensorflow.git>https://github.com/tensorflow/tensorflow.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In affected versions of TensorFlow under certain cases a saved model can trigger use of uninitialized values during code execution. This is caused by having tensor buffers be filled with the default value of the type but forgetting to default initialize the quantized floating point types in Eigen. This is fixed in versions 1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, and 2.4.0.
<p>Publish Date: 2020-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26266>CVE-2020-26266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qhxx-j73r-qpm2">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qhxx-j73r-qpm2</a></p>
<p>Release Date: 2020-12-10</p>
<p>Fix Resolution: tensorflow-1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, 2.4.0,tensorflow-cpu-2.1.3, 2.2.2, 2.3.2, 2.4.0,tensorflow-gpu-1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, 2.4.0</p>
</p>
</details>
<p></p>
| True | CVE-2020-26266 (Medium) detected in tensorflowv2.0.2 - ## CVE-2020-26266 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflowv2.0.2</b></p></summary>
<p>
<p>An Open Source Machine Learning Framework for Everyone</p>
<p>Library home page: <a href=https://github.com/tensorflow/tensorflow.git>https://github.com/tensorflow/tensorflow.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/tensorflow-ios-framework/tensorflow.framework/Headers/third_party/eigen3/unsupported/Eigen/CXX11/src/FixedPoint/FixedPointTypes.h</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In affected versions of TensorFlow under certain cases a saved model can trigger use of uninitialized values during code execution. This is caused by having tensor buffers be filled with the default value of the type but forgetting to default initialize the quantized floating point types in Eigen. This is fixed in versions 1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, and 2.4.0.
<p>Publish Date: 2020-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26266>CVE-2020-26266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qhxx-j73r-qpm2">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-qhxx-j73r-qpm2</a></p>
<p>Release Date: 2020-12-10</p>
<p>Fix Resolution: tensorflow-1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, 2.4.0,tensorflow-cpu-2.1.3, 2.2.2, 2.3.2, 2.4.0,tensorflow-gpu-1.15.5, 2.0.4, 2.1.3, 2.2.2, 2.3.2, 2.4.0</p>
</p>
</details>
<p></p>
| non_perf | cve medium detected in cve medium severity vulnerability vulnerable library an open source machine learning framework for everyone library home page a href found in base branch master vulnerable source files tensorflow ios framework tensorflow framework headers third party unsupported eigen src fixedpoint fixedpointtypes h tensorflow ios framework tensorflow framework headers third party unsupported eigen src fixedpoint fixedpointtypes h tensorflow ios framework tensorflow framework headers third party unsupported eigen src fixedpoint fixedpointtypes h vulnerability details in affected versions of tensorflow under certain cases a saved model can trigger use of uninitialized values during code execution this is caused by having tensor buffers be filled with the default value of the type but forgetting to default initialize the quantized floating point types in eigen this is fixed in versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu | 0 |
45,112 | 18,412,318,165 | IssuesEvent | 2021-10-13 07:36:31 | craftbuy/uptime | https://api.github.com/repos/craftbuy/uptime | closed | 🛑 CraftBuy.ru Services is down | status craft-buy-ru-services | In [`3e9b722`](https://github.com/craftbuy/uptime/commit/3e9b722fda208fd98fe607008fcad6a54eb85567
), CraftBuy.ru Services ($CRAFTBUY_SERVICES) was **down**:
- HTTP code: 0
- Response time: 0 ms
| 1.0 | 🛑 CraftBuy.ru Services is down - In [`3e9b722`](https://github.com/craftbuy/uptime/commit/3e9b722fda208fd98fe607008fcad6a54eb85567
), CraftBuy.ru Services ($CRAFTBUY_SERVICES) was **down**:
- HTTP code: 0
- Response time: 0 ms
| non_perf | 🛑 craftbuy ru services is down in craftbuy ru services craftbuy services was down http code response time ms | 0 |
234,633 | 25,880,477,572 | IssuesEvent | 2022-12-14 10:54:45 | rsoreq/WebGoat | https://api.github.com/repos/rsoreq/WebGoat | closed | CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - autoclosed | security vulnerability | ## CVE-2020-7733 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p>
<p>Path to dependency file: /docs/package.json</p>
<p>Path to vulnerable library: /docs/node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA.
<p>Publish Date: 2020-09-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7733>CVE-2020-7733</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p>
<p>Release Date: 2020-09-16</p>
<p>Fix Resolution (ua-parser-js): 0.7.22</p>
<p>Direct dependency fix Resolution (browser-sync): 2.26.9</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | True | CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - autoclosed - ## CVE-2020-7733 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p>
<p>Path to dependency file: /docs/package.json</p>
<p>Path to vulnerable library: /docs/node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA.
<p>Publish Date: 2020-09-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7733>CVE-2020-7733</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p>
<p>Release Date: 2020-09-16</p>
<p>Fix Resolution (ua-parser-js): 0.7.22</p>
<p>Direct dependency fix Resolution (browser-sync): 2.26.9</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | non_perf | cve high detected in ua parser js tgz autoclosed cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file docs package json path to vulnerable library docs node modules ua parser js package json dependency hierarchy browser sync tgz root library x ua parser js tgz vulnerable library found in base branch develop vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos via the regex for redmi phones and mi pad tablets ua publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ua parser js direct dependency fix resolution browser sync rescue worker helmet automatic remediation is available for this issue | 0 |
58,413 | 8,257,538,104 | IssuesEvent | 2018-09-13 05:42:12 | Microsoft/pxt-microbit | https://api.github.com/repos/Microsoft/pxt-microbit | closed | "Tutorials" & "Games" except "Examples" don't have brief introduction on home page | documentation homescreen | **Repro steps:**
1. Navigate to https://makecode.microbit.org/beta
2. Scroll down the scroll bar and click examples under "Tutorials" & "Games" one by one
3. Observe the opened window
**Expected result:**
Each tutorial and game should have a brief introduction

**Actual result:**
"Tutorials" & "Games" except "Examples" don't have brief introduction

| 1.0 | "Tutorials" & "Games" except "Examples" don't have brief introduction on home page - **Repro steps:**
1. Navigate to https://makecode.microbit.org/beta
2. Scroll down the scroll bar and click examples under "Tutorials" & "Games" one by one
3. Observe the opened window
**Expected result:**
Each tutorial and game should have a brief introduction

**Actual result:**
"Tutorials" & "Games" except "Examples" don't have brief introduction

| non_perf | tutorials games except examples don t have brief introduction on home page repro steps navigate to scroll down the scroll bar and click examples under tutorials games one by one observe the opened window expected result each tutorial and game should have a brief introduction actual result tutorials games except examples don t have brief introduction | 0 |
137,986 | 18,769,555,189 | IssuesEvent | 2021-11-06 15:30:31 | samqws-marketing/box_box-ui-elements | https://api.github.com/repos/samqws-marketing/box_box-ui-elements | opened | WS-2019-0339 (High) detected in bin-links-1.1.3.tgz, bin-links-1.1.2.tgz | security vulnerability | ## WS-2019-0339 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bin-links-1.1.3.tgz</b>, <b>bin-links-1.1.2.tgz</b></p></summary>
<p>
<details><summary><b>bin-links-1.1.3.tgz</b></p></summary>
<p>JavaScript package binary linker</p>
<p>Library home page: <a href="https://registry.npmjs.org/bin-links/-/bin-links-1.1.3.tgz">https://registry.npmjs.org/bin-links/-/bin-links-1.1.3.tgz</a></p>
<p>Path to dependency file: box_box-ui-elements/package.json</p>
<p>Path to vulnerable library: box_box-ui-elements/node_modules/bin-links/package.json</p>
<p>
Dependency Hierarchy:
- semantic-release-16.0.2.tgz (Root Library)
- npm-6.0.0.tgz
- npm-6.13.1.tgz
- :x: **bin-links-1.1.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>bin-links-1.1.2.tgz</b></p></summary>
<p>JavaScript package binary linker</p>
<p>Library home page: <a href="https://registry.npmjs.org/bin-links/-/bin-links-1.1.2.tgz">https://registry.npmjs.org/bin-links/-/bin-links-1.1.2.tgz</a></p>
<p>Path to dependency file: box_box-ui-elements/package.json</p>
<p>Path to vulnerable library: box_box-ui-elements/node_modules/bin-links/package.json</p>
<p>
Dependency Hierarchy:
- storybook-5.3.9.tgz (Root Library)
- cli-5.3.9.tgz
- pkg-add-deps-0.1.0.tgz
- libnpm-2.0.1.tgz
- :x: **bin-links-1.1.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In bin-links, versions prior to v1.1.6 are vulnerable to a Global 'node_modules' Binary Overwrite. It fails to prevent globally-installed binaries to be overwritten by other package installs.
<p>Publish Date: 2019-12-11
<p>URL: <a href=https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69>WS-2019-0339</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69">https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69</a></p>
<p>Release Date: 2019-12-17</p>
<p>Fix Resolution: bin-links - 1.1.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"bin-links","packageVersion":"1.1.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"semantic-release:16.0.2;@semantic-release/npm:6.0.0;npm:6.13.1;bin-links:1.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bin-links - 1.1.6"},{"packageType":"javascript/Node.js","packageName":"bin-links","packageVersion":"1.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"storybook:5.3.9;@storybook/cli:5.3.9;pkg-add-deps:0.1.0;libnpm:2.0.1;bin-links:1.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bin-links - 1.1.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0339","vulnerabilityDetails":"In bin-links, versions prior to v1.1.6 are vulnerable to a Global \u0027node_modules\u0027 Binary Overwrite. It fails to prevent globally-installed binaries to be overwritten by other package installs.","vulnerabilityUrl":"https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | WS-2019-0339 (High) detected in bin-links-1.1.3.tgz, bin-links-1.1.2.tgz - ## WS-2019-0339 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bin-links-1.1.3.tgz</b>, <b>bin-links-1.1.2.tgz</b></p></summary>
<p>
<details><summary><b>bin-links-1.1.3.tgz</b></p></summary>
<p>JavaScript package binary linker</p>
<p>Library home page: <a href="https://registry.npmjs.org/bin-links/-/bin-links-1.1.3.tgz">https://registry.npmjs.org/bin-links/-/bin-links-1.1.3.tgz</a></p>
<p>Path to dependency file: box_box-ui-elements/package.json</p>
<p>Path to vulnerable library: box_box-ui-elements/node_modules/bin-links/package.json</p>
<p>
Dependency Hierarchy:
- semantic-release-16.0.2.tgz (Root Library)
- npm-6.0.0.tgz
- npm-6.13.1.tgz
- :x: **bin-links-1.1.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>bin-links-1.1.2.tgz</b></p></summary>
<p>JavaScript package binary linker</p>
<p>Library home page: <a href="https://registry.npmjs.org/bin-links/-/bin-links-1.1.2.tgz">https://registry.npmjs.org/bin-links/-/bin-links-1.1.2.tgz</a></p>
<p>Path to dependency file: box_box-ui-elements/package.json</p>
<p>Path to vulnerable library: box_box-ui-elements/node_modules/bin-links/package.json</p>
<p>
Dependency Hierarchy:
- storybook-5.3.9.tgz (Root Library)
- cli-5.3.9.tgz
- pkg-add-deps-0.1.0.tgz
- libnpm-2.0.1.tgz
- :x: **bin-links-1.1.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In bin-links, versions prior to v1.1.6 are vulnerable to a Global 'node_modules' Binary Overwrite. It fails to prevent globally-installed binaries to be overwritten by other package installs.
<p>Publish Date: 2019-12-11
<p>URL: <a href=https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69>WS-2019-0339</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69">https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69</a></p>
<p>Release Date: 2019-12-17</p>
<p>Fix Resolution: bin-links - 1.1.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"bin-links","packageVersion":"1.1.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"semantic-release:16.0.2;@semantic-release/npm:6.0.0;npm:6.13.1;bin-links:1.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bin-links - 1.1.6"},{"packageType":"javascript/Node.js","packageName":"bin-links","packageVersion":"1.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"storybook:5.3.9;@storybook/cli:5.3.9;pkg-add-deps:0.1.0;libnpm:2.0.1;bin-links:1.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bin-links - 1.1.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0339","vulnerabilityDetails":"In bin-links, versions prior to v1.1.6 are vulnerable to a Global \u0027node_modules\u0027 Binary Overwrite. It fails to prevent globally-installed binaries to be overwritten by other package installs.","vulnerabilityUrl":"https://github.com/npm/bin-links/commit/642cd1890d1e0ca32b5f9ceadc6a587d94a63d69","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_perf | ws high detected in bin links tgz bin links tgz ws high severity vulnerability vulnerable libraries bin links tgz bin links tgz bin links tgz javascript package binary linker library home page a href path to dependency file box box ui elements package json path to vulnerable library box box ui elements node modules bin links package json dependency hierarchy semantic release tgz root library npm tgz npm tgz x bin links tgz vulnerable library bin links tgz javascript package binary linker library home page a href path to dependency file box box ui elements package json path to vulnerable library box box ui elements node modules bin links package json dependency hierarchy storybook tgz root library cli tgz pkg add deps tgz libnpm tgz x bin links tgz vulnerable library found in head commit a href found in base branch master vulnerability details in bin links versions prior to are vulnerable to a global node modules binary overwrite it fails to prevent globally installed binaries to be overwritten by other package installs publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bin links isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree semantic release semantic release npm npm bin links isminimumfixversionavailable true minimumfixversion bin links packagetype javascript node js packagename bin links packageversion packagefilepaths istransitivedependency true dependencytree storybook storybook cli pkg add deps libnpm bin links isminimumfixversionavailable true minimumfixversion bin links basebranches vulnerabilityidentifier ws vulnerabilitydetails in bin links versions prior to are vulnerable to a global modules binary overwrite it fails to prevent globally installed binaries to be overwritten by other package installs vulnerabilityurl | 0 |
7,613 | 3,105,356,852 | IssuesEvent | 2015-08-31 20:32:56 | softlayer/sl-ember-components | https://api.github.com/repos/softlayer/sl-ember-components | opened | Evaluate "selectionDirection" property | 0 - Backlog documentation question research sl-textarea | The `selectionDirection` attribute of a `textarea` element can be set but according to the spec - http://www.w3.org/TR/html5/forms.html#dom-textarea/input-selectiondirection - this value does not actually do anything from a "setting" perspective, though it does contain a value during text selection to be used in a "getter" perspective. Do we want to adjust the description of this property at #/demos/sl-textarea to align with this?
This then leads into the next item which is this attribute is set inconsistently across browsers. Chrome on OS X at least does not have this value populated. Given this (and other browsers should be tested too) do we even want to expose this property? What role do we want this library to play? Normalize across things or provide support pass-though and those environments that do will?
| 1.0 | Evaluate "selectionDirection" property - The `selectionDirection` attribute of a `textarea` element can be set but according to the spec - http://www.w3.org/TR/html5/forms.html#dom-textarea/input-selectiondirection - this value does not actually do anything from a "setting" perspective, though it does contain a value during text selection to be used in a "getter" perspective. Do we want to adjust the description of this property at #/demos/sl-textarea to align with this?
This then leads into the next item which is this attribute is set inconsistently across browsers. Chrome on OS X at least does not have this value populated. Given this (and other browsers should be tested too) do we even want to expose this property? What role do we want this library to play? Normalize across things or provide support pass-though and those environments that do will?
| non_perf | evaluate selectiondirection property the selectiondirection attribute of a textarea element can be set but according to the spec this value does not actually do anything from a setting perspective though it does contain a value during text selection to be used in a getter perspective do we want to adjust the description of this property at demos sl textarea to align with this this then leads into the next item which is this attribute is set inconsistently across browsers chrome on os x at least does not have this value populated given this and other browsers should be tested too do we even want to expose this property what role do we want this library to play normalize across things or provide support pass though and those environments that do will | 0 |
44,759 | 23,757,210,725 | IssuesEvent | 2022-09-01 05:07:47 | datafuselabs/databend | https://api.github.com/repos/datafuselabs/databend | closed | optimizer_statistics_exact is missing in new planner | C-performance A-planner | **Summary**
`optimizer_statistics_exact` works in old planner, but it's missing in new planner | True | optimizer_statistics_exact is missing in new planner - **Summary**
`optimizer_statistics_exact` works in old planner, but it's missing in new planner | perf | optimizer statistics exact is missing in new planner summary optimizer statistics exact works in old planner but it s missing in new planner | 1 |
222,274 | 24,697,454,546 | IssuesEvent | 2022-10-19 13:09:09 | rsoreq/kendo-ui-core | https://api.github.com/repos/rsoreq/kendo-ui-core | opened | CVE-2022-3517 (High) detected in multiple libraries | security vulnerability | ## CVE-2022-3517 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimatch-0.2.14.tgz</b>, <b>minimatch-2.0.10.tgz</b>, <b>minimatch-3.0.4.tgz</b>, <b>minimatch-1.0.0.tgz</b>, <b>minimatch-0.3.0.tgz</b></p></summary>
<p>
<details><summary><b>minimatch-0.2.14.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.2.14.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.2.14.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/globule/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- amd-optimize-0.6.1.tgz (Root Library)
- vinyl-fs-0.3.14.tgz
- glob-watcher-0.0.6.tgz
- gaze-0.5.2.tgz
- globule-0.1.0.tgz
- :x: **minimatch-0.2.14.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimatch-2.0.10.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-2.0.10.tgz">https://registry.npmjs.org/minimatch/-/minimatch-2.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- glob-5.0.15.tgz (Root Library)
- :x: **minimatch-2.0.10.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimatch-3.0.4.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/karma/node_modules/minimatch/package.json,/node_modules/accord/node_modules/minimatch/package.json,/docs-aspnet/node_modules/minimatch/package.json,/node_modules/gulp-match/node_modules/minimatch/package.json,/node_modules/mocha/node_modules/minimatch/package.json,/docs/node_modules/minimatch/package.json,/node_modules/rimraf/node_modules/minimatch/package.json,/node_modules/multimatch/node_modules/minimatch/package.json,/node_modules/gulp-jshint/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- gulp-if-2.0.2.tgz (Root Library)
- gulp-match-1.1.0.tgz
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimatch-1.0.0.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-1.0.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-1.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/gulp-ignore/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- gulp-ignore-1.2.1.tgz (Root Library)
- gulp-match-0.2.1.tgz
- :x: **minimatch-1.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimatch-0.3.0.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/gulp-subset-process/node_modules/minimatch/package.json,/node_modules/cli/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- jshint-2.8.0.tgz (Root Library)
- cli-0.6.6.tgz
- glob-3.2.11.tgz
- :x: **minimatch-0.3.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/kendo-ui-core/commit/62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad">62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
</details>
<p></p>
| True | CVE-2022-3517 (High) detected in multiple libraries - ## CVE-2022-3517 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimatch-0.2.14.tgz</b>, <b>minimatch-2.0.10.tgz</b>, <b>minimatch-3.0.4.tgz</b>, <b>minimatch-1.0.0.tgz</b>, <b>minimatch-0.3.0.tgz</b></p></summary>
<p>
<details><summary><b>minimatch-0.2.14.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.2.14.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.2.14.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/globule/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- amd-optimize-0.6.1.tgz (Root Library)
- vinyl-fs-0.3.14.tgz
- glob-watcher-0.0.6.tgz
- gaze-0.5.2.tgz
- globule-0.1.0.tgz
- :x: **minimatch-0.2.14.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimatch-2.0.10.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-2.0.10.tgz">https://registry.npmjs.org/minimatch/-/minimatch-2.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- glob-5.0.15.tgz (Root Library)
- :x: **minimatch-2.0.10.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimatch-3.0.4.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/karma/node_modules/minimatch/package.json,/node_modules/accord/node_modules/minimatch/package.json,/docs-aspnet/node_modules/minimatch/package.json,/node_modules/gulp-match/node_modules/minimatch/package.json,/node_modules/mocha/node_modules/minimatch/package.json,/docs/node_modules/minimatch/package.json,/node_modules/rimraf/node_modules/minimatch/package.json,/node_modules/multimatch/node_modules/minimatch/package.json,/node_modules/gulp-jshint/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- gulp-if-2.0.2.tgz (Root Library)
- gulp-match-1.1.0.tgz
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimatch-1.0.0.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-1.0.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-1.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/gulp-ignore/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- gulp-ignore-1.2.1.tgz (Root Library)
- gulp-match-0.2.1.tgz
- :x: **minimatch-1.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimatch-0.3.0.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/gulp-subset-process/node_modules/minimatch/package.json,/node_modules/cli/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- jshint-2.8.0.tgz (Root Library)
- cli-0.6.6.tgz
- glob-3.2.11.tgz
- :x: **minimatch-0.3.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/kendo-ui-core/commit/62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad">62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
</details>
<p></p>
| non_perf | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries minimatch tgz minimatch tgz minimatch tgz minimatch tgz minimatch tgz minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules globule node modules minimatch package json dependency hierarchy amd optimize tgz root library vinyl fs tgz glob watcher tgz gaze tgz globule tgz x minimatch tgz vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules minimatch package json dependency hierarchy glob tgz root library x minimatch tgz vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules karma node modules minimatch package json node modules accord node modules minimatch package json docs aspnet node modules minimatch package json node modules gulp match node modules minimatch package json node modules mocha node modules minimatch package json docs node modules minimatch package json node modules rimraf node modules minimatch package json node modules multimatch node modules minimatch package json node modules gulp jshint node modules minimatch package json dependency hierarchy gulp if tgz root library gulp match tgz x minimatch tgz vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules gulp ignore node modules minimatch package json dependency hierarchy gulp ignore tgz root library gulp match tgz x minimatch tgz vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules gulp subset process node modules minimatch package json node modules cli node modules minimatch package json dependency hierarchy jshint tgz root library cli tgz glob tgz x minimatch tgz vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability was found in the minimatch package this flaw allows a regular expression denial of service redos when calling the braceexpand function with specific arguments resulting in a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimatch | 0 |
91,897 | 18,737,234,364 | IssuesEvent | 2021-11-04 09:17:31 | cosmos/ibc-go | https://api.github.com/repos/cosmos/ibc-go | closed | Nitpicks: ICA Audit (TrySendTxFlow) | good first issue code-hygiene 27-interchain-accounts audit-ica | <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Please also ensure that this is not a duplicate issue :)
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary of Issue
The following nits arose as part of the audit.
- [x] TrySendTx (relay.go) channel not found should return for which port id
- [ ] ~~Return nil instead of []byte{} in all returns for keeper/relay.go~~
- [x] Require that we pass in an array of sdk.Msg instead of single sdk.Msg (keeper/keeper.go)
- [ ] ~~Add channel/port id in capability not found error (createOutgoingPacket - keeper/relay.go)~~
- [x] Add channel/port id in get next send sequence not found error (createOutgoingPacket - keeper/relay.go)
- [x] Potentially panic on errors that indicate bugs in code? (keeper/keeper.go)
- [x] ErrUnkownPacketData - ErrUnknownDataType (errors.go)
- [x] relay.go: AuthenticateTx: Get Interchain address first then loop through signers and return an error if the expected signer is not ICA address
- [x] relay.go: ExectureTx reduce code by returning when error occurs on executeMsg. Add comment for how cache context is functioning (atomic execution)
- [ ] ~~Fix error type and wrapping in module.go AcknowledgePacket~~
____
#### For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate contributors tagged/assigned
| 1.0 | Nitpicks: ICA Audit (TrySendTxFlow) - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Please also ensure that this is not a duplicate issue :)
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary of Issue
The following nits arose as part of the audit.
- [x] TrySendTx (relay.go) channel not found should return for which port id
- [ ] ~~Return nil instead of []byte{} in all returns for keeper/relay.go~~
- [x] Require that we pass in an array of sdk.Msg instead of single sdk.Msg (keeper/keeper.go)
- [ ] ~~Add channel/port id in capability not found error (createOutgoingPacket - keeper/relay.go)~~
- [x] Add channel/port id in get next send sequence not found error (createOutgoingPacket - keeper/relay.go)
- [x] Potentially panic on errors that indicate bugs in code? (keeper/keeper.go)
- [x] ErrUnkownPacketData - ErrUnknownDataType (errors.go)
- [x] relay.go: AuthenticateTx: Get Interchain address first then loop through signers and return an error if the expected signer is not ICA address
- [x] relay.go: ExectureTx reduce code by returning when error occurs on executeMsg. Add comment for how cache context is functioning (atomic execution)
- [ ] ~~Fix error type and wrapping in module.go AcknowledgePacket~~
____
#### For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate contributors tagged/assigned
| non_perf | nitpicks ica audit trysendtxflow ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ summary of issue the following nits arose as part of the audit trysendtx relay go channel not found should return for which port id return nil instead of byte in all returns for keeper relay go require that we pass in an array of sdk msg instead of single sdk msg keeper keeper go add channel port id in capability not found error createoutgoingpacket keeper relay go add channel port id in get next send sequence not found error createoutgoingpacket keeper relay go potentially panic on errors that indicate bugs in code keeper keeper go errunkownpacketdata errunknowndatatype errors go relay go authenticatetx get interchain address first then loop through signers and return an error if the expected signer is not ica address relay go execturetx reduce code by returning when error occurs on executemsg add comment for how cache context is functioning atomic execution fix error type and wrapping in module go acknowledgepacket for admin use not duplicate issue appropriate labels applied appropriate contributors tagged assigned | 0 |
221,732 | 17,026,196,841 | IssuesEvent | 2021-07-03 15:11:54 | WebDevJ/ski | https://api.github.com/repos/WebDevJ/ski | closed | Update README | documentation | * Comments; what was done, what wasn't, features added & known bugs If any.
* Provide a way to view the completed code and run it; local environment install/run instructions. | 1.0 | Update README - * Comments; what was done, what wasn't, features added & known bugs If any.
* Provide a way to view the completed code and run it; local environment install/run instructions. | non_perf | update readme comments what was done what wasn t features added known bugs if any provide a way to view the completed code and run it local environment install run instructions | 0 |
44,483 | 23,651,188,800 | IssuesEvent | 2022-08-26 06:49:28 | clj-commons/aleph | https://api.github.com/repos/clj-commons/aleph | closed | Improve wrap-future | performance good first issue | This was uncovered in the process of reviewing #608. See my comments there for more.
In short, `wrap-future` checks `.isSuccess` on a future, when it really means to check `.isDone`. If it succeeded, it creates an immediate success-deferred. If not, it attaches handlers. In the case of a failed/canceled future, it does a lot of unnecessary work before immediately invoking the handler: fn thread frame binding, handler object creation, and scheduling the handlers to run (potentially) on another thread. This is also kind of confusing because we're checking the wrong condition.
To do:
- [ ] Switch to `.isDone` from `.isSuccess` in `wrap-future`
- [ ] If `.isDone`, choose between returning success- and error-deferreds based on the future's status. For consistency, we can reuse the `operation-complete` handler, but add a flag to tell it to use/ignore executors (so we can run it on the current thread)
- [ ] Verify all tests still work, `test-classloader` in particular | True | Improve wrap-future - This was uncovered in the process of reviewing #608. See my comments there for more.
In short, `wrap-future` checks `.isSuccess` on a future, when it really means to check `.isDone`. If it succeeded, it creates an immediate success-deferred. If not, it attaches handlers. In the case of a failed/canceled future, it does a lot of unnecessary work before immediately invoking the handler: fn thread frame binding, handler object creation, and scheduling the handlers to run (potentially) on another thread. This is also kind of confusing because we're checking the wrong condition.
To do:
- [ ] Switch to `.isDone` from `.isSuccess` in `wrap-future`
- [ ] If `.isDone`, choose between returning success- and error-deferreds based on the future's status. For consistency, we can reuse the `operation-complete` handler, but add a flag to tell it to use/ignore executors (so we can run it on the current thread)
- [ ] Verify all tests still work, `test-classloader` in particular | perf | improve wrap future this was uncovered in the process of reviewing see my comments there for more in short wrap future checks issuccess on a future when it really means to check isdone if it succeeded it creates an immediate success deferred if not it attaches handlers in the case of a failed canceled future it does a lot of unnecessary work before immediately invoking the handler fn thread frame binding handler object creation and scheduling the handlers to run potentially on another thread this is also kind of confusing because we re checking the wrong condition to do switch to isdone from issuccess in wrap future if isdone choose between returning success and error deferreds based on the future s status for consistency we can reuse the operation complete handler but add a flag to tell it to use ignore executors so we can run it on the current thread verify all tests still work test classloader in particular | 1 |
121,419 | 17,651,781,223 | IssuesEvent | 2021-08-20 14:07:12 | snowdensb/spring-security-oauth | https://api.github.com/repos/snowdensb/spring-security-oauth | opened | CVE-2020-10673 (High) detected in jackson-databind-2.8.11.3.jar, jackson-databind-2.6.6.jar | security vulnerability | ## CVE-2020-10673 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.11.3.jar</b>, <b>jackson-databind-2.6.6.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.11.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.22.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.6/jackson-databind-2.6.6.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.3.5.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.6.6.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/spring-security-oauth/commit/62c6485ab8fc644a76523cfb55746e549555797c">62c6485ab8fc644a76523cfb55746e549555797c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.3","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.22.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.6","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.3.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-10673","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-10673 (High) detected in jackson-databind-2.8.11.3.jar, jackson-databind-2.6.6.jar - ## CVE-2020-10673 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.11.3.jar</b>, <b>jackson-databind-2.6.6.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.11.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.22.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.6/jackson-databind-2.6.6.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.3.5.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.6.6.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/spring-security-oauth/commit/62c6485ab8fc644a76523cfb55746e549555797c">62c6485ab8fc644a76523cfb55746e549555797c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.3","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.22.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.6","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.3.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-10673","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_perf | cve high detected in jackson databind jar jackson databind jar cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus vulnerabilityurl | 0 |
211,511 | 16,447,875,419 | IssuesEvent | 2021-05-20 22:16:20 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | Minor documentation bugs in quick-start section | area/documentation | Quick Start -> Java -> YCQL 4.6 section
POM file section is missing Java source and target versions
Fix whitespace in dependencies section
Quick Start -> Java -> YSQL JDBC section
Maven command needs additional target: package
Quick Start -> Java -> YSQL Spring Data section
Fix path for source
| 1.0 | Minor documentation bugs in quick-start section - Quick Start -> Java -> YCQL 4.6 section
POM file section is missing Java source and target versions
Fix whitespace in dependencies section
Quick Start -> Java -> YSQL JDBC section
Maven command needs additional target: package
Quick Start -> Java -> YSQL Spring Data section
Fix path for source
| non_perf | minor documentation bugs in quick start section quick start java ycql section pom file section is missing java source and target versions fix whitespace in dependencies section quick start java ysql jdbc section maven command needs additional target package quick start java ysql spring data section fix path for source | 0 |
36,382 | 17,650,209,064 | IssuesEvent | 2021-08-20 12:11:31 | socialsensingbot/frontend | https://api.github.com/repos/socialsensingbot/frontend | opened | Export Tweets to CSV is inefficent | internal performance | This can now be done entirely in SQL with a single call to the REST API. This should be done prior to 2.2 release. | True | Export Tweets to CSV is inefficent - This can now be done entirely in SQL with a single call to the REST API. This should be done prior to 2.2 release. | perf | export tweets to csv is inefficent this can now be done entirely in sql with a single call to the rest api this should be done prior to release | 1 |
314,482 | 9,597,762,876 | IssuesEvent | 2019-05-09 22:22:55 | CypherpunkArmory/UserLAnd | https://api.github.com/repos/CypherpunkArmory/UserLAnd | closed | Open Ubuntu Lxde but Debian starts | app request enhancement low priority | Samsung Note 9 installed Debian first with lxde and started using "start lxde &"
At some point starting debian began opening lxde automatically (all tight vnc).
(In the past Used to use gnu root with xsdl but i see it is not supported and tight vnc works fine and more easily. )
I then installed ubuntu file system with xfce4 and started it using command line to start xfce i forget the command something like startxfce4& or something. It worked but firefox was black.
At some point the lxde and xfce showed up on userland home screen.
I deleted ubuntu and reinstalled with lxde to see if firefox would run. Now i cannot get ubuntu to run lxde gui.
When i click lxde or xfce from userland menu only debian opens. IS there a way to select the file system that will be opened with lxde or xfce userland home button?
IS there a way to start lxde in ubuntu command line?
Is installing lxde/xfce in terminal still necessary with the home screen buttons? O notice if i click xfce debian will run in xfce even tho i installed only lxde in debian command line in the beginning.
Thank you for your time.
| 1.0 | Open Ubuntu Lxde but Debian starts - Samsung Note 9 installed Debian first with lxde and started using "start lxde &"
At some point starting debian began opening lxde automatically (all tight vnc).
(In the past Used to use gnu root with xsdl but i see it is not supported and tight vnc works fine and more easily. )
I then installed ubuntu file system with xfce4 and started it using command line to start xfce i forget the command something like startxfce4& or something. It worked but firefox was black.
At some point the lxde and xfce showed up on userland home screen.
I deleted ubuntu and reinstalled with lxde to see if firefox would run. Now i cannot get ubuntu to run lxde gui.
When i click lxde or xfce from userland menu only debian opens. IS there a way to select the file system that will be opened with lxde or xfce userland home button?
IS there a way to start lxde in ubuntu command line?
Is installing lxde/xfce in terminal still necessary with the home screen buttons? O notice if i click xfce debian will run in xfce even tho i installed only lxde in debian command line in the beginning.
Thank you for your time.
| non_perf | open ubuntu lxde but debian starts samsung note installed debian first with lxde and started using start lxde at some point starting debian began opening lxde automatically all tight vnc in the past used to use gnu root with xsdl but i see it is not supported and tight vnc works fine and more easily i then installed ubuntu file system with and started it using command line to start xfce i forget the command something like or something it worked but firefox was black at some point the lxde and xfce showed up on userland home screen i deleted ubuntu and reinstalled with lxde to see if firefox would run now i cannot get ubuntu to run lxde gui when i click lxde or xfce from userland menu only debian opens is there a way to select the file system that will be opened with lxde or xfce userland home button is there a way to start lxde in ubuntu command line is installing lxde xfce in terminal still necessary with the home screen buttons o notice if i click xfce debian will run in xfce even tho i installed only lxde in debian command line in the beginning thank you for your time | 0 |
153,713 | 24,175,424,453 | IssuesEvent | 2022-09-23 00:44:11 | enigmalea/enigmalea | https://api.github.com/repos/enigmalea/enigmalea | closed | Explore new color palettes | design | I'm not in love with the current color palettes for either light or dark mode. I need to explore more options. | 1.0 | Explore new color palettes - I'm not in love with the current color palettes for either light or dark mode. I need to explore more options. | non_perf | explore new color palettes i m not in love with the current color palettes for either light or dark mode i need to explore more options | 0 |
23,353 | 4,932,397,177 | IssuesEvent | 2016-11-28 13:33:41 | Jumpscale/jscockpit | https://api.github.com/repos/Jumpscale/jscockpit | opened | cockpit-doc: Instances Walkthrough | type_documentation type_feature | ## GOAL:
Explain what the **Instances** page is all about
## DESCRIPTION:
Placeholder: https://github.com/Jumpscale/jscockpit/blob/master/docs/walkthrough/Instances/Instances.md | 1.0 | cockpit-doc: Instances Walkthrough - ## GOAL:
Explain what the **Instances** page is all about
## DESCRIPTION:
Placeholder: https://github.com/Jumpscale/jscockpit/blob/master/docs/walkthrough/Instances/Instances.md | non_perf | cockpit doc instances walkthrough goal explain what the instances page is all about description placeholder | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.