Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
40,204
5,282,141,166
IssuesEvent
2017-02-07 18:06:18
institutotim/timtec
https://api.github.com/repos/institutotim/timtec
closed
bug: ao sair do sistema com o usuário admin
bug waiting test
flow: - editar curso - minha área > sair usuário: admin analisar log
1.0
bug: ao sair do sistema com o usuário admin - flow: - editar curso - minha área > sair usuário: admin analisar log
test
bug ao sair do sistema com o usuário admin flow editar curso minha área sair usuário admin analisar log
1
550,901
16,134,260,929
IssuesEvent
2021-04-29 09:41:32
mozilla/addons-server
https://api.github.com/repos/mozilla/addons-server
opened
don't include sitemap eula/privacy pages for empty policies
priority: p3
follow-up for #16907 the sitemap adds a page for each addon that has a `.eula` or `.privacy_policy` set. As these are translated fields the code currently just checks if the field_id has a non-`None` value, i.e. it has been set. This turned out to be a bit naive as we have less than ideal data in a our database - some of the translations are set to `''`. (In an ideal world translations instances of '' wouldn't exist. It's near-impossible to get an accurate number of how many eula or privacy policy specifically this has happened for, but addons-dev has 1.6M translations set to '', for all translated fields, including some old deleted content)
1.0
don't include sitemap eula/privacy pages for empty policies - follow-up for #16907 the sitemap adds a page for each addon that has a `.eula` or `.privacy_policy` set. As these are translated fields the code currently just checks if the field_id has a non-`None` value, i.e. it has been set. This turned out to be a bit naive as we have less than ideal data in a our database - some of the translations are set to `''`. (In an ideal world translations instances of '' wouldn't exist. It's near-impossible to get an accurate number of how many eula or privacy policy specifically this has happened for, but addons-dev has 1.6M translations set to '', for all translated fields, including some old deleted content)
non_test
don t include sitemap eula privacy pages for empty policies follow up for the sitemap adds a page for each addon that has a eula or privacy policy set as these are translated fields the code currently just checks if the field id has a non none value i e it has been set this turned out to be a bit naive as we have less than ideal data in a our database some of the translations are set to in an ideal world translations instances of wouldn t exist it s near impossible to get an accurate number of how many eula or privacy policy specifically this has happened for but addons dev has translations set to for all translated fields including some old deleted content
0
76,843
14,688,372,746
IssuesEvent
2021-01-02 02:17:22
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] Media Manager UX/UI - resizing/jumping of content
No Code Attached Yet
### Steps to reproduce the issue Go to **Media Manager** - **Resize** tab, click to change the width and change the width interactively. ### Expected result The source image stays in some fixed box (size will be changed inside the box) and the content outside the source image will not change - will be not moved, will not jump, etc. ### Actual result Now, all the content moves and jumps, see GIF: ![img](https://i.imgur.com/785Dw8W.gif) https://i.imgur.com/YhYqXfu.gif ### System information (as much as possible) Joomla! 4.0 ‎- 4.0.0-beta5 ### Additional comments
1.0
[4.0] Media Manager UX/UI - resizing/jumping of content - ### Steps to reproduce the issue Go to **Media Manager** - **Resize** tab, click to change the width and change the width interactively. ### Expected result The source image stays in some fixed box (size will be changed inside the box) and the content outside the source image will not change - will be not moved, will not jump, etc. ### Actual result Now, all the content moves and jumps, see GIF: ![img](https://i.imgur.com/785Dw8W.gif) https://i.imgur.com/YhYqXfu.gif ### System information (as much as possible) Joomla! 4.0 ‎- 4.0.0-beta5 ### Additional comments
non_test
media manager ux ui resizing jumping of content steps to reproduce the issue go to media manager resize tab click to change the width and change the width interactively expected result the source image stays in some fixed box size will be changed inside the box and the content outside the source image will not change will be not moved will not jump etc actual result now all the content moves and jumps see gif system information as much as possible joomla ‎ additional comments
0
125,714
26,717,502,059
IssuesEvent
2023-01-28 18:09:31
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
opened
Low Code UI: Missing field_pointer in the YAML
type/bug needs-triage low-code
<!-- Welcome to Airbyte! We appreciate your report and know that this will help us build an amazing tool. If you want to contribute yourself, you can find a good place to start by searching for the good-first-issues label or maybe... by trying to solve this one? (we can help debug this with you!) It's really important to have all the information and context. You can remove the examples below and fill out with your information. --> ## Environment Low Code UI 40.29 ## Current Behavior Setting a record selector/field pointer works in the UI, but it doesn't autogenerate in the browser-hosted YAML. However, if you download the YAML though it does appear in the config. 🤷 ## Expected Behavior Once you set the record selector in the UI it should show up in the YAML file. ## Logs ## Steps to Reproduce 1. Set the record selector 2. Check the YAML in the UI ## Are you willing to submit a PR? <!--- We accept contributions! Don't feel pressured, but if you want to contribute we can help you by giving some tips, highlighting the necessary code change or explaining any relevant point your feature will impact. You can also send questions on #dev Slack channel. We understand if you can't submit a PR and we're tremendously grateful that you've already contributed by suggesting a new feature. --> Remove this with your answer.
1.0
Low Code UI: Missing field_pointer in the YAML - <!-- Welcome to Airbyte! We appreciate your report and know that this will help us build an amazing tool. If you want to contribute yourself, you can find a good place to start by searching for the good-first-issues label or maybe... by trying to solve this one? (we can help debug this with you!) It's really important to have all the information and context. You can remove the examples below and fill out with your information. --> ## Environment Low Code UI 40.29 ## Current Behavior Setting a record selector/field pointer works in the UI, but it doesn't autogenerate in the browser-hosted YAML. However, if you download the YAML though it does appear in the config. 🤷 ## Expected Behavior Once you set the record selector in the UI it should show up in the YAML file. ## Logs ## Steps to Reproduce 1. Set the record selector 2. Check the YAML in the UI ## Are you willing to submit a PR? <!--- We accept contributions! Don't feel pressured, but if you want to contribute we can help you by giving some tips, highlighting the necessary code change or explaining any relevant point your feature will impact. You can also send questions on #dev Slack channel. We understand if you can't submit a PR and we're tremendously grateful that you've already contributed by suggesting a new feature. --> Remove this with your answer.
non_test
low code ui missing field pointer in the yaml welcome to airbyte we appreciate your report and know that this will help us build an amazing tool if you want to contribute yourself you can find a good place to start by searching for the good first issues label or maybe by trying to solve this one we can help debug this with you it s really important to have all the information and context you can remove the examples below and fill out with your information environment low code ui current behavior setting a record selector field pointer works in the ui but it doesn t autogenerate in the browser hosted yaml however if you download the yaml though it does appear in the config 🤷 expected behavior once you set the record selector in the ui it should show up in the yaml file logs steps to reproduce set the record selector check the yaml in the ui are you willing to submit a pr we accept contributions don t feel pressured but if you want to contribute we can help you by giving some tips highlighting the necessary code change or explaining any relevant point your feature will impact you can also send questions on dev slack channel we understand if you can t submit a pr and we re tremendously grateful that you ve already contributed by suggesting a new feature remove this with your answer
0
85,063
24,502,384,307
IssuesEvent
2022-10-10 13:46:28
spack/spack
https://api.github.com/repos/spack/spack
closed
hypre@2.25.0 +rocm fails: _hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found
build-error e4s ROCm/hip
### Steps to reproduce the issue Reproducible using: * `spack@develop` (4c2f357f0e8c5a2ce5609e654d5bfc435b6caee5 from `Mon Aug 15 18:00:33 2022 -0700`) * `ecpe4s/ubuntu22.04-runner-x86_64:2022-08-01` * `%gcc@11.2.0` * ROCm 5.2.0 Concrete spec: [hypre.spec.json.txt](https://github.com/spack/spack/files/9354617/hypre.spec.json.txt) <details><summary>Concretization</summary> <pre> - hypre@2.25.0%gcc@11.2.0~complex~cuda~debug+fortran~gptune~int64~internal-superlu~mixedint+mpi~openmp+rocm+shared~superlu-dist~unified-memory amdgpu_target=gfx90a arch=linux-ubuntu22.04-x86_64 [+] ^hip@5.2.0%gcc@11.2.0~ipo build_type=Release patches=959d1fe arch=linux-ubuntu22.04-x86_64 [+] ^cmake@3.23.3%gcc@11.2.0~doc+ncurses+ownlibs~qt build_type=Release arch=linux-ubuntu22.04-x86_64 [+] ^ncurses@6.2%gcc@11.2.0~symlinks+termlib abi=none arch=linux-ubuntu22.04-x86_64 [+] ^pkgconf@1.8.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^openssl@1.1.1q%gcc@11.2.0~docs~shared certs=mozilla patches=3fdcf2d arch=linux-ubuntu22.04-x86_64 [+] ^ca-certificates-mozilla@2022-07-19%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^perl@5.34.1%gcc@11.2.0+cpanm+shared+threads arch=linux-ubuntu22.04-x86_64 [+] ^berkeley-db@18.1.40%gcc@11.2.0+cxx~docs+stl patches=b231fcc arch=linux-ubuntu22.04-x86_64 [+] ^bzip2@1.0.8%gcc@11.2.0~debug~pic+shared arch=linux-ubuntu22.04-x86_64 [+] ^diffutils@3.8%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libiconv@1.16%gcc@11.2.0 libs=shared,static arch=linux-ubuntu22.04-x86_64 [+] ^gdbm@1.19%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^readline@8.1.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^zlib@1.2.12%gcc@11.2.0+optimize+pic+shared patches=0d38234 arch=linux-ubuntu22.04-x86_64 [+] ^comgr@5.2.0%gcc@11.2.0~ipo build_type=Release arch=linux-ubuntu22.04-x86_64 [+] ^llvm-amdgpu@5.2.0%gcc@11.2.0~ipo~link_llvm_dylib~llvm_dylib~openmp+rocm-device-libs build_type=Release patches=a08bbe1 arch=linux-ubuntu22.04-x86_64 [+] ^python@3.8.13%gcc@11.2.0+bz2+ctypes+dbm~debug+ensurepip+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3+ssl~tix~tkinter~ucs4+uuid+zlib patches=0d98e93,4c24573,f2fd060 arch=linux-ubuntu22.04-x86_64 [+] ^expat@2.4.8%gcc@11.2.0+libbsd arch=linux-ubuntu22.04-x86_64 [+] ^libbsd@0.11.5%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libmd@1.0.4%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^gettext@0.21%gcc@11.2.0+bzip2+curses+git~libunistring+libxml2+tar+xz arch=linux-ubuntu22.04-x86_64 [+] ^libxml2@2.9.13%gcc@11.2.0~python arch=linux-ubuntu22.04-x86_64 [+] ^xz@5.2.5%gcc@11.2.0+pic libs=shared,static arch=linux-ubuntu22.04-x86_64 [+] ^tar@1.34%gcc@11.2.0 zip=pigz arch=linux-ubuntu22.04-x86_64 [+] ^pigz@2.7%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^zstd@1.5.2%gcc@11.2.0+programs compression=none libs=shared,static arch=linux-ubuntu22.04-x86_64 [+] ^libffi@3.4.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^sqlite@3.38.5%gcc@11.2.0+column_metadata+dynamic_extensions+fts~functions+rtree arch=linux-ubuntu22.04-x86_64 [+] ^util-linux-uuid@2.37.4%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^z3@4.8.16%gcc@11.2.0~gmp~ipo~python build_type=RelWithDebInfo arch=linux-ubuntu22.04-x86_64 [+] ^rocm-cmake@5.2.0%gcc@11.2.0~ipo build_type=Release arch=linux-ubuntu22.04-x86_64 [+] ^glx@1.4%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^mesa@22.1.2%gcc@11.2.0+glx+llvm+opengl~opengles+osmesa~strip buildtype=release default_library=shared patches=ada85be arch=linux-ubuntu22.04-x86_64 [+] ^bison@3.8.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^m4@1.4.19%gcc@11.2.0+sigsegv patches=9dc5fbd,bfdffa7 arch=linux-ubuntu22.04-x86_64 [+] ^libsigsegv@2.13%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^flex@2.6.3%gcc@11.2.0+lex~nls arch=linux-ubuntu22.04-x86_64 [+] ^findutils@4.9.0%gcc@11.2.0 patches=440b954 arch=linux-ubuntu22.04-x86_64 [+] ^glproto@1.4.17%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^util-macros@1.19.3%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libunwind@1.6.2%gcc@11.2.0~block_signals~conservative_checks~cxx_exceptions~debug~debug_frame+docs+pic+tests+weak_backtrace+xz~zlib components=none libs=shared,static arch=linux-ubuntu22.04-x86_64 [+] ^libx11@1.7.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^inputproto@2.3.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^kbproto@1.0.7%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxcb@1.14%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libpthread-stubs@0.4%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxau@1.0.8%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xproto@7.0.31%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxdmcp@1.1.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xcb-proto@1.14.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xextproto@7.3.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xtrans@1.3.5%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxext@1.3.3%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxt@1.1.5%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libice@1.0.9%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libsm@1.2.3%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^meson@0.62.2%gcc@11.2.0 patches=9c87472,aa6c50d arch=linux-ubuntu22.04-x86_64 [+] ^ninja@1.11.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-pip@22.1.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-setuptools@63.0.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-wheel@0.37.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-mako@1.1.6%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-markupsafe@2.0.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xrandr@1.5.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxrandr@1.5.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxrender@0.9.10%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^renderproto@0.11.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^randrproto@1.5.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^hsa-rocr-dev@5.2.0%gcc@11.2.0+image~ipo+shared build_type=Release patches=71e6851 arch=linux-ubuntu22.04-x86_64 [+] ^elfutils@0.186%gcc@11.2.0+bzip2~debuginfod~nls+xz arch=linux-ubuntu22.04-x86_64 [+] ^hsakmt-roct@5.2.0%gcc@11.2.0~ipo+shared build_type=Release patches=f926273 arch=linux-ubuntu22.04-x86_64 [+] ^libdrm@2.4.110%gcc@11.2.0~docs arch=linux-ubuntu22.04-x86_64 [+] ^libpciaccess@0.16%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libtool@2.4.7%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^numactl@2.0.14%gcc@11.2.0 patches=4e1d78c,62fc8a8,ff37630 arch=linux-ubuntu22.04-x86_64 [+] ^autoconf@2.69%gcc@11.2.0 patches=35c4492,7793209,a49dd5b arch=linux-ubuntu22.04-x86_64 [+] ^automake@1.16.5%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xxd-standalone@8.2.1201%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^perl-file-which@1.22%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^perl-uri-encode@1.1.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^perl-module-build@0.4224%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^rocminfo@5.2.0%gcc@11.2.0~ipo build_type=Release arch=linux-ubuntu22.04-x86_64 [+] ^roctracer-dev-api@5.2.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^mpich@4.0.2%gcc@11.2.0~argobots~cuda+fortran+hwloc+hydra+libxml2+pci~rocm+romio~slurm~two_level_namespace~vci~verbs~wrapperrpath datatype-engine=auto device=ch4 netmod=ofi pmi=pmi arch=linux-ubuntu22.04-x86_64 [+] ^hwloc@2.8.0%gcc@11.2.0~cairo~cuda~gl~libudev+libxml2~netloc~nvml~oneapi-level-zero~opencl+pci~rocm+shared arch=linux-ubuntu22.04-x86_64 [+] ^libfabric@1.14.1%gcc@11.2.0~debug~kdreg fabrics=rxm,sockets,tcp,udp arch=linux-ubuntu22.04-x86_64 [+] ^yaksa@0.2%gcc@11.2.0~cuda~rocm arch=linux-ubuntu22.04-x86_64 [+] ^openblas@0.3.20%gcc@11.2.0~bignuma~consistent_fpcsr~ilp64+locking+pic+shared symbol_suffix=none threads=openmp arch=linux-ubuntu22.04-x86_64 </pre></details> Install error: ``` $> spack install -f ./hypre.spec.json ... ==> Installing hypre-2.25.0-6rsc2fdt7na5sqleojg3uug6klohx2f6 ==> No binary for hypre-2.25.0-6rsc2fdt7na5sqleojg3uug6klohx2f6 found: installing from source ==> Using cached archive: /spack/var/spack/cache/_source-cache/archive/f9/f9fc8371d91239fca694284dab17175bfda3821d7b7a871fd2e8f9d5930f303c.tar.gz ==> No patches needed for hypre ==> hypre: Executing phase: 'autoreconf' ==> hypre: Executing phase: 'configure' ==> hypre: Executing phase: 'build' ==> Error: ProcessError: Command exited with status 2: 'make' '-j16' 11 errors found in build log: 186 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c handle.c -o handle.ob j 187 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c int_array.c -o int_ar ray.obj 188 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c memory.c -o memory.ob j 189 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c omp_device.c -o omp_d evice.obj 190 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c nvtx.c -o nvtx.obj 191 In file included from device_utils.c:10: >> 192 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 193 #include <rocsparse.h> 194 ^~~~~~~~~~~~~ 195 In file included from general.c:9: >> 196 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 197 #include <rocsparse.h> 198 ^~~~~~~~~~~~~ 199 In file included from handle.c:15: >> 200 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 201 #include <rocsparse.h> 202 ^~~~~~~~~~~~~ 203 1 error generated when compiling for gfx90a. 204 1 error generated when compiling for gfx90a. 205 In file included from int_array.c:9: >> 206 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 207 #include <rocsparse.h> 208 ^~~~~~~~~~~~~ 209 1 error generated when compiling for gfx90a. >> 210 make[1]: *** [../config/Makefile.config:66: general.obj] Error 1 211 make[1]: *** Waiting for unfinished jobs.... >> 212 make[1]: *** [../config/Makefile.config:66: device_utils.obj] Error 1 213 1 error generated when compiling for gfx90a. >> 214 make[1]: *** [../config/Makefile.config:66: handle.obj] Error 1 >> 215 make[1]: *** [../config/Makefile.config:66: int_array.obj] Error 1 216 In file included from memory.c:15: >> 217 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 218 #include <rocsparse.h> 219 ^~~~~~~~~~~~~ 220 1 error generated when compiling for gfx90a. >> 221 make[1]: *** [../config/Makefile.config:66: memory.obj] Error 1 222 make[1]: Leaving directory '/tmp/root/spack-stage/spack-stage-hypre-2.25.0-6rsc2fdt7na5sqleojg3uug6klohx2f6/spack-src/src/utilities' >> 223 make: *** [Makefile:86: all] Error 1 ``` ### Error message <details><summary>Error message</summary><pre> ... see above </pre></details> ### Information on your system * **Spack:** 0.19.0.dev0 (9d5375a3ba150a3b216f06bc435b335b8f9bfc42) * **Python:** 3.10.4 * **Platform:** linux-ubuntu22.04-zen2 * **Concretizer:** clingo ### Additional information [spack-build-out.txt](https://github.com/spack/spack/files/9354620/spack-build-out.txt) [spack-build-env.txt](https://github.com/spack/spack/files/9354621/spack-build-env.txt) @balay @osborn9 @ulrikeyang @wspear ### General information - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded the build log and environment files - [X] I have searched the issues of this repo and believe this is not a duplicate
1.0
hypre@2.25.0 +rocm fails: _hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found - ### Steps to reproduce the issue Reproducible using: * `spack@develop` (4c2f357f0e8c5a2ce5609e654d5bfc435b6caee5 from `Mon Aug 15 18:00:33 2022 -0700`) * `ecpe4s/ubuntu22.04-runner-x86_64:2022-08-01` * `%gcc@11.2.0` * ROCm 5.2.0 Concrete spec: [hypre.spec.json.txt](https://github.com/spack/spack/files/9354617/hypre.spec.json.txt) <details><summary>Concretization</summary> <pre> - hypre@2.25.0%gcc@11.2.0~complex~cuda~debug+fortran~gptune~int64~internal-superlu~mixedint+mpi~openmp+rocm+shared~superlu-dist~unified-memory amdgpu_target=gfx90a arch=linux-ubuntu22.04-x86_64 [+] ^hip@5.2.0%gcc@11.2.0~ipo build_type=Release patches=959d1fe arch=linux-ubuntu22.04-x86_64 [+] ^cmake@3.23.3%gcc@11.2.0~doc+ncurses+ownlibs~qt build_type=Release arch=linux-ubuntu22.04-x86_64 [+] ^ncurses@6.2%gcc@11.2.0~symlinks+termlib abi=none arch=linux-ubuntu22.04-x86_64 [+] ^pkgconf@1.8.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^openssl@1.1.1q%gcc@11.2.0~docs~shared certs=mozilla patches=3fdcf2d arch=linux-ubuntu22.04-x86_64 [+] ^ca-certificates-mozilla@2022-07-19%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^perl@5.34.1%gcc@11.2.0+cpanm+shared+threads arch=linux-ubuntu22.04-x86_64 [+] ^berkeley-db@18.1.40%gcc@11.2.0+cxx~docs+stl patches=b231fcc arch=linux-ubuntu22.04-x86_64 [+] ^bzip2@1.0.8%gcc@11.2.0~debug~pic+shared arch=linux-ubuntu22.04-x86_64 [+] ^diffutils@3.8%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libiconv@1.16%gcc@11.2.0 libs=shared,static arch=linux-ubuntu22.04-x86_64 [+] ^gdbm@1.19%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^readline@8.1.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^zlib@1.2.12%gcc@11.2.0+optimize+pic+shared patches=0d38234 arch=linux-ubuntu22.04-x86_64 [+] ^comgr@5.2.0%gcc@11.2.0~ipo build_type=Release arch=linux-ubuntu22.04-x86_64 [+] ^llvm-amdgpu@5.2.0%gcc@11.2.0~ipo~link_llvm_dylib~llvm_dylib~openmp+rocm-device-libs build_type=Release patches=a08bbe1 arch=linux-ubuntu22.04-x86_64 [+] ^python@3.8.13%gcc@11.2.0+bz2+ctypes+dbm~debug+ensurepip+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3+ssl~tix~tkinter~ucs4+uuid+zlib patches=0d98e93,4c24573,f2fd060 arch=linux-ubuntu22.04-x86_64 [+] ^expat@2.4.8%gcc@11.2.0+libbsd arch=linux-ubuntu22.04-x86_64 [+] ^libbsd@0.11.5%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libmd@1.0.4%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^gettext@0.21%gcc@11.2.0+bzip2+curses+git~libunistring+libxml2+tar+xz arch=linux-ubuntu22.04-x86_64 [+] ^libxml2@2.9.13%gcc@11.2.0~python arch=linux-ubuntu22.04-x86_64 [+] ^xz@5.2.5%gcc@11.2.0+pic libs=shared,static arch=linux-ubuntu22.04-x86_64 [+] ^tar@1.34%gcc@11.2.0 zip=pigz arch=linux-ubuntu22.04-x86_64 [+] ^pigz@2.7%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^zstd@1.5.2%gcc@11.2.0+programs compression=none libs=shared,static arch=linux-ubuntu22.04-x86_64 [+] ^libffi@3.4.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^sqlite@3.38.5%gcc@11.2.0+column_metadata+dynamic_extensions+fts~functions+rtree arch=linux-ubuntu22.04-x86_64 [+] ^util-linux-uuid@2.37.4%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^z3@4.8.16%gcc@11.2.0~gmp~ipo~python build_type=RelWithDebInfo arch=linux-ubuntu22.04-x86_64 [+] ^rocm-cmake@5.2.0%gcc@11.2.0~ipo build_type=Release arch=linux-ubuntu22.04-x86_64 [+] ^glx@1.4%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^mesa@22.1.2%gcc@11.2.0+glx+llvm+opengl~opengles+osmesa~strip buildtype=release default_library=shared patches=ada85be arch=linux-ubuntu22.04-x86_64 [+] ^bison@3.8.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^m4@1.4.19%gcc@11.2.0+sigsegv patches=9dc5fbd,bfdffa7 arch=linux-ubuntu22.04-x86_64 [+] ^libsigsegv@2.13%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^flex@2.6.3%gcc@11.2.0+lex~nls arch=linux-ubuntu22.04-x86_64 [+] ^findutils@4.9.0%gcc@11.2.0 patches=440b954 arch=linux-ubuntu22.04-x86_64 [+] ^glproto@1.4.17%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^util-macros@1.19.3%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libunwind@1.6.2%gcc@11.2.0~block_signals~conservative_checks~cxx_exceptions~debug~debug_frame+docs+pic+tests+weak_backtrace+xz~zlib components=none libs=shared,static arch=linux-ubuntu22.04-x86_64 [+] ^libx11@1.7.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^inputproto@2.3.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^kbproto@1.0.7%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxcb@1.14%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libpthread-stubs@0.4%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxau@1.0.8%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xproto@7.0.31%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxdmcp@1.1.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xcb-proto@1.14.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xextproto@7.3.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xtrans@1.3.5%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxext@1.3.3%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxt@1.1.5%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libice@1.0.9%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libsm@1.2.3%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^meson@0.62.2%gcc@11.2.0 patches=9c87472,aa6c50d arch=linux-ubuntu22.04-x86_64 [+] ^ninja@1.11.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-pip@22.1.2%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-setuptools@63.0.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-wheel@0.37.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-mako@1.1.6%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^py-markupsafe@2.0.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xrandr@1.5.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxrandr@1.5.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libxrender@0.9.10%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^renderproto@0.11.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^randrproto@1.5.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^hsa-rocr-dev@5.2.0%gcc@11.2.0+image~ipo+shared build_type=Release patches=71e6851 arch=linux-ubuntu22.04-x86_64 [+] ^elfutils@0.186%gcc@11.2.0+bzip2~debuginfod~nls+xz arch=linux-ubuntu22.04-x86_64 [+] ^hsakmt-roct@5.2.0%gcc@11.2.0~ipo+shared build_type=Release patches=f926273 arch=linux-ubuntu22.04-x86_64 [+] ^libdrm@2.4.110%gcc@11.2.0~docs arch=linux-ubuntu22.04-x86_64 [+] ^libpciaccess@0.16%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^libtool@2.4.7%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^numactl@2.0.14%gcc@11.2.0 patches=4e1d78c,62fc8a8,ff37630 arch=linux-ubuntu22.04-x86_64 [+] ^autoconf@2.69%gcc@11.2.0 patches=35c4492,7793209,a49dd5b arch=linux-ubuntu22.04-x86_64 [+] ^automake@1.16.5%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^xxd-standalone@8.2.1201%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^perl-file-which@1.22%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^perl-uri-encode@1.1.1%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^perl-module-build@0.4224%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^rocminfo@5.2.0%gcc@11.2.0~ipo build_type=Release arch=linux-ubuntu22.04-x86_64 [+] ^roctracer-dev-api@5.2.0%gcc@11.2.0 arch=linux-ubuntu22.04-x86_64 [+] ^mpich@4.0.2%gcc@11.2.0~argobots~cuda+fortran+hwloc+hydra+libxml2+pci~rocm+romio~slurm~two_level_namespace~vci~verbs~wrapperrpath datatype-engine=auto device=ch4 netmod=ofi pmi=pmi arch=linux-ubuntu22.04-x86_64 [+] ^hwloc@2.8.0%gcc@11.2.0~cairo~cuda~gl~libudev+libxml2~netloc~nvml~oneapi-level-zero~opencl+pci~rocm+shared arch=linux-ubuntu22.04-x86_64 [+] ^libfabric@1.14.1%gcc@11.2.0~debug~kdreg fabrics=rxm,sockets,tcp,udp arch=linux-ubuntu22.04-x86_64 [+] ^yaksa@0.2%gcc@11.2.0~cuda~rocm arch=linux-ubuntu22.04-x86_64 [+] ^openblas@0.3.20%gcc@11.2.0~bignuma~consistent_fpcsr~ilp64+locking+pic+shared symbol_suffix=none threads=openmp arch=linux-ubuntu22.04-x86_64 </pre></details> Install error: ``` $> spack install -f ./hypre.spec.json ... ==> Installing hypre-2.25.0-6rsc2fdt7na5sqleojg3uug6klohx2f6 ==> No binary for hypre-2.25.0-6rsc2fdt7na5sqleojg3uug6klohx2f6 found: installing from source ==> Using cached archive: /spack/var/spack/cache/_source-cache/archive/f9/f9fc8371d91239fca694284dab17175bfda3821d7b7a871fd2e8f9d5930f303c.tar.gz ==> No patches needed for hypre ==> hypre: Executing phase: 'autoreconf' ==> hypre: Executing phase: 'configure' ==> hypre: Executing phase: 'build' ==> Error: ProcessError: Command exited with status 2: 'make' '-j16' 11 errors found in build log: 186 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c handle.c -o handle.ob j 187 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c int_array.c -o int_ar ray.obj 188 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c memory.c -o memory.ob j 189 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c omp_device.c -o omp_d evice.obj 190 hipcc --amdgpu-target=gfx90a -fPIC -O2 -x hip -std=c++14 -DHAVE_CONFIG_H -I.. -I./.. -I./../struct_mv -I. -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-ow pe5je3fhes7ivmri3nbrw6l44qe3ak/rocthrust/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocprim/include -I/spack/opt/s pack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fhes7ivmri3nbrw6l44qe3ak/rocsparse/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/hip-5.2.0-owpe5je3fh es7ivmri3nbrw6l44qe3ak/rocrand/include -I/spack/opt/spack/linux-ubuntu22.04-x86_64/gcc-11.2.0/mpich-4.0.2-gpzf75yhrekurnz3v63fmevfjez2x2q2/include -c nvtx.c -o nvtx.obj 191 In file included from device_utils.c:10: >> 192 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 193 #include <rocsparse.h> 194 ^~~~~~~~~~~~~ 195 In file included from general.c:9: >> 196 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 197 #include <rocsparse.h> 198 ^~~~~~~~~~~~~ 199 In file included from handle.c:15: >> 200 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 201 #include <rocsparse.h> 202 ^~~~~~~~~~~~~ 203 1 error generated when compiling for gfx90a. 204 1 error generated when compiling for gfx90a. 205 In file included from int_array.c:9: >> 206 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 207 #include <rocsparse.h> 208 ^~~~~~~~~~~~~ 209 1 error generated when compiling for gfx90a. >> 210 make[1]: *** [../config/Makefile.config:66: general.obj] Error 1 211 make[1]: *** Waiting for unfinished jobs.... >> 212 make[1]: *** [../config/Makefile.config:66: device_utils.obj] Error 1 213 1 error generated when compiling for gfx90a. >> 214 make[1]: *** [../config/Makefile.config:66: handle.obj] Error 1 >> 215 make[1]: *** [../config/Makefile.config:66: int_array.obj] Error 1 216 In file included from memory.c:15: >> 217 ./_hypre_utilities.hpp:116:10: fatal error: 'rocsparse.h' file not found 218 #include <rocsparse.h> 219 ^~~~~~~~~~~~~ 220 1 error generated when compiling for gfx90a. >> 221 make[1]: *** [../config/Makefile.config:66: memory.obj] Error 1 222 make[1]: Leaving directory '/tmp/root/spack-stage/spack-stage-hypre-2.25.0-6rsc2fdt7na5sqleojg3uug6klohx2f6/spack-src/src/utilities' >> 223 make: *** [Makefile:86: all] Error 1 ``` ### Error message <details><summary>Error message</summary><pre> ... see above </pre></details> ### Information on your system * **Spack:** 0.19.0.dev0 (9d5375a3ba150a3b216f06bc435b335b8f9bfc42) * **Python:** 3.10.4 * **Platform:** linux-ubuntu22.04-zen2 * **Concretizer:** clingo ### Additional information [spack-build-out.txt](https://github.com/spack/spack/files/9354620/spack-build-out.txt) [spack-build-env.txt](https://github.com/spack/spack/files/9354621/spack-build-env.txt) @balay @osborn9 @ulrikeyang @wspear ### General information - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded the build log and environment files - [X] I have searched the issues of this repo and believe this is not a duplicate
non_test
hypre rocm fails hypre utilities hpp fatal error rocsparse h file not found steps to reproduce the issue reproducible using spack develop from mon aug runner gcc rocm concrete spec concretization hypre gcc complex cuda debug fortran gptune internal superlu mixedint mpi openmp rocm shared superlu dist unified memory amdgpu target arch linux hip gcc ipo build type release patches arch linux cmake gcc doc ncurses ownlibs qt build type release arch linux ncurses gcc symlinks termlib abi none arch linux pkgconf gcc arch linux openssl gcc docs shared certs mozilla patches arch linux ca certificates mozilla gcc arch linux perl gcc cpanm shared threads arch linux berkeley db gcc cxx docs stl patches arch linux gcc debug pic shared arch linux diffutils gcc arch linux libiconv gcc libs shared static arch linux gdbm gcc arch linux readline gcc arch linux zlib gcc optimize pic shared patches arch linux comgr gcc ipo build type release arch linux llvm amdgpu gcc ipo link llvm dylib llvm dylib openmp rocm device libs build type release patches arch linux python gcc ctypes dbm debug ensurepip lzma nis optimizations pic pyexpat pythoncmd readline shared ssl tix tkinter uuid zlib patches arch linux expat gcc libbsd arch linux libbsd gcc arch linux libmd gcc arch linux gettext gcc curses git libunistring tar xz arch linux gcc python arch linux xz gcc pic libs shared static arch linux tar gcc zip pigz arch linux pigz gcc arch linux zstd gcc programs compression none libs shared static arch linux libffi gcc arch linux sqlite gcc column metadata dynamic extensions fts functions rtree arch linux util linux uuid gcc arch linux gcc gmp ipo python build type relwithdebinfo arch linux rocm cmake gcc ipo build type release arch linux glx gcc arch linux mesa gcc glx llvm opengl opengles osmesa strip buildtype release default library shared patches arch linux bison gcc arch linux gcc sigsegv patches arch linux libsigsegv gcc arch linux flex gcc lex nls arch linux findutils gcc patches arch linux glproto gcc arch linux util macros gcc arch linux libunwind gcc block signals conservative checks cxx exceptions debug debug frame docs pic tests weak backtrace xz zlib components none libs shared static arch linux gcc arch linux inputproto gcc arch linux kbproto gcc arch linux libxcb gcc arch linux libpthread stubs gcc arch linux libxau gcc arch linux xproto gcc arch linux libxdmcp gcc arch linux xcb proto gcc arch linux xextproto gcc arch linux xtrans gcc arch linux libxext gcc arch linux libxt gcc arch linux libice gcc arch linux libsm gcc arch linux meson gcc patches arch linux ninja gcc arch linux py pip gcc arch linux py setuptools gcc arch linux py wheel gcc arch linux py mako gcc arch linux py markupsafe gcc arch linux xrandr gcc arch linux libxrandr gcc arch linux libxrender gcc arch linux renderproto gcc arch linux randrproto gcc arch linux hsa rocr dev gcc image ipo shared build type release patches arch linux elfutils gcc debuginfod nls xz arch linux hsakmt roct gcc ipo shared build type release patches arch linux libdrm gcc docs arch linux libpciaccess gcc arch linux libtool gcc arch linux numactl gcc patches arch linux autoconf gcc patches arch linux automake gcc arch linux xxd standalone gcc arch linux perl file which gcc arch linux perl uri encode gcc arch linux perl module build gcc arch linux rocminfo gcc ipo build type release arch linux roctracer dev api gcc arch linux mpich gcc argobots cuda fortran hwloc hydra pci rocm romio slurm two level namespace vci verbs wrapperrpath datatype engine auto device netmod ofi pmi pmi arch linux hwloc gcc cairo cuda gl libudev netloc nvml oneapi level zero opencl pci rocm shared arch linux libfabric gcc debug kdreg fabrics rxm sockets tcp udp arch linux yaksa gcc cuda rocm arch linux openblas gcc bignuma consistent fpcsr locking pic shared symbol suffix none threads openmp arch linux install error spack install f hypre spec json installing hypre no binary for hypre found installing from source using cached archive spack var spack cache source cache archive tar gz no patches needed for hypre hypre executing phase autoreconf hypre executing phase configure hypre executing phase build error processerror command exited with status make errors found in build log hipcc amdgpu target fpic x hip std c dhave config h i i i struct mv i i spack opt spack linux gcc hip ow rocthrust include i spack opt spack linux gcc hip rocprim include i spack opt s pack linux gcc hip rocsparse include i spack opt spack linux gcc hip rocrand include i spack opt spack linux gcc mpich include c handle c o handle ob j hipcc amdgpu target fpic x hip std c dhave config h i i i struct mv i i spack opt spack linux gcc hip ow rocthrust include i spack opt spack linux gcc hip rocprim include i spack opt s pack linux gcc hip rocsparse include i spack opt spack linux gcc hip rocrand include i spack opt spack linux gcc mpich include c int array c o int ar ray obj hipcc amdgpu target fpic x hip std c dhave config h i i i struct mv i i spack opt spack linux gcc hip ow rocthrust include i spack opt spack linux gcc hip rocprim include i spack opt s pack linux gcc hip rocsparse include i spack opt spack linux gcc hip rocrand include i spack opt spack linux gcc mpich include c memory c o memory ob j hipcc amdgpu target fpic x hip std c dhave config h i i i struct mv i i spack opt spack linux gcc hip ow rocthrust include i spack opt spack linux gcc hip rocprim include i spack opt s pack linux gcc hip rocsparse include i spack opt spack linux gcc hip rocrand include i spack opt spack linux gcc mpich include c omp device c o omp d evice obj hipcc amdgpu target fpic x hip std c dhave config h i i i struct mv i i spack opt spack linux gcc hip ow rocthrust include i spack opt spack linux gcc hip rocprim include i spack opt s pack linux gcc hip rocsparse include i spack opt spack linux gcc hip rocrand include i spack opt spack linux gcc mpich include c nvtx c o nvtx obj in file included from device utils c hypre utilities hpp fatal error rocsparse h file not found include in file included from general c hypre utilities hpp fatal error rocsparse h file not found include in file included from handle c hypre utilities hpp fatal error rocsparse h file not found include error generated when compiling for error generated when compiling for in file included from int array c hypre utilities hpp fatal error rocsparse h file not found include error generated when compiling for make error make waiting for unfinished jobs make error error generated when compiling for make error make error in file included from memory c hypre utilities hpp fatal error rocsparse h file not found include error generated when compiling for make error make leaving directory tmp root spack stage spack stage hypre spack src src utilities make error error message error message see above information on your system spack python platform linux concretizer clingo additional information balay ulrikeyang wspear general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
0
522,377
15,158,805,380
IssuesEvent
2021-02-12 02:12:59
sonia-auv/sonia_common
https://api.github.com/repos/sonia-auv/sonia_common
closed
Bug with tag for docker image push
Priority: High Type: Bug
**Warning :** Before creating an issue or task, make sure that it does not already exists in the [issue tracker](../). Thank you. ## Expected Behavior Push the required images with the right tag ## Current Behavior Crashed with the docker push command
1.0
Bug with tag for docker image push - **Warning :** Before creating an issue or task, make sure that it does not already exists in the [issue tracker](../). Thank you. ## Expected Behavior Push the required images with the right tag ## Current Behavior Crashed with the docker push command
non_test
bug with tag for docker image push warning before creating an issue or task make sure that it does not already exists in the thank you expected behavior push the required images with the right tag current behavior crashed with the docker push command
0
299,441
25,902,455,852
IssuesEvent
2022-12-15 07:22:26
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: copy/bank/rows=100000,nodes=5,txn=true failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.copy/bank/rows=100000,nodes=5,txn=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7969261?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7969261?buildTab=artifacts#/copy/bank/rows=100000,nodes=5,txn=true) on master @ [d6f98e90684894fd36f53596e6aac355676d232e](https://github.com/cockroachdb/cockroach/commits/d6f98e90684894fd36f53596e6aac355676d232e): ``` test artifacts and logs in: /artifacts/copy/bank/rows=100000_nodes=5_txn=true/run_1 (test_impl.go:291).Fatal: monitor failure: monitor task failed: empty bank table split over multiple ranges ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*copy/bank/rows=100000,nodes=5,txn=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: copy/bank/rows=100000,nodes=5,txn=true failed - roachtest.copy/bank/rows=100000,nodes=5,txn=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7969261?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7969261?buildTab=artifacts#/copy/bank/rows=100000,nodes=5,txn=true) on master @ [d6f98e90684894fd36f53596e6aac355676d232e](https://github.com/cockroachdb/cockroach/commits/d6f98e90684894fd36f53596e6aac355676d232e): ``` test artifacts and logs in: /artifacts/copy/bank/rows=100000_nodes=5_txn=true/run_1 (test_impl.go:291).Fatal: monitor failure: monitor task failed: empty bank table split over multiple ranges ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*copy/bank/rows=100000,nodes=5,txn=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest copy bank rows nodes txn true failed roachtest copy bank rows nodes txn true with on master test artifacts and logs in artifacts copy bank rows nodes txn true run test impl go fatal monitor failure monitor task failed empty bank table split over multiple ranges parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage
1
35,802
14,884,695,966
IssuesEvent
2021-01-20 14:53:15
opstrace/opstrace
https://api.github.com/repos/opstrace/opstrace
opened
ci: dns service quota reached
area: ci area: dns-service priority: 0
``` [2021-01-20T14:45:11Z] 2021-01-20T14:45:11.432Z debug: cluster name not found in opstraceClient.GetAll() response, call opstraceClient.Create() [2021-01-20T14:45:11Z] 2021-01-20T14:45:11.432Z debug: do DNS service client:POST [2021-01-20T14:45:11Z] 2021-01-20T14:45:11.503Z warning: DNS service client:POST failed: Response code 429 (Too Many Requests) [2021-01-20T14:45:11Z] 2021-01-20T14:45:11.503Z debug: HTTP resp to POST(https://dns-api.opstrace.net/dns/): [2021-01-20T14:45:11Z] status: 429 [2021-01-20T14:45:11Z] body[:500]: {"error":"Maximum number of names reached."} [2021-01-20T14:45:11Z] headers: {"x-powered-by":"Express","content-type":"application/json; charset=utf-8","content-length":"44","etag":"W/\"2c-ks+6Eln1r58Hkj7sNd9LudRGmZM\"","date":"Wed, 20 Jan 2021 14:45:11 GMT","via":"1.1 google","alt-svc":"clear","connection":"close"} [2021-01-20T14:45:11Z] totalTime: 0.07 s [2021-01-20T14:45:11Z] dnsDone->TCPconnectDone: 0.006 s [2021-01-20T14:45:11Z] connectDone->reqSent 0 s [2021-01-20T14:45:11Z] reqSent->firstResponseByte: 0.05 s ```
1.0
ci: dns service quota reached - ``` [2021-01-20T14:45:11Z] 2021-01-20T14:45:11.432Z debug: cluster name not found in opstraceClient.GetAll() response, call opstraceClient.Create() [2021-01-20T14:45:11Z] 2021-01-20T14:45:11.432Z debug: do DNS service client:POST [2021-01-20T14:45:11Z] 2021-01-20T14:45:11.503Z warning: DNS service client:POST failed: Response code 429 (Too Many Requests) [2021-01-20T14:45:11Z] 2021-01-20T14:45:11.503Z debug: HTTP resp to POST(https://dns-api.opstrace.net/dns/): [2021-01-20T14:45:11Z] status: 429 [2021-01-20T14:45:11Z] body[:500]: {"error":"Maximum number of names reached."} [2021-01-20T14:45:11Z] headers: {"x-powered-by":"Express","content-type":"application/json; charset=utf-8","content-length":"44","etag":"W/\"2c-ks+6Eln1r58Hkj7sNd9LudRGmZM\"","date":"Wed, 20 Jan 2021 14:45:11 GMT","via":"1.1 google","alt-svc":"clear","connection":"close"} [2021-01-20T14:45:11Z] totalTime: 0.07 s [2021-01-20T14:45:11Z] dnsDone->TCPconnectDone: 0.006 s [2021-01-20T14:45:11Z] connectDone->reqSent 0 s [2021-01-20T14:45:11Z] reqSent->firstResponseByte: 0.05 s ```
non_test
ci dns service quota reached debug cluster name not found in opstraceclient getall response call opstraceclient create debug do dns service client post warning dns service client post failed response code too many requests debug http resp to post status body error maximum number of names reached headers x powered by express content type application json charset utf content length etag w ks date wed jan gmt via google alt svc clear connection close totaltime s dnsdone tcpconnectdone s connectdone reqsent s reqsent firstresponsebyte s
0
394,549
11,645,053,734
IssuesEvent
2020-02-29 22:27:45
TeamCautionRobotics/2020-FRC-robot
https://api.github.com/repos/TeamCautionRobotics/2020-FRC-robot
closed
Implement Climb code in Robot.java
Timed Robot high priority
Ports and controls are temporary - Motor ports: winch port 7, arm port 8 - Piston port 3 - Limit switch port 9 - Use whatever controls you want. We'll change them later.
1.0
Implement Climb code in Robot.java - Ports and controls are temporary - Motor ports: winch port 7, arm port 8 - Piston port 3 - Limit switch port 9 - Use whatever controls you want. We'll change them later.
non_test
implement climb code in robot java ports and controls are temporary motor ports winch port arm port piston port limit switch port use whatever controls you want we ll change them later
0
36,162
9,762,193,599
IssuesEvent
2019-06-05 10:40:14
ShaikASK/Testing
https://api.github.com/repos/ShaikASK/Testing
closed
Edit Folders Screen : Document : Non dynamic document is getting displayed in the edit folders with web icon symbol
Defect Documents Dynamic Webform Folders P2 Release #4 Build # 1
Steps To Replicate : 1.Launch the URL 2.Sign in as HR admin user 3.Upload a ”Document” with “Dynamic Webform” settings 4.Navigate to Edit Dynamic Webform screen 5.Provide mapping from “Edit Dynamic webform’ screen and save it 6.Check the above uploaded document in the folders (displayed with web icon symbol) 7.Edit the above uploaded document by removing ‘Dynamic Webform” settings 8.Go to Folders an check the document in the folders screen (document will be displayed expected) 9.Add the document in the folder and save it 10..Edit the above created folder Experienced Behavior : Observed that Non dynamic document is getting displayed in the edit folders with web icon symbol Expected Behavior : Ensure that Non dynamic document should not display with web icon symbol in the edit folders screen
1.0
Edit Folders Screen : Document : Non dynamic document is getting displayed in the edit folders with web icon symbol - Steps To Replicate : 1.Launch the URL 2.Sign in as HR admin user 3.Upload a ”Document” with “Dynamic Webform” settings 4.Navigate to Edit Dynamic Webform screen 5.Provide mapping from “Edit Dynamic webform’ screen and save it 6.Check the above uploaded document in the folders (displayed with web icon symbol) 7.Edit the above uploaded document by removing ‘Dynamic Webform” settings 8.Go to Folders an check the document in the folders screen (document will be displayed expected) 9.Add the document in the folder and save it 10..Edit the above created folder Experienced Behavior : Observed that Non dynamic document is getting displayed in the edit folders with web icon symbol Expected Behavior : Ensure that Non dynamic document should not display with web icon symbol in the edit folders screen
non_test
edit folders screen document non dynamic document is getting displayed in the edit folders with web icon symbol steps to replicate launch the url sign in as hr admin user upload a ”document” with “dynamic webform” settings navigate to edit dynamic webform screen provide mapping from “edit dynamic webform’ screen and save it check the above uploaded document in the folders displayed with web icon symbol edit the above uploaded document by removing ‘dynamic webform” settings go to folders an check the document in the folders screen document will be displayed expected add the document in the folder and save it edit the above created folder experienced behavior observed that non dynamic document is getting displayed in the edit folders with web icon symbol expected behavior ensure that non dynamic document should not display with web icon symbol in the edit folders screen
0
87,831
10,556,367,646
IssuesEvent
2019-10-04 01:29:44
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
[web]: Adding new dependency causes error on reload
d: examples documentation severe: crash tool ☸ platform-web
In VS Code 1.38 with the Flutter plugin installed. After a clean install, following the instructions at [Write your first app - Step 2: Use an external package](https://flutter.dev/docs/get-started/codelab#step-2-use-an-external-package). Adding the `english_words: ^3.1.0` dependency in `pubspec.yaml` triggers `flutter pub get` as expected. After adding the import and changing `lib/main.dart`, I'm getting the following error: ``` Unable to find modules for some sources, this is usually the result of either a bad import, a missing dependency in a package (or possibly a dev_dependency needs to move to a real dependency), or a build failure (if importing a generated file). Please check the following imports: `import 'package:english_words/english_words.dart';` from flutter_hello|lib/main.dart at 6:1 Failed after 76ms Performing hot restart... 584ms Try again after fixing the above error(s). ``` Stopping the process and restarting `flutter run -d chrome` fixes the issue with the following message: ``` Throwing away cached asset graph because the build phases have changed. This most commonly would happen as a result of adding a new dependency or updating your dependencies. ```
1.0
[web]: Adding new dependency causes error on reload - In VS Code 1.38 with the Flutter plugin installed. After a clean install, following the instructions at [Write your first app - Step 2: Use an external package](https://flutter.dev/docs/get-started/codelab#step-2-use-an-external-package). Adding the `english_words: ^3.1.0` dependency in `pubspec.yaml` triggers `flutter pub get` as expected. After adding the import and changing `lib/main.dart`, I'm getting the following error: ``` Unable to find modules for some sources, this is usually the result of either a bad import, a missing dependency in a package (or possibly a dev_dependency needs to move to a real dependency), or a build failure (if importing a generated file). Please check the following imports: `import 'package:english_words/english_words.dart';` from flutter_hello|lib/main.dart at 6:1 Failed after 76ms Performing hot restart... 584ms Try again after fixing the above error(s). ``` Stopping the process and restarting `flutter run -d chrome` fixes the issue with the following message: ``` Throwing away cached asset graph because the build phases have changed. This most commonly would happen as a result of adding a new dependency or updating your dependencies. ```
non_test
adding new dependency causes error on reload in vs code with the flutter plugin installed after a clean install following the instructions at adding the english words dependency in pubspec yaml triggers flutter pub get as expected after adding the import and changing lib main dart i m getting the following error unable to find modules for some sources this is usually the result of either a bad import a missing dependency in a package or possibly a dev dependency needs to move to a real dependency or a build failure if importing a generated file please check the following imports import package english words english words dart from flutter hello lib main dart at failed after performing hot restart try again after fixing the above error s stopping the process and restarting flutter run d chrome fixes the issue with the following message throwing away cached asset graph because the build phases have changed this most commonly would happen as a result of adding a new dependency or updating your dependencies
0
225,938
17,931,404,056
IssuesEvent
2021-09-10 09:43:24
apache/shardingsphere
https://api.github.com/repos/apache/shardingsphere
closed
Add unit test for shardingsphere-sql-parser-engine
in: test good first issue project: OSD2021
Hi community, This issue is for #10887. ## Aim Add unit test for shardingsphere-sql-parser-engine project to test its public functions. ## Basic Qualifications - Java - Maven - Junit.Test ## Classes Add unit test for classes to test its public functions. - [ ] SQLParserEngine - [ ] SQLVisitorEngine - [ ] SQLParserFactory ## Examples FYI - SQLParserEngineTest
1.0
Add unit test for shardingsphere-sql-parser-engine - Hi community, This issue is for #10887. ## Aim Add unit test for shardingsphere-sql-parser-engine project to test its public functions. ## Basic Qualifications - Java - Maven - Junit.Test ## Classes Add unit test for classes to test its public functions. - [ ] SQLParserEngine - [ ] SQLVisitorEngine - [ ] SQLParserFactory ## Examples FYI - SQLParserEngineTest
test
add unit test for shardingsphere sql parser engine hi community this issue is for aim add unit test for shardingsphere sql parser engine project to test its public functions basic qualifications java maven junit test classes add unit test for classes to test its public functions sqlparserengine sqlvisitorengine sqlparserfactory examples fyi sqlparserenginetest
1
202,393
15,828,993,324
IssuesEvent
2021-04-06 10:35:03
enthought/traitsui
https://api.github.com/repos/enthought/traitsui
closed
Use "viewcode" Sphinx extension in documentation.
component: documentation difficulty: easy type: enhancement
I've found it really helpful to have direct links from the documentation to the source in many cases. We're already doing this for some ETS packages, but not others. Could we add the source links for TraitsUI and Pyface? This should be as easy as adding `"sphinx.ext.viewcode"` to the list of Sphinx extensions in the `conf.py` configuration file.
1.0
Use "viewcode" Sphinx extension in documentation. - I've found it really helpful to have direct links from the documentation to the source in many cases. We're already doing this for some ETS packages, but not others. Could we add the source links for TraitsUI and Pyface? This should be as easy as adding `"sphinx.ext.viewcode"` to the list of Sphinx extensions in the `conf.py` configuration file.
non_test
use viewcode sphinx extension in documentation i ve found it really helpful to have direct links from the documentation to the source in many cases we re already doing this for some ets packages but not others could we add the source links for traitsui and pyface this should be as easy as adding sphinx ext viewcode to the list of sphinx extensions in the conf py configuration file
0
179,341
14,702,430,567
IssuesEvent
2021-01-04 13:35:27
connectomicslab/connectomemapper3
https://api.github.com/repos/connectomicslab/connectomemapper3
closed
[TASK] Write missing docstrings and documentation
documentation good first issue help wanted
*Tasks* - [x] Write docstrings where missing to describe modules / classes / functions. Add citations to softwares and/or papers whenever it applies - [x] Insert API documentation in the docs *Resources* * Sphinx Read-The-Docs tutorial on docstrings: https://sphinx-rtd-tutorial.readthedocs.io/en/latest/docstrings.html * An other link that shows how to use bibtex for citation/biblio management in docstrings: https://lightlab.readthedocs.io/en/master/_static/developers/docYourCode.html * Demo for API documentation in Read-The-Docs: https://sphinx-rtd-theme.readthedocs.io/en/stable/demo/api.html
1.0
[TASK] Write missing docstrings and documentation - *Tasks* - [x] Write docstrings where missing to describe modules / classes / functions. Add citations to softwares and/or papers whenever it applies - [x] Insert API documentation in the docs *Resources* * Sphinx Read-The-Docs tutorial on docstrings: https://sphinx-rtd-tutorial.readthedocs.io/en/latest/docstrings.html * An other link that shows how to use bibtex for citation/biblio management in docstrings: https://lightlab.readthedocs.io/en/master/_static/developers/docYourCode.html * Demo for API documentation in Read-The-Docs: https://sphinx-rtd-theme.readthedocs.io/en/stable/demo/api.html
non_test
write missing docstrings and documentation tasks write docstrings where missing to describe modules classes functions add citations to softwares and or papers whenever it applies insert api documentation in the docs resources sphinx read the docs tutorial on docstrings an other link that shows how to use bibtex for citation biblio management in docstrings demo for api documentation in read the docs
0
275,752
23,936,646,953
IssuesEvent
2022-09-11 10:26:47
dnd-side-project/dnd-7th-2-android
https://api.github.com/repos/dnd-side-project/dnd-7th-2-android
closed
이메일 인증 로직(이메일 전송 및 이메일 인증번호 확인)
feature test
### 구현해야할 내용 - [x] 이메일 인증 로직 처리 - [x] 이메일 전송 로직 - [x] 이메일 인증번호 확인 로직 - [x] 각 로직 테스트코드 작성
1.0
이메일 인증 로직(이메일 전송 및 이메일 인증번호 확인) - ### 구현해야할 내용 - [x] 이메일 인증 로직 처리 - [x] 이메일 전송 로직 - [x] 이메일 인증번호 확인 로직 - [x] 각 로직 테스트코드 작성
test
이메일 인증 로직 이메일 전송 및 이메일 인증번호 확인 구현해야할 내용 이메일 인증 로직 처리 이메일 전송 로직 이메일 인증번호 확인 로직 각 로직 테스트코드 작성
1
239,041
19,806,626,126
IssuesEvent
2022-01-19 07:40:49
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: sqlsmith/setup=rand-tables/setting=no-mutations failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.sqlsmith/setup=rand-tables/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4149622&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4149622&tab=artifacts#/sqlsmith/setup=rand-tables/setting=no-mutations) on master @ [912964e02ddd951c77d4f71981ae18b3894e9084](https://github.com/cockroachdb/cockroach/commits/912964e02ddd951c77d4f71981ae18b3894e9084): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/sqlsmith/setup=rand-tables/setting=no-mutations/run_1 sqlsmith.go:259,sqlsmith.go:298,test_runner.go:780: error: pq: internal error: no volatility for cast regnamespace::regrole stmt: SELECT 'R':::STRING AS col_11597, CASE WHEN ('192.54.138.135/30':::INET::INET >> '45e2:63e:5c62:9c2c:a729:1cab:caf:f293/114':::INET::INET)::BOOL THEN tab_4943.col1_1 ELSE tab_4943.col1_4 END AS col_11598, tab_4943.tableoid AS col_11599, 2931691121:::OID AS col_11600, tab_4943.col1_1 AS col_11601, tab_4943.col1_5 AS col_11602, '':::STRING AS col_11603 FROM defaultdb.public.table1@table1_pkey AS tab_4943 WHERE (false OR true) LIMIT 3:::INT8; ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=rand-tables/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: sqlsmith/setup=rand-tables/setting=no-mutations failed - roachtest.sqlsmith/setup=rand-tables/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4149622&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4149622&tab=artifacts#/sqlsmith/setup=rand-tables/setting=no-mutations) on master @ [912964e02ddd951c77d4f71981ae18b3894e9084](https://github.com/cockroachdb/cockroach/commits/912964e02ddd951c77d4f71981ae18b3894e9084): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/sqlsmith/setup=rand-tables/setting=no-mutations/run_1 sqlsmith.go:259,sqlsmith.go:298,test_runner.go:780: error: pq: internal error: no volatility for cast regnamespace::regrole stmt: SELECT 'R':::STRING AS col_11597, CASE WHEN ('192.54.138.135/30':::INET::INET >> '45e2:63e:5c62:9c2c:a729:1cab:caf:f293/114':::INET::INET)::BOOL THEN tab_4943.col1_1 ELSE tab_4943.col1_4 END AS col_11598, tab_4943.tableoid AS col_11599, 2931691121:::OID AS col_11600, tab_4943.col1_1 AS col_11601, tab_4943.col1_5 AS col_11602, '':::STRING AS col_11603 FROM defaultdb.public.table1@table1_pkey AS tab_4943 WHERE (false OR true) LIMIT 3:::INT8; ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=rand-tables/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest sqlsmith setup rand tables setting no mutations failed roachtest sqlsmith setup rand tables setting no mutations with on master the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts sqlsmith setup rand tables setting no mutations run sqlsmith go sqlsmith go test runner go error pq internal error no volatility for cast regnamespace regrole stmt select r string as col case when inet inet caf inet inet bool then tab else tab end as col tab tableoid as col oid as col tab as col tab as col string as col from defaultdb public pkey as tab where false or true limit help see see cc cockroachdb sql queries
1
204,720
15,528,792,153
IssuesEvent
2021-03-13 12:33:23
B3Partners/brmo
https://api.github.com/repos/B3Partners/brmo
closed
Sommige integratie tests van de brmo-loader falen op Windows
bouw- en testinfrastuctuur bug :bomb: windows
### Probleem Sommige integratie tests van de brmo-loader falen op Windows als zo'n test een bericht moet transformeren en de tests met java 11 en Maven draaien het gaat om: Failures: - `BRKLocatiebeschrijvingIntegrationTest#testLocatieBeschrijving` - `BrkToStagingToRsgbIntegrationTest#testBrkXMLToStaging` - `GH789GroteWaardeIDIntegrationTest#testTransformBerichten` - `Mantis10315IntegrationTest#testTransformBerichten` - `Mantis10547IntegrationTest#testAll` - `Mantis11180IntegrationTest#testTransformBerichten` - `Mantis6098IntegrationTest#testAll` - `Mantis6098IntegrationTest#testStand` - `Mantis6098IntegrationTest#testStandDelete` - `Mantis6098IntegrationTest#testStandDeleteMutatie` - `Mantis6098IntegrationTest#testStandMutatie` - `Mantis6166IntegrationTest#testAll` - `Mantis6166IntegrationTest#testStand` - `Mantis6166IntegrationTest#testStandDelete` - `Mantis6166IntegrationTest#testStandDeleteMutatie` - `Mantis6166IntegrationTest#testStandMutatie` - `Mantis6380IntegrationTest#testTransformBerichten` - `Mantis6380IntegrationTest#testTransformStand` - `Mantis6380IntegrationTest#testTransformStandDaarnaMutatieBerichten` - `VerminderenStukdelenIntegrationTest#testMinderStukdelenInMutatie` - `ZakRechtArchiefIntegrationTest#transformeerberichten` - `BagBerichtIntegrationTest#loadAll` ontdekt tijdens maken van #1025 (dus vooralsnog ronden we die af met java 8...) de kern van het probleem blijkt uit de melding: ``` BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#updateBerichtException:422) - Mutatieverwerking en foutief bericht, dus stoppen! java.nio.file.InvalidPathException: Illegal char <:> at index 21: org\geotools\gml3\jar:file:\C:\Users\mark\.m2\repository\org\geotools\xsd\gt-xsd-gml3\24.2\gt-xsd-gml3-24.2.jar!\org\geotools\gml3\gml.xsd at java.base/sun.nio.fs.WindowsPathParser.normalize(WindowsPathParser.java:182) ``` Er gaat dus iets mis met de constructie van het pad naar de xsd. De tests draaien wel goed vanuit de IDE / Intellij. ### stacktrace <details><summary>maven log:</summary> ``` [INFO] Running nl.b3p.brmo.loader.entity.BagBerichtIntegrationTest BRMO-LOADER-TEST: INFO 17:40:30 (n.b.AbstractDatabaseIntegrationTest#startTest:103) - ==== Start test methode: loadInStaging() BRMO-LOADER-TEST: INFO 17:40:31 (n.b.AbstractDatabaseIntegrationTest#endTest:111) - ==== Einde test methode: loadInStaging() BRMO-LOADER-TEST: INFO 17:40:31 (n.b.AbstractDatabaseIntegrationTest#startTest:103) - ==== Start test methode: loadAll() BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#updateBerichtException:422) - Mutatieverwerking en foutief bericht, dus stoppen! java.nio.file.InvalidPathException: Illegal char <:> at index 21: org\geotools\gml3\jar:file:\C:\Users\mark\.m2\repository\org\geotools\xsd\gt-xsd-gml3\24.2\gt-xsd-gml3-24.2.jar!\org\geotools\gml3\gml.xsd at java.base/sun.nio.fs.WindowsPathParser.normalize(WindowsPathParser.java:182) at java.base/sun.nio.fs.WindowsPathParser.parse(WindowsPathParser.java:153) at java.base/sun.nio.fs.WindowsPathParser.parse(WindowsPathParser.java:77) at java.base/sun.nio.fs.WindowsPath.parse(WindowsPath.java:92) at java.base/sun.nio.fs.WindowsFileSystem.getPath(WindowsFileSystem.java:229) at java.base/jdk.internal.module.Resources.toSafeFilePath(Resources.java:143) at java.base/jdk.internal.module.Resources.toFilePath(Resources.java:97) at java.base/jdk.internal.module.ModulePatcher$ExplodedResourceFinder.find(ModulePatcher.java:505) at java.base/jdk.internal.module.ModulePatcher$PatchedModuleReader.findResourceInPatch(ModulePatcher.java:301) at java.base/jdk.internal.module.ModulePatcher$PatchedModuleReader.find(ModulePatcher.java:357) at java.base/jdk.internal.loader.BuiltinClassLoader$2.run(BuiltinClassLoader.java:407) at java.base/jdk.internal.loader.BuiltinClassLoader$2.run(BuiltinClassLoader.java:402) at java.base/java.security.AccessController.doPrivileged(Native Method) at java.base/jdk.internal.loader.BuiltinClassLoader.findMiscResource(BuiltinClassLoader.java:401) at java.base/jdk.internal.loader.BuiltinClassLoader.findResource(BuiltinClassLoader.java:288) at java.base/java.lang.ClassLoader.getResource(ClassLoader.java:1401) at java.base/java.lang.Class.getResource(Class.java:2740) at org.geotools.xsd.SchemaLocationResolver.resolveLocationToResource(SchemaLocationResolver.java:111) at org.geotools.xsd.SchemaLocationResolver.canHandle(SchemaLocationResolver.java:102) at org.geotools.xsd.SchemaLocationResolver.resolveSchemaLocation(SchemaLocationResolver.java:151) at org.geotools.xsd.impl.ParserHandler.startElement(ParserHandler.java:399) at java.xml/com.sun.org.apache.xerces.internal.parsers.AbstractSAXParser.startElement(AbstractSAXParser.java:510) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.scanStartElement(XMLNSDocumentScannerImpl.java:374) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl$NSContentDriver.scanRootElementHook(XMLNSDocumentScannerImpl.java:613) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl$FragmentContentDriver.next(XMLDocumentFragmentScannerImpl.java:3063) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl$PrologDriver.next(XMLDocumentScannerImpl.java:836) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl.next(XMLDocumentScannerImpl.java:605) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.next(XMLNSDocumentScannerImpl.java:112) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:534) at java.xml/com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:888) at java.xml/com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:824) at java.xml/com.sun.org.apache.xerces.internal.parsers.XMLParser.parse(XMLParser.java:141) at java.xml/com.sun.org.apache.xerces.internal.parsers.AbstractSAXParser.parse(AbstractSAXParser.java:1216) at java.xml/com.sun.org.apache.xerces.internal.jaxp.SAXParserImpl$JAXPSAXParser.parse(SAXParserImpl.java:635) at java.xml/com.sun.org.apache.xerces.internal.jaxp.SAXParserImpl.parse(SAXParserImpl.java:324) at org.geotools.xsd.Parser.parse(Parser.java:184) at org.geotools.xsd.Parser.parse(Parser.java:139) at brmo.loader/nl.b3p.brmo.loader.util.DataComfortXMLReader.readDataXML(DataComfortXMLReader.java:185) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.transformToTableData(RsgbProxy.java:362) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.handle(RsgbProxy.java:485) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:410) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:377) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:388) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:266) at brmo.loader/nl.b3p.brmo.loader.StagingProxy.handleBerichtenByJob(StagingProxy.java:377) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.run(RsgbProxy.java:289) at java.base/java.lang.Thread.run(Thread.java:834) BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#handle:560) - Fout bij verwerking bericht met id 41402969931, melding: null BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#updateBerichtException:422) - Mutatieverwerking en foutief bericht, dus stoppen! java.lang.NullPointerException at brmo.loader/nl.b3p.brmo.loader.StagingProxy.updateBerichtProcessing(StagingProxy.java:689) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.updateProcessingResult(RsgbProxy.java:336) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.transformToTableData(RsgbProxy.java:372) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.handle(RsgbProxy.java:485) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:410) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:377) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:388) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:266) at brmo.loader/nl.b3p.brmo.loader.StagingProxy.handleBerichtenByJob(StagingProxy.java:377) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.run(RsgbProxy.java:289) at java.base/java.lang.Thread.run(Thread.java:834) BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#run:293) - Fout tijdens verwerken berichten java.lang.NullPointerException at brmo.loader/nl.b3p.brmo.loader.StagingProxy.updateBerichtProcessing(StagingProxy.java:689) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.updateProcessingResult(RsgbProxy.java:336) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.handle(RsgbProxy.java:581) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:410) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:377) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:388) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:266) at brmo.loader/nl.b3p.brmo.loader.StagingProxy.handleBerichtenByJob(StagingProxy.java:377) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.run(RsgbProxy.java:289) at java.base/java.lang.Thread.run(Thread.java:834) BRMO-LOADER-TEST: INFO 17:40:32 (n.b.AbstractDatabaseIntegrationTest#endTest:111) - ==== Einde test methode: loadAll() [ERROR] Tests run: 2, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.926 s <<< FAILURE! - in nl.b3p.brmo.loader.entity.BagBerichtIntegrationTest ``` </details>
1.0
Sommige integratie tests van de brmo-loader falen op Windows - ### Probleem Sommige integratie tests van de brmo-loader falen op Windows als zo'n test een bericht moet transformeren en de tests met java 11 en Maven draaien het gaat om: Failures: - `BRKLocatiebeschrijvingIntegrationTest#testLocatieBeschrijving` - `BrkToStagingToRsgbIntegrationTest#testBrkXMLToStaging` - `GH789GroteWaardeIDIntegrationTest#testTransformBerichten` - `Mantis10315IntegrationTest#testTransformBerichten` - `Mantis10547IntegrationTest#testAll` - `Mantis11180IntegrationTest#testTransformBerichten` - `Mantis6098IntegrationTest#testAll` - `Mantis6098IntegrationTest#testStand` - `Mantis6098IntegrationTest#testStandDelete` - `Mantis6098IntegrationTest#testStandDeleteMutatie` - `Mantis6098IntegrationTest#testStandMutatie` - `Mantis6166IntegrationTest#testAll` - `Mantis6166IntegrationTest#testStand` - `Mantis6166IntegrationTest#testStandDelete` - `Mantis6166IntegrationTest#testStandDeleteMutatie` - `Mantis6166IntegrationTest#testStandMutatie` - `Mantis6380IntegrationTest#testTransformBerichten` - `Mantis6380IntegrationTest#testTransformStand` - `Mantis6380IntegrationTest#testTransformStandDaarnaMutatieBerichten` - `VerminderenStukdelenIntegrationTest#testMinderStukdelenInMutatie` - `ZakRechtArchiefIntegrationTest#transformeerberichten` - `BagBerichtIntegrationTest#loadAll` ontdekt tijdens maken van #1025 (dus vooralsnog ronden we die af met java 8...) de kern van het probleem blijkt uit de melding: ``` BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#updateBerichtException:422) - Mutatieverwerking en foutief bericht, dus stoppen! java.nio.file.InvalidPathException: Illegal char <:> at index 21: org\geotools\gml3\jar:file:\C:\Users\mark\.m2\repository\org\geotools\xsd\gt-xsd-gml3\24.2\gt-xsd-gml3-24.2.jar!\org\geotools\gml3\gml.xsd at java.base/sun.nio.fs.WindowsPathParser.normalize(WindowsPathParser.java:182) ``` Er gaat dus iets mis met de constructie van het pad naar de xsd. De tests draaien wel goed vanuit de IDE / Intellij. ### stacktrace <details><summary>maven log:</summary> ``` [INFO] Running nl.b3p.brmo.loader.entity.BagBerichtIntegrationTest BRMO-LOADER-TEST: INFO 17:40:30 (n.b.AbstractDatabaseIntegrationTest#startTest:103) - ==== Start test methode: loadInStaging() BRMO-LOADER-TEST: INFO 17:40:31 (n.b.AbstractDatabaseIntegrationTest#endTest:111) - ==== Einde test methode: loadInStaging() BRMO-LOADER-TEST: INFO 17:40:31 (n.b.AbstractDatabaseIntegrationTest#startTest:103) - ==== Start test methode: loadAll() BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#updateBerichtException:422) - Mutatieverwerking en foutief bericht, dus stoppen! java.nio.file.InvalidPathException: Illegal char <:> at index 21: org\geotools\gml3\jar:file:\C:\Users\mark\.m2\repository\org\geotools\xsd\gt-xsd-gml3\24.2\gt-xsd-gml3-24.2.jar!\org\geotools\gml3\gml.xsd at java.base/sun.nio.fs.WindowsPathParser.normalize(WindowsPathParser.java:182) at java.base/sun.nio.fs.WindowsPathParser.parse(WindowsPathParser.java:153) at java.base/sun.nio.fs.WindowsPathParser.parse(WindowsPathParser.java:77) at java.base/sun.nio.fs.WindowsPath.parse(WindowsPath.java:92) at java.base/sun.nio.fs.WindowsFileSystem.getPath(WindowsFileSystem.java:229) at java.base/jdk.internal.module.Resources.toSafeFilePath(Resources.java:143) at java.base/jdk.internal.module.Resources.toFilePath(Resources.java:97) at java.base/jdk.internal.module.ModulePatcher$ExplodedResourceFinder.find(ModulePatcher.java:505) at java.base/jdk.internal.module.ModulePatcher$PatchedModuleReader.findResourceInPatch(ModulePatcher.java:301) at java.base/jdk.internal.module.ModulePatcher$PatchedModuleReader.find(ModulePatcher.java:357) at java.base/jdk.internal.loader.BuiltinClassLoader$2.run(BuiltinClassLoader.java:407) at java.base/jdk.internal.loader.BuiltinClassLoader$2.run(BuiltinClassLoader.java:402) at java.base/java.security.AccessController.doPrivileged(Native Method) at java.base/jdk.internal.loader.BuiltinClassLoader.findMiscResource(BuiltinClassLoader.java:401) at java.base/jdk.internal.loader.BuiltinClassLoader.findResource(BuiltinClassLoader.java:288) at java.base/java.lang.ClassLoader.getResource(ClassLoader.java:1401) at java.base/java.lang.Class.getResource(Class.java:2740) at org.geotools.xsd.SchemaLocationResolver.resolveLocationToResource(SchemaLocationResolver.java:111) at org.geotools.xsd.SchemaLocationResolver.canHandle(SchemaLocationResolver.java:102) at org.geotools.xsd.SchemaLocationResolver.resolveSchemaLocation(SchemaLocationResolver.java:151) at org.geotools.xsd.impl.ParserHandler.startElement(ParserHandler.java:399) at java.xml/com.sun.org.apache.xerces.internal.parsers.AbstractSAXParser.startElement(AbstractSAXParser.java:510) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.scanStartElement(XMLNSDocumentScannerImpl.java:374) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl$NSContentDriver.scanRootElementHook(XMLNSDocumentScannerImpl.java:613) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl$FragmentContentDriver.next(XMLDocumentFragmentScannerImpl.java:3063) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl$PrologDriver.next(XMLDocumentScannerImpl.java:836) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl.next(XMLDocumentScannerImpl.java:605) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.next(XMLNSDocumentScannerImpl.java:112) at java.xml/com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:534) at java.xml/com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:888) at java.xml/com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:824) at java.xml/com.sun.org.apache.xerces.internal.parsers.XMLParser.parse(XMLParser.java:141) at java.xml/com.sun.org.apache.xerces.internal.parsers.AbstractSAXParser.parse(AbstractSAXParser.java:1216) at java.xml/com.sun.org.apache.xerces.internal.jaxp.SAXParserImpl$JAXPSAXParser.parse(SAXParserImpl.java:635) at java.xml/com.sun.org.apache.xerces.internal.jaxp.SAXParserImpl.parse(SAXParserImpl.java:324) at org.geotools.xsd.Parser.parse(Parser.java:184) at org.geotools.xsd.Parser.parse(Parser.java:139) at brmo.loader/nl.b3p.brmo.loader.util.DataComfortXMLReader.readDataXML(DataComfortXMLReader.java:185) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.transformToTableData(RsgbProxy.java:362) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.handle(RsgbProxy.java:485) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:410) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:377) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:388) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:266) at brmo.loader/nl.b3p.brmo.loader.StagingProxy.handleBerichtenByJob(StagingProxy.java:377) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.run(RsgbProxy.java:289) at java.base/java.lang.Thread.run(Thread.java:834) BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#handle:560) - Fout bij verwerking bericht met id 41402969931, melding: null BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#updateBerichtException:422) - Mutatieverwerking en foutief bericht, dus stoppen! java.lang.NullPointerException at brmo.loader/nl.b3p.brmo.loader.StagingProxy.updateBerichtProcessing(StagingProxy.java:689) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.updateProcessingResult(RsgbProxy.java:336) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.transformToTableData(RsgbProxy.java:372) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.handle(RsgbProxy.java:485) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:410) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:377) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:388) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:266) at brmo.loader/nl.b3p.brmo.loader.StagingProxy.handleBerichtenByJob(StagingProxy.java:377) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.run(RsgbProxy.java:289) at java.base/java.lang.Thread.run(Thread.java:834) BRMO-LOADER-TEST: ERROR 17:40:32 (n.b.b.l.RsgbProxy#run:293) - Fout tijdens verwerken berichten java.lang.NullPointerException at brmo.loader/nl.b3p.brmo.loader.StagingProxy.updateBerichtProcessing(StagingProxy.java:689) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.updateProcessingResult(RsgbProxy.java:336) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.handle(RsgbProxy.java:581) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:410) at brmo.loader/nl.b3p.brmo.loader.StagingProxy$1.handle(StagingProxy.java:377) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:388) at org.apache.commons.dbutils.QueryRunner.query(QueryRunner.java:266) at brmo.loader/nl.b3p.brmo.loader.StagingProxy.handleBerichtenByJob(StagingProxy.java:377) at brmo.loader/nl.b3p.brmo.loader.RsgbProxy.run(RsgbProxy.java:289) at java.base/java.lang.Thread.run(Thread.java:834) BRMO-LOADER-TEST: INFO 17:40:32 (n.b.AbstractDatabaseIntegrationTest#endTest:111) - ==== Einde test methode: loadAll() [ERROR] Tests run: 2, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.926 s <<< FAILURE! - in nl.b3p.brmo.loader.entity.BagBerichtIntegrationTest ``` </details>
test
sommige integratie tests van de brmo loader falen op windows probleem sommige integratie tests van de brmo loader falen op windows als zo n test een bericht moet transformeren en de tests met java en maven draaien het gaat om failures brklocatiebeschrijvingintegrationtest testlocatiebeschrijving brktostagingtorsgbintegrationtest testbrkxmltostaging testtransformberichten testtransformberichten testall testtransformberichten testall teststand teststanddelete teststanddeletemutatie teststandmutatie testall teststand teststanddelete teststanddeletemutatie teststandmutatie testtransformberichten testtransformstand testtransformstanddaarnamutatieberichten verminderenstukdelenintegrationtest testminderstukdeleninmutatie zakrechtarchiefintegrationtest transformeerberichten bagberichtintegrationtest loadall ontdekt tijdens maken van dus vooralsnog ronden we die af met java de kern van het probleem blijkt uit de melding brmo loader test error n b b l rsgbproxy updateberichtexception mutatieverwerking en foutief bericht dus stoppen java nio file invalidpathexception illegal char at index org geotools jar file c users mark repository org geotools xsd gt xsd gt xsd jar org geotools gml xsd at java base sun nio fs windowspathparser normalize windowspathparser java er gaat dus iets mis met de constructie van het pad naar de xsd de tests draaien wel goed vanuit de ide intellij stacktrace maven log running nl brmo loader entity bagberichtintegrationtest brmo loader test info n b abstractdatabaseintegrationtest starttest start test methode loadinstaging brmo loader test info n b abstractdatabaseintegrationtest endtest einde test methode loadinstaging brmo loader test info n b abstractdatabaseintegrationtest starttest start test methode loadall brmo loader test error n b b l rsgbproxy updateberichtexception mutatieverwerking en foutief bericht dus stoppen java nio file invalidpathexception illegal char at index org geotools jar file c users mark repository org geotools xsd gt xsd gt xsd jar org geotools gml xsd at java base sun nio fs windowspathparser normalize windowspathparser java at java base sun nio fs windowspathparser parse windowspathparser java at java base sun nio fs windowspathparser parse windowspathparser java at java base sun nio fs windowspath parse windowspath java at java base sun nio fs windowsfilesystem getpath windowsfilesystem java at java base jdk internal module resources tosafefilepath resources java at java base jdk internal module resources tofilepath resources java at java base jdk internal module modulepatcher explodedresourcefinder find modulepatcher java at java base jdk internal module modulepatcher patchedmodulereader findresourceinpatch modulepatcher java at java base jdk internal module modulepatcher patchedmodulereader find modulepatcher java at java base jdk internal loader builtinclassloader run builtinclassloader java at java base jdk internal loader builtinclassloader run builtinclassloader java at java base java security accesscontroller doprivileged native method at java base jdk internal loader builtinclassloader findmiscresource builtinclassloader java at java base jdk internal loader builtinclassloader findresource builtinclassloader java at java base java lang classloader getresource classloader java at java base java lang class getresource class java at org geotools xsd schemalocationresolver resolvelocationtoresource schemalocationresolver java at org geotools xsd schemalocationresolver canhandle schemalocationresolver java at org geotools xsd schemalocationresolver resolveschemalocation schemalocationresolver java at org geotools xsd impl parserhandler startelement parserhandler java at java xml com sun org apache xerces internal parsers abstractsaxparser startelement abstractsaxparser java at java xml com sun org apache xerces internal impl xmlnsdocumentscannerimpl scanstartelement xmlnsdocumentscannerimpl java at java xml com sun org apache xerces internal impl xmlnsdocumentscannerimpl nscontentdriver scanrootelementhook xmlnsdocumentscannerimpl java at java xml com sun org apache xerces internal impl xmldocumentfragmentscannerimpl fragmentcontentdriver next xmldocumentfragmentscannerimpl java at java xml com sun org apache xerces internal impl xmldocumentscannerimpl prologdriver next xmldocumentscannerimpl java at java xml com sun org apache xerces internal impl xmldocumentscannerimpl next xmldocumentscannerimpl java at java xml com sun org apache xerces internal impl xmlnsdocumentscannerimpl next xmlnsdocumentscannerimpl java at java xml com sun org apache xerces internal impl xmldocumentfragmentscannerimpl scandocument xmldocumentfragmentscannerimpl java at java xml com sun org apache xerces internal parsers parse java at java xml com sun org apache xerces internal parsers parse java at java xml com sun org apache xerces internal parsers xmlparser parse xmlparser java at java xml com sun org apache xerces internal parsers abstractsaxparser parse abstractsaxparser java at java xml com sun org apache xerces internal jaxp saxparserimpl jaxpsaxparser parse saxparserimpl java at java xml com sun org apache xerces internal jaxp saxparserimpl parse saxparserimpl java at org geotools xsd parser parse parser java at org geotools xsd parser parse parser java at brmo loader nl brmo loader util datacomfortxmlreader readdataxml datacomfortxmlreader java at brmo loader nl brmo loader rsgbproxy transformtotabledata rsgbproxy java at brmo loader nl brmo loader rsgbproxy handle rsgbproxy java at brmo loader nl brmo loader stagingproxy handle stagingproxy java at brmo loader nl brmo loader stagingproxy handle stagingproxy java at org apache commons dbutils queryrunner query queryrunner java at org apache commons dbutils queryrunner query queryrunner java at brmo loader nl brmo loader stagingproxy handleberichtenbyjob stagingproxy java at brmo loader nl brmo loader rsgbproxy run rsgbproxy java at java base java lang thread run thread java brmo loader test error n b b l rsgbproxy handle fout bij verwerking bericht met id melding null brmo loader test error n b b l rsgbproxy updateberichtexception mutatieverwerking en foutief bericht dus stoppen java lang nullpointerexception at brmo loader nl brmo loader stagingproxy updateberichtprocessing stagingproxy java at brmo loader nl brmo loader rsgbproxy updateprocessingresult rsgbproxy java at brmo loader nl brmo loader rsgbproxy transformtotabledata rsgbproxy java at brmo loader nl brmo loader rsgbproxy handle rsgbproxy java at brmo loader nl brmo loader stagingproxy handle stagingproxy java at brmo loader nl brmo loader stagingproxy handle stagingproxy java at org apache commons dbutils queryrunner query queryrunner java at org apache commons dbutils queryrunner query queryrunner java at brmo loader nl brmo loader stagingproxy handleberichtenbyjob stagingproxy java at brmo loader nl brmo loader rsgbproxy run rsgbproxy java at java base java lang thread run thread java brmo loader test error n b b l rsgbproxy run fout tijdens verwerken berichten java lang nullpointerexception at brmo loader nl brmo loader stagingproxy updateberichtprocessing stagingproxy java at brmo loader nl brmo loader rsgbproxy updateprocessingresult rsgbproxy java at brmo loader nl brmo loader rsgbproxy handle rsgbproxy java at brmo loader nl brmo loader stagingproxy handle stagingproxy java at brmo loader nl brmo loader stagingproxy handle stagingproxy java at org apache commons dbutils queryrunner query queryrunner java at org apache commons dbutils queryrunner query queryrunner java at brmo loader nl brmo loader stagingproxy handleberichtenbyjob stagingproxy java at brmo loader nl brmo loader rsgbproxy run rsgbproxy java at java base java lang thread run thread java brmo loader test info n b abstractdatabaseintegrationtest endtest einde test methode loadall tests run failures errors skipped time elapsed s failure in nl brmo loader entity bagberichtintegrationtest
1
8,736
2,611,542,499
IssuesEvent
2015-02-27 06:11:20
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
opened
Saving a custom weapon scheme resets selection to Default
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Create a custom weapon scheme 2. Leave the menu, thus saving the weapon scheme 3. See what weapon scheme is now selected What is the expected output? What do you see instead? I expect a) That the frontend goes without further ado back to the game menu b) That my custom weapon scheme is selected in the game menu. Actually a) The frontend seems to behave a bit “messy”: For a brief time, all fields (weapon scheme name, number of weapons, wheather the fields are “grayed out” or not) wildly change. This takes about 1-2 seconds b) The game menu has *always* reset my selection of the weapon scheme to Default What version of the product are you using? On what operating system? I use GNU/Linux and can reproduce this in the following Hedgewars versions: - (probably some earlier versions than 0.9.20) - 0.9.20 - Git f488421… Please provide any additional information below. If you repeat the steps but leave the weapon scheme editor with a default (non-custom) weapon scheme, the weapon editor seems to behave OK. So make sure you do the test with a custom weapon scheme. ``` Original issue reported on code.google.com by `almikes@aol.com` on 10 Dec 2014 at 5:58
1.0
Saving a custom weapon scheme resets selection to Default - ``` What steps will reproduce the problem? 1. Create a custom weapon scheme 2. Leave the menu, thus saving the weapon scheme 3. See what weapon scheme is now selected What is the expected output? What do you see instead? I expect a) That the frontend goes without further ado back to the game menu b) That my custom weapon scheme is selected in the game menu. Actually a) The frontend seems to behave a bit “messy”: For a brief time, all fields (weapon scheme name, number of weapons, wheather the fields are “grayed out” or not) wildly change. This takes about 1-2 seconds b) The game menu has *always* reset my selection of the weapon scheme to Default What version of the product are you using? On what operating system? I use GNU/Linux and can reproduce this in the following Hedgewars versions: - (probably some earlier versions than 0.9.20) - 0.9.20 - Git f488421… Please provide any additional information below. If you repeat the steps but leave the weapon scheme editor with a default (non-custom) weapon scheme, the weapon editor seems to behave OK. So make sure you do the test with a custom weapon scheme. ``` Original issue reported on code.google.com by `almikes@aol.com` on 10 Dec 2014 at 5:58
non_test
saving a custom weapon scheme resets selection to default what steps will reproduce the problem create a custom weapon scheme leave the menu thus saving the weapon scheme see what weapon scheme is now selected what is the expected output what do you see instead i expect a that the frontend goes without further ado back to the game menu b that my custom weapon scheme is selected in the game menu actually a the frontend seems to behave a bit “messy” for a brief time all fields weapon scheme name number of weapons wheather the fields are “grayed out” or not wildly change this takes about seconds b the game menu has always reset my selection of the weapon scheme to default what version of the product are you using on what operating system i use gnu linux and can reproduce this in the following hedgewars versions probably some earlier versions than git … please provide any additional information below if you repeat the steps but leave the weapon scheme editor with a default non custom weapon scheme the weapon editor seems to behave ok so make sure you do the test with a custom weapon scheme original issue reported on code google com by almikes aol com on dec at
0
168,362
13,081,976,483
IssuesEvent
2020-08-01 12:57:36
ValveSoftware/SteamVR-for-Linux
https://api.github.com/repos/ValveSoftware/SteamVR-for-Linux
closed
Unable to Get Past Vive Room Setup on Ubuntu 17.10
Need Retest
#### Your system information * Steam client version (build number or date): Oct 30 2017, at 19:09:46 * Distribution (e.g. Ubuntu): Ubuntu 17.10 * Graphics driver version (run nvidia-settings): 387.34 * Gist for SteamVR System Information: https://gist.github.com/georgewsinger/429d151ef68cb4f303377595f31e00b2 * Opted into Steam client beta?: No * Opted into SteamVR beta?: Yes * Have you checked for system updates?: Yes #### Please describe your issue in as much detail as possible: I'm unable to complete the Vive Room Setup on Ubuntu 17.10 using `nvidia-387` drivers. Specifically, it says "Headset Not Visible" even though (i) the headset does seem to be recognized in the SteamVR Status menu and (ii) the headset is visible when I launch the SteamVR Room Setup in Windows using the same machine. Here's what it looks like in Ubuntu: ![pic](https://i.imgur.com/bmDlJ1u.png) And here is the tail end of my `vrserver.txt` log: ``` Thu Dec 07 2017 11:47:59.125159 - lighthouse: VRC-D4B105EBBE C: ----- BOOTSTRAPPED base 775ADEB0 (immediate) distance 3.13m velocity 0.09m/s recorded pitch ~24.0 deg roll ~-15.9 deg ----- Thu Dec 07 2017 11:48:00.316710 - lighthouse: LHR-FF421D45 C: Trying to add a secondary base E2F61B3C: Samples didn't yield successful bootstrap pose Thu Dec 07 2017 11:48:01.176479 - lighthouse: VRC-D4B105EBBE C: Calibration failed: no optical samples from base 775ADEB0 for 2001ms Thu Dec 07 2017 11:48:06.184635 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base 775ADEB0: Not enough contiguous samples for a bootstrap pose Thu Dec 07 2017 11:48:06.184669 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base E2F61B3C: Samples didn't yield successful bootstrap pose Thu Dec 07 2017 11:48:07.370937 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base E2F61B3C: Could boot, but disallowed by LighthouseDB Thu Dec 07 2017 11:48:08.924515 - 1 - in use time less than now: 0x486e707ae3fe < 0x486e707ae04c Thu Dec 07 2017 11:48:14.909500 - lighthouse: LHR-FF421D45 C: Resetting tracking: no optical samples from base 775ADEB0 for 2000ms Thu Dec 07 2017 11:48:14.909542 - lighthouse: Stopped tracking with universe 1512584546 Thu Dec 07 2017 11:48:16.125024 - lighthouse: LHR-FF421D45 C: ----- BOOTSTRAPPED base 775ADEB0 (immediate) distance 3.16m velocity 0.01m/s recorded pitch ~24.0 deg roll ~-15.9 deg ----- Thu Dec 07 2017 11:48:21.226713 - lighthouse: LHR-FF421D45 C: ----- CALIBRATED base 775ADEB0 at pitch 23.59 deg roll -15.90 deg ----- Thu Dec 07 2017 11:48:21.226747 - lighthouse: Updating tilt for 775ADEB0.1. Old= 24.01, -15.94, 3.27743e-06 <- 0.42 deg -> new= 23.59, -15.90, 4.05965e-06 result=23.59, -15.90, 4.05965e-06 Thu Dec 07 2017 11:48:21.226756 - lighthouse: BootstrapFinished setting tilt base to 775ADEB0 Thu Dec 07 2017 11:48:21.226761 - lighthouse: Selected existing universe 1512584546 (775ADEB0 is primary) Thu Dec 07 2017 11:48:21.226770 - lighthouse: Setting tilt from 775ADEB0 via transform to global: pitch 23.59 deg roll -15.90 deg Thu Dec 07 2017 11:48:21.230145 - lighthouse: Saved LighthouseDB rev 198 to /home/carl/.steam/steam/config/lighthouse/lighthousedb.json Thu Dec 07 2017 11:48:24.911844 - lighthouse: LHR-FF421D45 C: Trying to add a secondary base E2F61B3C: Samples didn't yield successful bootstrap pose Thu Dec 07 2017 11:48:31.189461 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base 775ADEB0: Samples didn't yield successful bootstrap pose Thu Dec 07 2017 11:48:32.058936 - lighthouse: VRC-D4B105EBBE C: ----- BOOTSTRAPPED base 775ADEB0 (immediate) distance 3.13m velocity 0.01m/s recorded pitch ~23.6 deg roll ~-15.9 deg ----- Thu Dec 07 2017 11:48:32.135386 - 2 - in use time less than now: 0x4873d7f4a5e0 < 0x4873d7f4a567 Thu Dec 07 2017 11:48:33.913399 - 2 - in use time less than now: 0x487441eef6fe < 0x487441eef504 Thu Dec 07 2017 11:48:34.146779 - lighthouse: VRC-D4B105EBBE C: Calibration failed: no optical samples from base 775ADEB0 for 2004ms Thu Dec 07 2017 11:48:36.775038 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base E2F61B3C: Could boot, but disallowed by LighthouseDB Thu Dec 07 2017 11:48:39.153008 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base 775ADEB0: Not enough contiguous samples for a bootstrap pose Thu Dec 07 2017 11:48:44.845123 - lighthouse: LHR-FF421D45 C: Resetting tracking: no optical samples from base 775ADEB0 for 2002ms Thu Dec 07 2017 11:48:44.845166 - lighthouse: Stopped tracking with universe 1512584546 Thu Dec 07 2017 11:48:49.853283 - lighthouse: LHR-FF421D45 C: Trying to start tracking from base 775ADEB0: Not enough contiguous samples for a bootstrap pose ``` **NOTE:** I also confirmed that my Vive udev rules have been written properly, restarted, and am still getting this error. #### Steps for reproducing this issue: 1. Update Ubuntu 16.04 to Ubuntu 17.10, and install `nvidia-387` package. 2. Launch Vive Room Setup 3. Receive error in picture above.
1.0
Unable to Get Past Vive Room Setup on Ubuntu 17.10 - #### Your system information * Steam client version (build number or date): Oct 30 2017, at 19:09:46 * Distribution (e.g. Ubuntu): Ubuntu 17.10 * Graphics driver version (run nvidia-settings): 387.34 * Gist for SteamVR System Information: https://gist.github.com/georgewsinger/429d151ef68cb4f303377595f31e00b2 * Opted into Steam client beta?: No * Opted into SteamVR beta?: Yes * Have you checked for system updates?: Yes #### Please describe your issue in as much detail as possible: I'm unable to complete the Vive Room Setup on Ubuntu 17.10 using `nvidia-387` drivers. Specifically, it says "Headset Not Visible" even though (i) the headset does seem to be recognized in the SteamVR Status menu and (ii) the headset is visible when I launch the SteamVR Room Setup in Windows using the same machine. Here's what it looks like in Ubuntu: ![pic](https://i.imgur.com/bmDlJ1u.png) And here is the tail end of my `vrserver.txt` log: ``` Thu Dec 07 2017 11:47:59.125159 - lighthouse: VRC-D4B105EBBE C: ----- BOOTSTRAPPED base 775ADEB0 (immediate) distance 3.13m velocity 0.09m/s recorded pitch ~24.0 deg roll ~-15.9 deg ----- Thu Dec 07 2017 11:48:00.316710 - lighthouse: LHR-FF421D45 C: Trying to add a secondary base E2F61B3C: Samples didn't yield successful bootstrap pose Thu Dec 07 2017 11:48:01.176479 - lighthouse: VRC-D4B105EBBE C: Calibration failed: no optical samples from base 775ADEB0 for 2001ms Thu Dec 07 2017 11:48:06.184635 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base 775ADEB0: Not enough contiguous samples for a bootstrap pose Thu Dec 07 2017 11:48:06.184669 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base E2F61B3C: Samples didn't yield successful bootstrap pose Thu Dec 07 2017 11:48:07.370937 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base E2F61B3C: Could boot, but disallowed by LighthouseDB Thu Dec 07 2017 11:48:08.924515 - 1 - in use time less than now: 0x486e707ae3fe < 0x486e707ae04c Thu Dec 07 2017 11:48:14.909500 - lighthouse: LHR-FF421D45 C: Resetting tracking: no optical samples from base 775ADEB0 for 2000ms Thu Dec 07 2017 11:48:14.909542 - lighthouse: Stopped tracking with universe 1512584546 Thu Dec 07 2017 11:48:16.125024 - lighthouse: LHR-FF421D45 C: ----- BOOTSTRAPPED base 775ADEB0 (immediate) distance 3.16m velocity 0.01m/s recorded pitch ~24.0 deg roll ~-15.9 deg ----- Thu Dec 07 2017 11:48:21.226713 - lighthouse: LHR-FF421D45 C: ----- CALIBRATED base 775ADEB0 at pitch 23.59 deg roll -15.90 deg ----- Thu Dec 07 2017 11:48:21.226747 - lighthouse: Updating tilt for 775ADEB0.1. Old= 24.01, -15.94, 3.27743e-06 <- 0.42 deg -> new= 23.59, -15.90, 4.05965e-06 result=23.59, -15.90, 4.05965e-06 Thu Dec 07 2017 11:48:21.226756 - lighthouse: BootstrapFinished setting tilt base to 775ADEB0 Thu Dec 07 2017 11:48:21.226761 - lighthouse: Selected existing universe 1512584546 (775ADEB0 is primary) Thu Dec 07 2017 11:48:21.226770 - lighthouse: Setting tilt from 775ADEB0 via transform to global: pitch 23.59 deg roll -15.90 deg Thu Dec 07 2017 11:48:21.230145 - lighthouse: Saved LighthouseDB rev 198 to /home/carl/.steam/steam/config/lighthouse/lighthousedb.json Thu Dec 07 2017 11:48:24.911844 - lighthouse: LHR-FF421D45 C: Trying to add a secondary base E2F61B3C: Samples didn't yield successful bootstrap pose Thu Dec 07 2017 11:48:31.189461 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base 775ADEB0: Samples didn't yield successful bootstrap pose Thu Dec 07 2017 11:48:32.058936 - lighthouse: VRC-D4B105EBBE C: ----- BOOTSTRAPPED base 775ADEB0 (immediate) distance 3.13m velocity 0.01m/s recorded pitch ~23.6 deg roll ~-15.9 deg ----- Thu Dec 07 2017 11:48:32.135386 - 2 - in use time less than now: 0x4873d7f4a5e0 < 0x4873d7f4a567 Thu Dec 07 2017 11:48:33.913399 - 2 - in use time less than now: 0x487441eef6fe < 0x487441eef504 Thu Dec 07 2017 11:48:34.146779 - lighthouse: VRC-D4B105EBBE C: Calibration failed: no optical samples from base 775ADEB0 for 2004ms Thu Dec 07 2017 11:48:36.775038 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base E2F61B3C: Could boot, but disallowed by LighthouseDB Thu Dec 07 2017 11:48:39.153008 - lighthouse: VRC-D4B105EBBE C: Trying to start tracking from base 775ADEB0: Not enough contiguous samples for a bootstrap pose Thu Dec 07 2017 11:48:44.845123 - lighthouse: LHR-FF421D45 C: Resetting tracking: no optical samples from base 775ADEB0 for 2002ms Thu Dec 07 2017 11:48:44.845166 - lighthouse: Stopped tracking with universe 1512584546 Thu Dec 07 2017 11:48:49.853283 - lighthouse: LHR-FF421D45 C: Trying to start tracking from base 775ADEB0: Not enough contiguous samples for a bootstrap pose ``` **NOTE:** I also confirmed that my Vive udev rules have been written properly, restarted, and am still getting this error. #### Steps for reproducing this issue: 1. Update Ubuntu 16.04 to Ubuntu 17.10, and install `nvidia-387` package. 2. Launch Vive Room Setup 3. Receive error in picture above.
test
unable to get past vive room setup on ubuntu your system information steam client version build number or date oct at distribution e g ubuntu ubuntu graphics driver version run nvidia settings gist for steamvr system information opted into steam client beta no opted into steamvr beta yes have you checked for system updates yes please describe your issue in as much detail as possible i m unable to complete the vive room setup on ubuntu using nvidia drivers specifically it says headset not visible even though i the headset does seem to be recognized in the steamvr status menu and ii the headset is visible when i launch the steamvr room setup in windows using the same machine here s what it looks like in ubuntu and here is the tail end of my vrserver txt log thu dec lighthouse vrc c bootstrapped base immediate distance velocity s recorded pitch deg roll deg thu dec lighthouse lhr c trying to add a secondary base samples didn t yield successful bootstrap pose thu dec lighthouse vrc c calibration failed no optical samples from base for thu dec lighthouse vrc c trying to start tracking from base not enough contiguous samples for a bootstrap pose thu dec lighthouse vrc c trying to start tracking from base samples didn t yield successful bootstrap pose thu dec lighthouse vrc c trying to start tracking from base could boot but disallowed by lighthousedb thu dec in use time less than now thu dec lighthouse lhr c resetting tracking no optical samples from base for thu dec lighthouse stopped tracking with universe thu dec lighthouse lhr c bootstrapped base immediate distance velocity s recorded pitch deg roll deg thu dec lighthouse lhr c calibrated base at pitch deg roll deg thu dec lighthouse updating tilt for old new result thu dec lighthouse bootstrapfinished setting tilt base to thu dec lighthouse selected existing universe is primary thu dec lighthouse setting tilt from via transform to global pitch deg roll deg thu dec lighthouse saved lighthousedb rev to home carl steam steam config lighthouse lighthousedb json thu dec lighthouse lhr c trying to add a secondary base samples didn t yield successful bootstrap pose thu dec lighthouse vrc c trying to start tracking from base samples didn t yield successful bootstrap pose thu dec lighthouse vrc c bootstrapped base immediate distance velocity s recorded pitch deg roll deg thu dec in use time less than now thu dec in use time less than now thu dec lighthouse vrc c calibration failed no optical samples from base for thu dec lighthouse vrc c trying to start tracking from base could boot but disallowed by lighthousedb thu dec lighthouse vrc c trying to start tracking from base not enough contiguous samples for a bootstrap pose thu dec lighthouse lhr c resetting tracking no optical samples from base for thu dec lighthouse stopped tracking with universe thu dec lighthouse lhr c trying to start tracking from base not enough contiguous samples for a bootstrap pose note i also confirmed that my vive udev rules have been written properly restarted and am still getting this error steps for reproducing this issue update ubuntu to ubuntu and install nvidia package launch vive room setup receive error in picture above
1
306,604
26,483,763,744
IssuesEvent
2023-01-17 16:26:03
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix exponents_and_logarithms.test_numpy_log2
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3924549296/jobs/6708877461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3924549296/jobs/6708877461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3924549296/jobs/6708877461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3924549296/jobs/6708877461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_mathematical_functions/test_exponents_and_logarithms.py::test_numpy_log2[cpu-ivy.functional.backends.tensorflow-False-False]</summary> 2023-01-15T17:58:27.8336559Z E AssertionError: [-inf] != [nan] 2023-01-15T17:58:27.8336826Z E Falsifying example: test_numpy_log2( 2023-01-15T17:58:27.8337152Z E dtypes_values_casting=(['float64'], 2023-01-15T17:58:27.8337452Z E [array(-5.87747175e-39)], 2023-01-15T17:58:27.8337692Z E 'unsafe', 2023-01-15T17:58:27.8337922Z E 'float32'), 2023-01-15T17:58:27.8338114Z E where=True, 2023-01-15T17:58:27.8338326Z E num_positional_args=1, 2023-01-15T17:58:27.8338548Z E as_variable=[False], 2023-01-15T17:58:27.8338764Z E native_array=[False], 2023-01-15T17:58:27.8339173Z E with_out=False, 2023-01-15T17:58:27.8339524Z E fn_tree='ivy.functional.frontends.numpy.log2', 2023-01-15T17:58:27.8339837Z E frontend='numpy', 2023-01-15T17:58:27.8340084Z E on_device='cpu', 2023-01-15T17:58:27.8340276Z E ) 2023-01-15T17:58:27.8340441Z E 2023-01-15T17:58:27.8340970Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2FiAALG4w8YoIARTAIAGkYBsA==') as a decorator on your test case </details>
1.0
Fix exponents_and_logarithms.test_numpy_log2 - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3924549296/jobs/6708877461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3924549296/jobs/6708877461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3924549296/jobs/6708877461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3924549296/jobs/6708877461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_mathematical_functions/test_exponents_and_logarithms.py::test_numpy_log2[cpu-ivy.functional.backends.tensorflow-False-False]</summary> 2023-01-15T17:58:27.8336559Z E AssertionError: [-inf] != [nan] 2023-01-15T17:58:27.8336826Z E Falsifying example: test_numpy_log2( 2023-01-15T17:58:27.8337152Z E dtypes_values_casting=(['float64'], 2023-01-15T17:58:27.8337452Z E [array(-5.87747175e-39)], 2023-01-15T17:58:27.8337692Z E 'unsafe', 2023-01-15T17:58:27.8337922Z E 'float32'), 2023-01-15T17:58:27.8338114Z E where=True, 2023-01-15T17:58:27.8338326Z E num_positional_args=1, 2023-01-15T17:58:27.8338548Z E as_variable=[False], 2023-01-15T17:58:27.8338764Z E native_array=[False], 2023-01-15T17:58:27.8339173Z E with_out=False, 2023-01-15T17:58:27.8339524Z E fn_tree='ivy.functional.frontends.numpy.log2', 2023-01-15T17:58:27.8339837Z E frontend='numpy', 2023-01-15T17:58:27.8340084Z E on_device='cpu', 2023-01-15T17:58:27.8340276Z E ) 2023-01-15T17:58:27.8340441Z E 2023-01-15T17:58:27.8340970Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2FiAALG4w8YoIARTAIAGkYBsA==') as a decorator on your test case </details>
test
fix exponents and logarithms test numpy tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test numpy test mathematical functions test exponents and logarithms py test numpy e assertionerror e falsifying example test numpy e dtypes values casting e e unsafe e e where true e num positional args e as variable e native array e with out false e fn tree ivy functional frontends numpy e frontend numpy e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
349,353
31,795,299,494
IssuesEvent
2023-09-13 07:36:08
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
pkg/sql/logictest/tests/local/local_test: TestLogic_zone_config_system_tenant failed
C-test-failure O-robot release-blocker T-sql-queries branch-release-23.1.11-rc
pkg/sql/logictest/tests/local/local_test.TestLogic_zone_config_system_tenant [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SqlLogicTestHighVModuleNightlyBazel/11754215?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SqlLogicTestHighVModuleNightlyBazel/11754215?buildTab=artifacts#/) on release-23.1.11-rc @ [81bc6bd4352cc022db883327b122bab24e26ed29](https://github.com/cockroachdb/cockroach/commits/81bc6bd4352cc022db883327b122bab24e26ed29): ``` Purge passes: 0 (0 / sec) Purges: 0 (0 / sec) Hugeifies: 0 (0 / sec) Dehugifies: 0 (0 / sec) In full slabs: npageslabs: 0 huge, 0 nonhuge nactive: 0 huge, 0 nonhuge ndirty: 0 huge, 0 nonhuge nretained: 0 huge, 0 nonhuge In empty slabs: npageslabs: 0 huge, 0 nonhuge nactive: 0 huge, 0 nonhuge ndirty: 0 huge, 0 nonhuge nretained: 0 huge, 0 nonhuge size ind npageslabs_huge nactive_huge ndirty_huge npageslabs_nonhuge nactive_nonhuge ndirty_nonhuge nretained_nonhuge --- --- End jemalloc statistics --- [07:27:11] -- OK; [07:27:11] SELECT crdb_internal.pb_to_json('cockroach.roachpb.SpanConfig', config)->'gcPolicy'->>'ttlSeconds' FROM system.span_configurations WHERE start_key >= (SELECT crdb_internal.table_span(100)[1]) ORDER BY start_key; [07:27:11] -- FAIL logic.go:2887: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3950/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/logictest/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/logictest/testdata/logic_test/zone_config_system_tenant:153: SELECT crdb_internal.pb_to_json('cockroach.roachpb.SpanConfig', config)->'gcPolicy'->>'ttlSeconds' FROM system.span_configurations WHERE start_key >= (SELECT crdb_internal.table_span(100)[1]) ORDER BY start_key expected: 14400 14400 90001 90001 but found (query options: "") : 14400 14400 90001 90001 14400 [07:27:11] -- FAIL logic.go:2146: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3950/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/logictest/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/logictest/testdata/logic_test/zone_config_system_tenant:164: too many errors encountered, skipping the rest of the input [07:27:11] --- done: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3950/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/logictest/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/logictest/testdata/logic_test/zone_config_system_tenant with config local: 31 tests, 2 failures [07:27:15] --- total progress: 31 statements --- total: 31 tests, 2 failures --- FAIL: TestLogic_zone_config_system_tenant/dropped_relation_gc_ttl (3.64s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLogic_zone_config_system_tenant.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
pkg/sql/logictest/tests/local/local_test: TestLogic_zone_config_system_tenant failed - pkg/sql/logictest/tests/local/local_test.TestLogic_zone_config_system_tenant [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SqlLogicTestHighVModuleNightlyBazel/11754215?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SqlLogicTestHighVModuleNightlyBazel/11754215?buildTab=artifacts#/) on release-23.1.11-rc @ [81bc6bd4352cc022db883327b122bab24e26ed29](https://github.com/cockroachdb/cockroach/commits/81bc6bd4352cc022db883327b122bab24e26ed29): ``` Purge passes: 0 (0 / sec) Purges: 0 (0 / sec) Hugeifies: 0 (0 / sec) Dehugifies: 0 (0 / sec) In full slabs: npageslabs: 0 huge, 0 nonhuge nactive: 0 huge, 0 nonhuge ndirty: 0 huge, 0 nonhuge nretained: 0 huge, 0 nonhuge In empty slabs: npageslabs: 0 huge, 0 nonhuge nactive: 0 huge, 0 nonhuge ndirty: 0 huge, 0 nonhuge nretained: 0 huge, 0 nonhuge size ind npageslabs_huge nactive_huge ndirty_huge npageslabs_nonhuge nactive_nonhuge ndirty_nonhuge nretained_nonhuge --- --- End jemalloc statistics --- [07:27:11] -- OK; [07:27:11] SELECT crdb_internal.pb_to_json('cockroach.roachpb.SpanConfig', config)->'gcPolicy'->>'ttlSeconds' FROM system.span_configurations WHERE start_key >= (SELECT crdb_internal.table_span(100)[1]) ORDER BY start_key; [07:27:11] -- FAIL logic.go:2887: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3950/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/logictest/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/logictest/testdata/logic_test/zone_config_system_tenant:153: SELECT crdb_internal.pb_to_json('cockroach.roachpb.SpanConfig', config)->'gcPolicy'->>'ttlSeconds' FROM system.span_configurations WHERE start_key >= (SELECT crdb_internal.table_span(100)[1]) ORDER BY start_key expected: 14400 14400 90001 90001 but found (query options: "") : 14400 14400 90001 90001 14400 [07:27:11] -- FAIL logic.go:2146: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3950/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/logictest/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/logictest/testdata/logic_test/zone_config_system_tenant:164: too many errors encountered, skipping the rest of the input [07:27:11] --- done: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3950/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/logictest/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/logictest/testdata/logic_test/zone_config_system_tenant with config local: 31 tests, 2 failures [07:27:15] --- total progress: 31 statements --- total: 31 tests, 2 failures --- FAIL: TestLogic_zone_config_system_tenant/dropped_relation_gc_ttl (3.64s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLogic_zone_config_system_tenant.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
pkg sql logictest tests local local test testlogic zone config system tenant failed pkg sql logictest tests local local test testlogic zone config system tenant with on release rc purge passes sec purges sec hugeifies sec dehugifies sec in full slabs npageslabs huge nonhuge nactive huge nonhuge ndirty huge nonhuge nretained huge nonhuge in empty slabs npageslabs huge nonhuge nactive huge nonhuge ndirty huge nonhuge nretained huge nonhuge size ind npageslabs huge nactive huge ndirty huge npageslabs nonhuge nactive nonhuge ndirty nonhuge nretained nonhuge end jemalloc statistics ok select crdb internal pb to json cockroach roachpb spanconfig config gcpolicy ttlseconds from system span configurations where start key select crdb internal table span order by start key fail logic go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql logictest tests local local test local test runfiles com github cockroachdb cockroach pkg sql logictest testdata logic test zone config system tenant select crdb internal pb to json cockroach roachpb spanconfig config gcpolicy ttlseconds from system span configurations where start key select crdb internal table span order by start key expected but found query options fail logic go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql logictest tests local local test local test runfiles com github cockroachdb cockroach pkg sql logictest testdata logic test zone config system tenant too many errors encountered skipping the rest of the input done home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql logictest tests local local test local test runfiles com github cockroachdb cockroach pkg sql logictest testdata logic test zone config system tenant with config local tests failures total progress statements total tests failures fail testlogic zone config system tenant dropped relation gc ttl help see also cc cockroachdb sql queries
1
239,285
18,268,199,231
IssuesEvent
2021-10-04 10:57:40
ConorMFinn/CO600-SeriousCyberSec
https://api.github.com/repos/ConorMFinn/CO600-SeriousCyberSec
closed
Create a Video showcasing the game
documentation report
## Expected Behaviour Part of the final corpus of materials There should be a video created that showcases how the game plays and how it works. ## Actual Behaviour Missing ## Steps to Reproduce Behaviour 1. Check not there.
1.0
Create a Video showcasing the game - ## Expected Behaviour Part of the final corpus of materials There should be a video created that showcases how the game plays and how it works. ## Actual Behaviour Missing ## Steps to Reproduce Behaviour 1. Check not there.
non_test
create a video showcasing the game expected behaviour part of the final corpus of materials there should be a video created that showcases how the game plays and how it works actual behaviour missing steps to reproduce behaviour check not there
0
424,144
29,045,246,911
IssuesEvent
2023-05-13 13:19:54
SergiyKochenko/pp5-house-stock-ecommerce
https://api.github.com/repos/SergiyKochenko/pp5-house-stock-ecommerce
closed
USER STORY: Add authentication and Email backends for Email account verification
documentation Should Have
As a Purchasing Customer User, I want to verify my account by email in order to improve the security of my account and receive feedback that the registration process was successful. Acceptance Criteria 1: The user is asked to verify their email address upon login if unverified. 2: The user is redirected to a temporary success URL if the email address is verified. Unit Tasks * [x] Task/Chore : Create email verification functionality by making changes to AllAuth Account settings in settings.py * [x] Task : Test Feature is functioning correctly by adding 'success' to redirect urls.
1.0
USER STORY: Add authentication and Email backends for Email account verification - As a Purchasing Customer User, I want to verify my account by email in order to improve the security of my account and receive feedback that the registration process was successful. Acceptance Criteria 1: The user is asked to verify their email address upon login if unverified. 2: The user is redirected to a temporary success URL if the email address is verified. Unit Tasks * [x] Task/Chore : Create email verification functionality by making changes to AllAuth Account settings in settings.py * [x] Task : Test Feature is functioning correctly by adding 'success' to redirect urls.
non_test
user story add authentication and email backends for email account verification as a purchasing customer user i want to verify my account by email in order to improve the security of my account and receive feedback that the registration process was successful acceptance criteria the user is asked to verify their email address upon login if unverified the user is redirected to a temporary success url if the email address is verified unit tasks task chore create email verification functionality by making changes to allauth account settings in settings py task test feature is functioning correctly by adding success to redirect urls
0
22,462
3,960,395,455
IssuesEvent
2016-05-02 06:18:11
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
closed
Copy in File Explorer does not paste into Code Explorer
files question v-test
- VSCode Version: 1.1.0-alpha - OS Version: Windows 10 Steps to Reproduce: 1. Right click > copy or Ctrl+C a file in File Explorer 2. Go to Code, Right click > Paste or Ctrl+V in Explorer Actual: Nothing happens unless you already copied a file in Code, then it will paste that instead. I read some other issues similar to this, but I'm wondering why dragging files from File Explorer into Code works and this doesn't. I'm guessing an Electron issue?
1.0
Copy in File Explorer does not paste into Code Explorer - - VSCode Version: 1.1.0-alpha - OS Version: Windows 10 Steps to Reproduce: 1. Right click > copy or Ctrl+C a file in File Explorer 2. Go to Code, Right click > Paste or Ctrl+V in Explorer Actual: Nothing happens unless you already copied a file in Code, then it will paste that instead. I read some other issues similar to this, but I'm wondering why dragging files from File Explorer into Code works and this doesn't. I'm guessing an Electron issue?
test
copy in file explorer does not paste into code explorer vscode version alpha os version windows steps to reproduce right click copy or ctrl c a file in file explorer go to code right click paste or ctrl v in explorer actual nothing happens unless you already copied a file in code then it will paste that instead i read some other issues similar to this but i m wondering why dragging files from file explorer into code works and this doesn t i m guessing an electron issue
1
373,975
11,053,593,672
IssuesEvent
2019-12-10 11:43:22
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
eksisozluk.com - see bug description
browser-fenix engine-gecko priority-important
<!-- @browser: Firefox Mobile 72.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:72.0) Gecko/72.0 Firefox/72.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://eksisozluk.com/basliklar/m/populer **Browser / Version**: Firefox Mobile 72.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: No **Problem type**: Something else **Description**: I need click twice to open links **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
eksisozluk.com - see bug description - <!-- @browser: Firefox Mobile 72.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:72.0) Gecko/72.0 Firefox/72.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://eksisozluk.com/basliklar/m/populer **Browser / Version**: Firefox Mobile 72.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: No **Problem type**: Something else **Description**: I need click twice to open links **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
eksisozluk com see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description i need click twice to open links steps to reproduce browser configuration none from with ❤️
0
104,632
8,990,602,200
IssuesEvent
2019-02-01 05:55:36
owncloud/user_ldap
https://api.github.com/repos/owncloud/user_ldap
closed
Add api and webUI acceptance tests for modifying ldap user's attributes
QA-team dev:acceptance-tests
check what can/cannot happen with respect to modifying attributes of LDAP users (clues in comment https://github.com/owncloud/core/pull/32731#issuecomment-428204236 ) - Similar Issue for occ - #292 - [x] provisioning API #346 - [x] using the webUI (might be more difficult to test - for core master the user_management app would need to be installed/enabled, for core stable10 the user_management app should already be there) #360
1.0
Add api and webUI acceptance tests for modifying ldap user's attributes - check what can/cannot happen with respect to modifying attributes of LDAP users (clues in comment https://github.com/owncloud/core/pull/32731#issuecomment-428204236 ) - Similar Issue for occ - #292 - [x] provisioning API #346 - [x] using the webUI (might be more difficult to test - for core master the user_management app would need to be installed/enabled, for core stable10 the user_management app should already be there) #360
test
add api and webui acceptance tests for modifying ldap user s attributes check what can cannot happen with respect to modifying attributes of ldap users clues in comment similar issue for occ provisioning api using the webui might be more difficult to test for core master the user management app would need to be installed enabled for core the user management app should already be there
1
336,908
30,227,266,371
IssuesEvent
2023-07-06 02:19:53
harvester/harvester
https://api.github.com/repos/harvester/harvester
closed
[ENHANCEMENT] maintain the packages on OBS
kind/enhancement priority/2 area/os not-require/test-plan backport-needed/1.1.3 required-for-rc/v1.2.0
**Is your enhancement related to a problem? Please describe.** We have some packages installed from luet, but most of them need to be updated. We should maintain our package list on our OBS. **Describe the solution you'd like** Maintain packages ourselves on OBS. **Describe alternatives you've considered** **Additional context** For the short term, we can download the newer version directly.
1.0
[ENHANCEMENT] maintain the packages on OBS - **Is your enhancement related to a problem? Please describe.** We have some packages installed from luet, but most of them need to be updated. We should maintain our package list on our OBS. **Describe the solution you'd like** Maintain packages ourselves on OBS. **Describe alternatives you've considered** **Additional context** For the short term, we can download the newer version directly.
test
maintain the packages on obs is your enhancement related to a problem please describe we have some packages installed from luet but most of them need to be updated we should maintain our package list on our obs describe the solution you d like maintain packages ourselves on obs describe alternatives you ve considered additional context for the short term we can download the newer version directly
1
329,983
28,321,483,838
IssuesEvent
2023-04-11 01:49:41
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
ccl/changefeedccl: TestChangefeedBasics failed
C-test-failure O-robot branch-release-23.1
ccl/changefeedccl.TestChangefeedBasics [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9535855?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9535855?buildTab=artifacts#/) on release-23.1 @ [02cbcb36015ab1191192198b3892dea4b719d386](https://github.com/cockroachdb/cockroach/commits/02cbcb36015ab1191192198b3892dea4b719d386): ``` === RUN TestChangefeedBasics test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/51fe861426dc77645f2622be784d4d3c/logTestChangefeedBasics2829921528 test_log_scope.go:79: use -show-logs to present logs inline === CONT TestChangefeedBasics changefeed_test.go:265: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/51fe861426dc77645f2622be784d4d3c/logTestChangefeedBasics2829921528 --- FAIL: TestChangefeedBasics (23.89s) === RUN TestChangefeedBasics/pubsub helpers_test.go:817: making server as secondary tenant helpers_test.go:895: making pubsub feed factory changefeed_test.go:229: failed to start feed for job 0: pq: opening client: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information. --- FAIL: TestChangefeedBasics/pubsub (5.63s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestChangefeedBasics.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-26841
1.0
ccl/changefeedccl: TestChangefeedBasics failed - ccl/changefeedccl.TestChangefeedBasics [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9535855?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9535855?buildTab=artifacts#/) on release-23.1 @ [02cbcb36015ab1191192198b3892dea4b719d386](https://github.com/cockroachdb/cockroach/commits/02cbcb36015ab1191192198b3892dea4b719d386): ``` === RUN TestChangefeedBasics test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/51fe861426dc77645f2622be784d4d3c/logTestChangefeedBasics2829921528 test_log_scope.go:79: use -show-logs to present logs inline === CONT TestChangefeedBasics changefeed_test.go:265: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/51fe861426dc77645f2622be784d4d3c/logTestChangefeedBasics2829921528 --- FAIL: TestChangefeedBasics (23.89s) === RUN TestChangefeedBasics/pubsub helpers_test.go:817: making server as secondary tenant helpers_test.go:895: making pubsub feed factory changefeed_test.go:229: failed to start feed for job 0: pq: opening client: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information. --- FAIL: TestChangefeedBasics/pubsub (5.63s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestChangefeedBasics.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-26841
test
ccl changefeedccl testchangefeedbasics failed ccl changefeedccl testchangefeedbasics with on release run testchangefeedbasics test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline cont testchangefeedbasics changefeed test go test log scope end test logs left over in artifacts tmp tmp fail testchangefeedbasics run testchangefeedbasics pubsub helpers test go making server as secondary tenant helpers test go making pubsub feed factory changefeed test go failed to start feed for job pq opening client google could not find default credentials see for more information fail testchangefeedbasics pubsub help see also cc cockroachdb cdc jira issue crdb
1
169,286
13,132,489,801
IssuesEvent
2020-08-06 19:00:41
lum-ai/odinson
https://api.github.com/repos/lum-ai/odinson
opened
Tests for REST API
area: REST API testing
We need some foundational tests for the REST API. This will require ... - A test index to hit and inspect responses to ensure they conform to the [OpenAPI specification](https://github.com/lum-ai/odinson/blob/3531b5676297d7cf959fb78b127043181c44ede2/backend/public/schema/odinson.yaml) - This will require injecting the config at test time. - Establishing a testing framework for the `backend` subproject ([likely using `org.scalatestplus.play`](https://www.playframework.com/documentation/2.8.x/ScalaTestingWithScalaTest)). _Minimally_ (or to start anyway), we need to test the following endpoints: - [`/api/execute/pattern`](https://github.com/lum-ai/odinson/blob/3531b5676297d7cf959fb78b127043181c44ede2/backend/conf/routes#L15) - [`/api/execute/grammar`](https://github.com/lum-ai/odinson/blob/3531b5676297d7cf959fb78b127043181c44ede2/backend/conf/routes#L17)
1.0
Tests for REST API - We need some foundational tests for the REST API. This will require ... - A test index to hit and inspect responses to ensure they conform to the [OpenAPI specification](https://github.com/lum-ai/odinson/blob/3531b5676297d7cf959fb78b127043181c44ede2/backend/public/schema/odinson.yaml) - This will require injecting the config at test time. - Establishing a testing framework for the `backend` subproject ([likely using `org.scalatestplus.play`](https://www.playframework.com/documentation/2.8.x/ScalaTestingWithScalaTest)). _Minimally_ (or to start anyway), we need to test the following endpoints: - [`/api/execute/pattern`](https://github.com/lum-ai/odinson/blob/3531b5676297d7cf959fb78b127043181c44ede2/backend/conf/routes#L15) - [`/api/execute/grammar`](https://github.com/lum-ai/odinson/blob/3531b5676297d7cf959fb78b127043181c44ede2/backend/conf/routes#L17)
test
tests for rest api we need some foundational tests for the rest api this will require a test index to hit and inspect responses to ensure they conform to the this will require injecting the config at test time establishing a testing framework for the backend subproject minimally or to start anyway we need to test the following endpoints
1
200,672
15,128,939,098
IssuesEvent
2021-02-10 00:18:15
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Revisting test states
testing
Currently the model of test states works great to show the current state of thing, but browsing and navigating test results gets more challenging. The responsibility to persist the most recent test result is on the extension. For example, if I run tests in a document, close the document, and reopen it (creating a new test subscription) the extension needs to re-hydrate the state of tests at that time. My workaround for this was to expose the test results to the extension, but this is a poor solution. https://github.com/microsoft/vscode/blob/b7e87c1bf2780f6b9338721369e422dcb398224d/src/vs/vscode.proposed.d.ts#L2084-L2101 So, it appears the idea of tests containing their state is flawed. Instead, I think it is reasonable to say that test states only change during a test run. An approach to this is to modify the signature to `runTests(run: TestRun, cancellationToken)`, where `TestRun` is: ```ts /** * Class given to `TestProvider.runTests` */ export class TestRun<T extends TestItem = TestItem> { // existing properties of TestRunOptions tests: T[]; debug: boolean; /** * Updates the state of the test in the run. By default, all tests involved * in the run will have a "queued" state until they are updated by this method. */ update(test: T, state: TestState): void; } ``` ...and then there would be no `state` on TestItems themselves. This is closer to the API that @hbenl has in the test explorer API. TestItems are purely metadata about the location and runnability of tests, and VS Code handles the display and persistence of test states. Something I like about this API is that it opens the door to streaming output (as @matepak requested in https://github.com/microsoft/vscode/issues/107467#issuecomment-766085288) via some `appendOutput(test: T, contents: string | TestMessage)` method, as well as extra information like coverage later on. It also removes the awkward event emitter loopback I mentioned in https://github.com/microsoft/vscode/issues/107467#issuecomment-730706370 One use case that the current API provides that this change does not is the ability to open test results without previously running tests. For that we can introduce some a new optional `provideTestResults()` which can return a result summary.
1.0
Revisting test states - Currently the model of test states works great to show the current state of thing, but browsing and navigating test results gets more challenging. The responsibility to persist the most recent test result is on the extension. For example, if I run tests in a document, close the document, and reopen it (creating a new test subscription) the extension needs to re-hydrate the state of tests at that time. My workaround for this was to expose the test results to the extension, but this is a poor solution. https://github.com/microsoft/vscode/blob/b7e87c1bf2780f6b9338721369e422dcb398224d/src/vs/vscode.proposed.d.ts#L2084-L2101 So, it appears the idea of tests containing their state is flawed. Instead, I think it is reasonable to say that test states only change during a test run. An approach to this is to modify the signature to `runTests(run: TestRun, cancellationToken)`, where `TestRun` is: ```ts /** * Class given to `TestProvider.runTests` */ export class TestRun<T extends TestItem = TestItem> { // existing properties of TestRunOptions tests: T[]; debug: boolean; /** * Updates the state of the test in the run. By default, all tests involved * in the run will have a "queued" state until they are updated by this method. */ update(test: T, state: TestState): void; } ``` ...and then there would be no `state` on TestItems themselves. This is closer to the API that @hbenl has in the test explorer API. TestItems are purely metadata about the location and runnability of tests, and VS Code handles the display and persistence of test states. Something I like about this API is that it opens the door to streaming output (as @matepak requested in https://github.com/microsoft/vscode/issues/107467#issuecomment-766085288) via some `appendOutput(test: T, contents: string | TestMessage)` method, as well as extra information like coverage later on. It also removes the awkward event emitter loopback I mentioned in https://github.com/microsoft/vscode/issues/107467#issuecomment-730706370 One use case that the current API provides that this change does not is the ability to open test results without previously running tests. For that we can introduce some a new optional `provideTestResults()` which can return a result summary.
test
revisting test states currently the model of test states works great to show the current state of thing but browsing and navigating test results gets more challenging the responsibility to persist the most recent test result is on the extension for example if i run tests in a document close the document and reopen it creating a new test subscription the extension needs to re hydrate the state of tests at that time my workaround for this was to expose the test results to the extension but this is a poor solution so it appears the idea of tests containing their state is flawed instead i think it is reasonable to say that test states only change during a test run an approach to this is to modify the signature to runtests run testrun cancellationtoken where testrun is ts class given to testprovider runtests export class testrun existing properties of testrunoptions tests t debug boolean updates the state of the test in the run by default all tests involved in the run will have a queued state until they are updated by this method update test t state teststate void and then there would be no state on testitems themselves this is closer to the api that hbenl has in the test explorer api testitems are purely metadata about the location and runnability of tests and vs code handles the display and persistence of test states something i like about this api is that it opens the door to streaming output as matepak requested in via some appendoutput test t contents string testmessage method as well as extra information like coverage later on it also removes the awkward event emitter loopback i mentioned in one use case that the current api provides that this change does not is the ability to open test results without previously running tests for that we can introduce some a new optional providetestresults which can return a result summary
1
132,519
10,757,714,357
IssuesEvent
2019-10-31 13:47:13
kcigeospatial/Fred_Co_Land-Management
https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management
closed
Building Permits - Fire Permit - Inspection Assignment
Ready for Test Env. Retest
The FIRE-FWFLW inspection type is not auto-assigning. ![image](https://user-images.githubusercontent.com/31770504/67878135-2f69f600-fb11-11e9-90a8-c69bcc0682ae.png)
2.0
Building Permits - Fire Permit - Inspection Assignment - The FIRE-FWFLW inspection type is not auto-assigning. ![image](https://user-images.githubusercontent.com/31770504/67878135-2f69f600-fb11-11e9-90a8-c69bcc0682ae.png)
test
building permits fire permit inspection assignment the fire fwflw inspection type is not auto assigning
1
204,017
15,397,843,577
IssuesEvent
2021-03-03 22:51:09
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
Test failure: org.elasticsearch.grok.GrokTests.testExponentialExpressions
:Core/Features/Ingest >test-failure Team:Core/Features
Given the rarity of this failure I am not muting. However, if this increases in occurrences this test should be muted. **Build scan**: https://gradle-enterprise.elastic.co/s/iak3abjnc4huu **Repro line**: ``` REPRODUCE WITH: ./gradlew ':libs:elasticsearch-grok:test' --tests "org.elasticsearch.grok.GrokTests.testExponentialExpressions" \ -Dtests.seed=2012EF738111106E \ -Dtests.security.manager=true \ -Dtests.locale=es-VE \ -Dtests.timezone=Africa/Niamey \ -Druntime.java=11 ``` **Reproduces locally?**: No. Ran a few thousand iterations (without the extra -D arguments ) and they all passed. **Applicable branches**: This failure is on master, but history shows it is not specific to master. **Failure history**: https://gradle-enterprise.elastic.co/scans/tests?list.size=50&list.sortColumn=startTime&list.sortOrder=desc&search.buildToolType=gradle&search.buildToolType=maven&search.relativeStartTime=P28D&search.timeZoneId=America/Chicago&tests.container=org.elasticsearch.grok.GrokTests&tests.sortField=FAILED&tests.test=testExponentialExpressions&tests.unstableOnly=true This is the only one in past 28 day. However, searching my email archives i see ~ 15 failures in the past 1.5 years across all branches. So this is pretty rare, but not also not likely caused by cosmic rays. I suspect that there is a subtle race condition in the test itself. **Failure excerpt**: ``` org.elasticsearch.grok.GrokTests > classMethod FAILED java.lang.Exception: Suite timeout exceeded (>= 1200000 msec). at __randomizedtesting.SeedInfo.seed([2012EF738111106E]:0) ``` and ``` 2> WARNING: Suite execution timed out: org.elasticsearch.grok.GrokTests ``` I believe that for this particular test, it can occur if the Watch dog thread does not kill the Grok match as it should.
1.0
Test failure: org.elasticsearch.grok.GrokTests.testExponentialExpressions - Given the rarity of this failure I am not muting. However, if this increases in occurrences this test should be muted. **Build scan**: https://gradle-enterprise.elastic.co/s/iak3abjnc4huu **Repro line**: ``` REPRODUCE WITH: ./gradlew ':libs:elasticsearch-grok:test' --tests "org.elasticsearch.grok.GrokTests.testExponentialExpressions" \ -Dtests.seed=2012EF738111106E \ -Dtests.security.manager=true \ -Dtests.locale=es-VE \ -Dtests.timezone=Africa/Niamey \ -Druntime.java=11 ``` **Reproduces locally?**: No. Ran a few thousand iterations (without the extra -D arguments ) and they all passed. **Applicable branches**: This failure is on master, but history shows it is not specific to master. **Failure history**: https://gradle-enterprise.elastic.co/scans/tests?list.size=50&list.sortColumn=startTime&list.sortOrder=desc&search.buildToolType=gradle&search.buildToolType=maven&search.relativeStartTime=P28D&search.timeZoneId=America/Chicago&tests.container=org.elasticsearch.grok.GrokTests&tests.sortField=FAILED&tests.test=testExponentialExpressions&tests.unstableOnly=true This is the only one in past 28 day. However, searching my email archives i see ~ 15 failures in the past 1.5 years across all branches. So this is pretty rare, but not also not likely caused by cosmic rays. I suspect that there is a subtle race condition in the test itself. **Failure excerpt**: ``` org.elasticsearch.grok.GrokTests > classMethod FAILED java.lang.Exception: Suite timeout exceeded (>= 1200000 msec). at __randomizedtesting.SeedInfo.seed([2012EF738111106E]:0) ``` and ``` 2> WARNING: Suite execution timed out: org.elasticsearch.grok.GrokTests ``` I believe that for this particular test, it can occur if the Watch dog thread does not kill the Grok match as it should.
test
test failure org elasticsearch grok groktests testexponentialexpressions given the rarity of this failure i am not muting however if this increases in occurrences this test should be muted build scan repro line reproduce with gradlew libs elasticsearch grok test tests org elasticsearch grok groktests testexponentialexpressions dtests seed dtests security manager true dtests locale es ve dtests timezone africa niamey druntime java reproduces locally no ran a few thousand iterations without the extra d arguments and they all passed applicable branches this failure is on master but history shows it is not specific to master failure history this is the only one in past day however searching my email archives i see failures in the past years across all branches so this is pretty rare but not also not likely caused by cosmic rays i suspect that there is a subtle race condition in the test itself failure excerpt org elasticsearch grok groktests classmethod failed java lang exception suite timeout exceeded msec at randomizedtesting seedinfo seed and warning suite execution timed out org elasticsearch grok groktests i believe that for this particular test it can occur if the watch dog thread does not kill the grok match as it should
1
10,915
8,214,534,495
IssuesEvent
2018-09-04 23:56:11
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
Lookup Realms / Authorizing Realms
:Security/Authorization >feature
# Support for lookup/authorizing realms There are times where authentication is performed by identity server but for user details, one might want to use an external authorization server fetching details like group information, metadata etc. ## Use Case: - As ES Admin, I would want to authenticate with one identity server/mechanism but queries that user on an external authorization server providing group information or other metadata. For eg. `PkiRealm` authenticates users based on the client presented certificate. In turn, the identified username could be used to query group information from an external LDAP server. Another eg. would be for `KerberosRealm` where after user authentication we might want to use external LDAP server to fetch group information. We have a chain of authentication realms which get invoked during the authentication process based on the request attributes. Current realms support `lookupUser()` for that particular realm used during `runAs` user impersonation. We can invoke this lookup user from the realms who want to depend on other realms for user information.
True
Lookup Realms / Authorizing Realms - # Support for lookup/authorizing realms There are times where authentication is performed by identity server but for user details, one might want to use an external authorization server fetching details like group information, metadata etc. ## Use Case: - As ES Admin, I would want to authenticate with one identity server/mechanism but queries that user on an external authorization server providing group information or other metadata. For eg. `PkiRealm` authenticates users based on the client presented certificate. In turn, the identified username could be used to query group information from an external LDAP server. Another eg. would be for `KerberosRealm` where after user authentication we might want to use external LDAP server to fetch group information. We have a chain of authentication realms which get invoked during the authentication process based on the request attributes. Current realms support `lookupUser()` for that particular realm used during `runAs` user impersonation. We can invoke this lookup user from the realms who want to depend on other realms for user information.
non_test
lookup realms authorizing realms support for lookup authorizing realms there are times where authentication is performed by identity server but for user details one might want to use an external authorization server fetching details like group information metadata etc use case as es admin i would want to authenticate with one identity server mechanism but queries that user on an external authorization server providing group information or other metadata for eg pkirealm authenticates users based on the client presented certificate in turn the identified username could be used to query group information from an external ldap server another eg would be for kerberosrealm where after user authentication we might want to use external ldap server to fetch group information we have a chain of authentication realms which get invoked during the authentication process based on the request attributes current realms support lookupuser for that particular realm used during runas user impersonation we can invoke this lookup user from the realms who want to depend on other realms for user information
0
114,907
9,764,379,216
IssuesEvent
2019-06-05 15:39:07
aces/Loris
https://api.github.com/repos/aces/Loris
closed
[Dicom Archive] Click on 'View Details' not working for one record
21.0.0 Testing Bug Fix PR sent
TestPlan Item 5. `Click on 'View Details' under Metadata to view the Details of MRI Scan [Manual Testing]` Problem: Record with `MTL022_300022_V1`, click on `View Details` leads to a blank page. Other records seems fine.
1.0
[Dicom Archive] Click on 'View Details' not working for one record - TestPlan Item 5. `Click on 'View Details' under Metadata to view the Details of MRI Scan [Manual Testing]` Problem: Record with `MTL022_300022_V1`, click on `View Details` leads to a blank page. Other records seems fine.
test
click on view details not working for one record testplan item click on view details under metadata to view the details of mri scan problem record with click on view details leads to a blank page other records seems fine
1
148,544
11,856,138,870
IssuesEvent
2020-03-25 06:41:20
appium/appium
https://api.github.com/repos/appium/appium
closed
Unable to click on an element in bluetooth screen of iOS device
ThirdParty XCUITest
## The problem WDA was not able to perform press action in some iterations on "More Info" element for a paired device on Bluetooth Screen. ## Environment * Appium version (or git revision) that exhibits the issue: Latest * Last Appium version that did not exhibit the issue (if applicable): N/A * Desktop OS/version used to run Appium: Mac OS Catalina/Xcode 11.3.1 * Node.js version (unless using Appium.app|exe): N/A * Npm or Yarn package manager: N/A * Mobile platform/version under test: iOS/13.3.1 * Real device or emulator/simulator: Real Device (iPhone XR) ## Details Paired a device using bluetooth and tried to click on "More Info" button using WDA.It worked for some iterations. But in some cases, WDA was not able to press the button.However the response was always the same. Response: { "value" : null, "sessionId" : "B95282F5-8096-4882-92B4-073D918E4453" } ## Link to Appium logs Snippet of logs for finding & pressing on "More Info" element taken from Xcode : t = 66.57s Find: Descendants matching type Alert t = 66.57s Synthesize event t = 68.75s Find the Application 'com.apple.Preferences' t = 68.75s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 68.88s Get all elements bound by accessibility element for: Elements matching predicate '"More Info" IN identifiers' t = 68.88s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 68.96s Find: Descendants matching type Any t = 68.96s Find: Elements matching predicate '"More Info" IN identifiers' t = 68.97s Find the "More Info" Button t = 68.98s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 69.05s Find: Descendants matching type Any t = 69.06s Find: Elements matching predicate '"More Info" IN identifiers' t = 69.06s Find: Identity Binding t = 71.29s Find the "More Info" Any t = 71.29s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 71.42s Find: Descendants matching type Any t = 71.42s Find: Elements matching predicate '"More Info" IN identifiers' t = 71.43s Find: Identity Binding t = 72.44s Find the "More Info" Any (retry 1) t = 72.45s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 72.56s Find: Descendants matching type Any t = 72.56s Find: Elements matching predicate '"More Info" IN identifiers' t = 72.57s Find: Identity Binding t = 73.58s Find the "More Info" Any (retry 2) t = 73.59s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 73.69s Find: Descendants matching type Any t = 73.69s Find: Elements matching predicate '"More Info" IN identifiers' t = 73.70s Find: Identity Binding t = 73.70s Collecting extra data to assist test failure triage t = 73.70s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 73.83s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 73.94s Find: Descendants matching type Any t = 73.95s Find: Elements matching predicate '"More Info" IN identifiers' t = 73.96s Find: Identity Binding 2020-03-19 11:19:04.625081+0530 WebDriverAgentRunner-Runner[312:13820] Enqueue Failure: Failed to get matching snapshot: No matches found for Identity Binding from input {( label: 'More Info' )} /Users/abilash/Documents/Projects/Appium/WebDriverAgentRunner/UITestingUITests.m 38 1 t = 73.99s Find the "More Info" Button t = 73.99s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 74.06s Find: Descendants matching type Any t = 74.06s Find: Elements matching predicate '"More Info" IN identifiers' t = 74.06s Find: Identity Binding t = 75.07s Find the "More Info" Button (retry 1) t = 75.08s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 75.21s Find: Descendants matching type Any t = 75.22s Find: Elements matching predicate '"More Info" IN identifiers' t = 75.22s Find: Identity Binding t = 76.23s Find the "More Info" Button (retry 2) t = 76.24s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 76.38s Find: Descendants matching type Any t = 76.38s Find: Elements matching predicate '"More Info" IN identifiers' t = 76.39s Find: Identity Binding t = 76.40s Collecting extra data to assist test failure triage t = 76.40s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 76.53s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 76.65s Find: Descendants matching type Any t = 76.66s Find: Elements matching predicate '"More Info" IN identifiers' t = 76.66s Find: Identity Binding 2020-03-19 11:19:07.297958+0530 WebDriverAgentRunner-Runner[312:13820] Enqueue Failure: Failed to get matching snapshot: No matches found for Identity Binding from input {( label: 'More Info' )} /Users/abilash/Documents/Projects/Appium/WebDriverAgentRunner/UITestingUITests.m 38 1 t = 76.77s Find the "More Info" Button t = 76.78s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 76.92s Find: Descendants matching type Any t = 76.92s Find: Elements matching predicate '"More Info" IN identifiers' t = 76.93s Find: Identity Binding t = 77.94s Find the "More Info" Button (retry 1) t = 77.94s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 78.08s Find: Descendants matching type Any t = 78.09s Find: Elements matching predicate '"More Info" IN identifiers' t = 78.09s Find: Identity Binding t = 79.10s Find the "More Info" Button (retry 2) t = 79.11s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 79.25s Find: Descendants matching type Any t = 79.26s Find: Elements matching predicate '"More Info" IN identifiers' t = 79.26s Find: Identity Binding t = 79.27s Collecting extra data to assist test failure triage t = 79.27s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 79.39s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 79.52s Find: Descendants matching type Any t = 79.53s Find: Elements matching predicate '"More Info" IN identifiers' t = 79.53s Find: Identity Binding 2020-03-19 11:19:10.166765+0530 WebDriverAgentRunner-Runner[312:13820] Enqueue Failure: Failed to get matching snapshot: No matches found for Identity Binding from input {( label: 'More Info' )} /Users/abilash/Documents/Projects/Appium/WebDriverAgentRunner/UITestingUITests.m 38 1 t = 79.53s Tap "More Info" Button t = 79.53s Find the "More Info" Button t = 79.54s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 79.62s Find: Descendants matching type Any t = 79.62s Find: Elements matching predicate '"More Info" IN identifiers' t = 79.63s Find: Identity Binding t = 80.63s Find the "More Info" Button (retry 1) t = 80.64s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 80.78s Find: Descendants matching type Any t = 80.78s Find: Elements matching predicate '"More Info" IN identifiers' t = 80.79s Find: Identity Binding t = 81.79s Find the "More Info" Button (retry 2) t = 81.80s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 81.94s Find: Descendants matching type Any t = 81.95s Find: Elements matching predicate '"More Info" IN identifiers' t = 81.96s Find: Identity Binding t = 81.96s Collecting extra data to assist test failure triage t = 81.96s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 82.08s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 82.20s Find: Descendants matching type Any t = 82.21s Find: Elements matching predicate '"More Info" IN identifiers' t = 82.22s Find: Identity Binding 2020-03-19 11:19:12.851526+0530 WebDriverAgentRunner-Runner[312:13820] Enqueue Failure: Failed to get matching snapshot: No matches found for Identity Binding from input {( label: 'More Info' )} ## Code To Reproduce Issue [ Good To Have ] Find & Press on "More Info" Element on paired device in bluetooth Screen.
1.0
Unable to click on an element in bluetooth screen of iOS device - ## The problem WDA was not able to perform press action in some iterations on "More Info" element for a paired device on Bluetooth Screen. ## Environment * Appium version (or git revision) that exhibits the issue: Latest * Last Appium version that did not exhibit the issue (if applicable): N/A * Desktop OS/version used to run Appium: Mac OS Catalina/Xcode 11.3.1 * Node.js version (unless using Appium.app|exe): N/A * Npm or Yarn package manager: N/A * Mobile platform/version under test: iOS/13.3.1 * Real device or emulator/simulator: Real Device (iPhone XR) ## Details Paired a device using bluetooth and tried to click on "More Info" button using WDA.It worked for some iterations. But in some cases, WDA was not able to press the button.However the response was always the same. Response: { "value" : null, "sessionId" : "B95282F5-8096-4882-92B4-073D918E4453" } ## Link to Appium logs Snippet of logs for finding & pressing on "More Info" element taken from Xcode : t = 66.57s Find: Descendants matching type Alert t = 66.57s Synthesize event t = 68.75s Find the Application 'com.apple.Preferences' t = 68.75s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 68.88s Get all elements bound by accessibility element for: Elements matching predicate '"More Info" IN identifiers' t = 68.88s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 68.96s Find: Descendants matching type Any t = 68.96s Find: Elements matching predicate '"More Info" IN identifiers' t = 68.97s Find the "More Info" Button t = 68.98s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 69.05s Find: Descendants matching type Any t = 69.06s Find: Elements matching predicate '"More Info" IN identifiers' t = 69.06s Find: Identity Binding t = 71.29s Find the "More Info" Any t = 71.29s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 71.42s Find: Descendants matching type Any t = 71.42s Find: Elements matching predicate '"More Info" IN identifiers' t = 71.43s Find: Identity Binding t = 72.44s Find the "More Info" Any (retry 1) t = 72.45s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 72.56s Find: Descendants matching type Any t = 72.56s Find: Elements matching predicate '"More Info" IN identifiers' t = 72.57s Find: Identity Binding t = 73.58s Find the "More Info" Any (retry 2) t = 73.59s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 73.69s Find: Descendants matching type Any t = 73.69s Find: Elements matching predicate '"More Info" IN identifiers' t = 73.70s Find: Identity Binding t = 73.70s Collecting extra data to assist test failure triage t = 73.70s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 73.83s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 73.94s Find: Descendants matching type Any t = 73.95s Find: Elements matching predicate '"More Info" IN identifiers' t = 73.96s Find: Identity Binding 2020-03-19 11:19:04.625081+0530 WebDriverAgentRunner-Runner[312:13820] Enqueue Failure: Failed to get matching snapshot: No matches found for Identity Binding from input {( label: 'More Info' )} /Users/abilash/Documents/Projects/Appium/WebDriverAgentRunner/UITestingUITests.m 38 1 t = 73.99s Find the "More Info" Button t = 73.99s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 74.06s Find: Descendants matching type Any t = 74.06s Find: Elements matching predicate '"More Info" IN identifiers' t = 74.06s Find: Identity Binding t = 75.07s Find the "More Info" Button (retry 1) t = 75.08s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 75.21s Find: Descendants matching type Any t = 75.22s Find: Elements matching predicate '"More Info" IN identifiers' t = 75.22s Find: Identity Binding t = 76.23s Find the "More Info" Button (retry 2) t = 76.24s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 76.38s Find: Descendants matching type Any t = 76.38s Find: Elements matching predicate '"More Info" IN identifiers' t = 76.39s Find: Identity Binding t = 76.40s Collecting extra data to assist test failure triage t = 76.40s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 76.53s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 76.65s Find: Descendants matching type Any t = 76.66s Find: Elements matching predicate '"More Info" IN identifiers' t = 76.66s Find: Identity Binding 2020-03-19 11:19:07.297958+0530 WebDriverAgentRunner-Runner[312:13820] Enqueue Failure: Failed to get matching snapshot: No matches found for Identity Binding from input {( label: 'More Info' )} /Users/abilash/Documents/Projects/Appium/WebDriverAgentRunner/UITestingUITests.m 38 1 t = 76.77s Find the "More Info" Button t = 76.78s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 76.92s Find: Descendants matching type Any t = 76.92s Find: Elements matching predicate '"More Info" IN identifiers' t = 76.93s Find: Identity Binding t = 77.94s Find the "More Info" Button (retry 1) t = 77.94s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 78.08s Find: Descendants matching type Any t = 78.09s Find: Elements matching predicate '"More Info" IN identifiers' t = 78.09s Find: Identity Binding t = 79.10s Find the "More Info" Button (retry 2) t = 79.11s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 79.25s Find: Descendants matching type Any t = 79.26s Find: Elements matching predicate '"More Info" IN identifiers' t = 79.26s Find: Identity Binding t = 79.27s Collecting extra data to assist test failure triage t = 79.27s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 79.39s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 79.52s Find: Descendants matching type Any t = 79.53s Find: Elements matching predicate '"More Info" IN identifiers' t = 79.53s Find: Identity Binding 2020-03-19 11:19:10.166765+0530 WebDriverAgentRunner-Runner[312:13820] Enqueue Failure: Failed to get matching snapshot: No matches found for Identity Binding from input {( label: 'More Info' )} /Users/abilash/Documents/Projects/Appium/WebDriverAgentRunner/UITestingUITests.m 38 1 t = 79.53s Tap "More Info" Button t = 79.53s Find the "More Info" Button t = 79.54s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 79.62s Find: Descendants matching type Any t = 79.62s Find: Elements matching predicate '"More Info" IN identifiers' t = 79.63s Find: Identity Binding t = 80.63s Find the "More Info" Button (retry 1) t = 80.64s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 80.78s Find: Descendants matching type Any t = 80.78s Find: Elements matching predicate '"More Info" IN identifiers' t = 80.79s Find: Identity Binding t = 81.79s Find the "More Info" Button (retry 2) t = 81.80s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 81.94s Find: Descendants matching type Any t = 81.95s Find: Elements matching predicate '"More Info" IN identifiers' t = 81.96s Find: Identity Binding t = 81.96s Collecting extra data to assist test failure triage t = 81.96s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 82.08s Requesting snapshot of accessibility hierarchy for app with pid 321 t = 82.20s Find: Descendants matching type Any t = 82.21s Find: Elements matching predicate '"More Info" IN identifiers' t = 82.22s Find: Identity Binding 2020-03-19 11:19:12.851526+0530 WebDriverAgentRunner-Runner[312:13820] Enqueue Failure: Failed to get matching snapshot: No matches found for Identity Binding from input {( label: 'More Info' )} ## Code To Reproduce Issue [ Good To Have ] Find & Press on "More Info" Element on paired device in bluetooth Screen.
test
unable to click on an element in bluetooth screen of ios device the problem wda was not able to perform press action in some iterations on more info element for a paired device on bluetooth screen environment appium version or git revision that exhibits the issue latest last appium version that did not exhibit the issue if applicable n a desktop os version used to run appium mac os catalina xcode node js version unless using appium app exe n a npm or yarn package manager n a mobile platform version under test ios real device or emulator simulator real device iphone xr details paired a device using bluetooth and tried to click on more info button using wda it worked for some iterations but in some cases wda was not able to press the button however the response was always the same response value null sessionid link to appium logs snippet of logs for finding pressing on more info element taken from xcode t find descendants matching type alert t synthesize event t find the application com apple preferences t requesting snapshot of accessibility hierarchy for app with pid t get all elements bound by accessibility element for elements matching predicate more info in identifiers t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find the more info button t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info any t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info any retry t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info any retry t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t collecting extra data to assist test failure triage t requesting snapshot of accessibility hierarchy for app with pid t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding webdriveragentrunner runner enqueue failure failed to get matching snapshot no matches found for identity binding from input label more info users abilash documents projects appium webdriveragentrunner uitestinguitests m t find the more info button t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info button retry t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info button retry t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t collecting extra data to assist test failure triage t requesting snapshot of accessibility hierarchy for app with pid t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding webdriveragentrunner runner enqueue failure failed to get matching snapshot no matches found for identity binding from input label more info users abilash documents projects appium webdriveragentrunner uitestinguitests m t find the more info button t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info button retry t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info button retry t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t collecting extra data to assist test failure triage t requesting snapshot of accessibility hierarchy for app with pid t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding webdriveragentrunner runner enqueue failure failed to get matching snapshot no matches found for identity binding from input label more info users abilash documents projects appium webdriveragentrunner uitestinguitests m t tap more info button t find the more info button t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info button retry t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t find the more info button retry t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding t collecting extra data to assist test failure triage t requesting snapshot of accessibility hierarchy for app with pid t requesting snapshot of accessibility hierarchy for app with pid t find descendants matching type any t find elements matching predicate more info in identifiers t find identity binding webdriveragentrunner runner enqueue failure failed to get matching snapshot no matches found for identity binding from input label more info code to reproduce issue find press on more info element on paired device in bluetooth screen
1
52,176
22,051,199,916
IssuesEvent
2022-05-30 08:49:18
polarismesh/polaris
https://api.github.com/repos/polarismesh/polaris
closed
【Function Design】支持服务实例的增量推送
enhancement apisever need discuss service xds good advanced issue
### Discussed in https://github.com/polarismesh/polaris/discussions/348 <div type='discussions-op-text'> <sup>Originally posted by **chuntaojun** April 24, 2022</sup> **What is the feature you want to add?** - 支持 xds 的增量推送协议 - 为 polaris-sdk 也提供支持增量推送的 api **Why do you want to add this feature?** - 降低每次推送的数据包大小,降低server自身序列化所带来的CPU开销 - 降低整个集群的网络负担 **How to implement this feature?** **Additional context** Add any other context or screenshots about the feature request here. </div>
1.0
【Function Design】支持服务实例的增量推送 - ### Discussed in https://github.com/polarismesh/polaris/discussions/348 <div type='discussions-op-text'> <sup>Originally posted by **chuntaojun** April 24, 2022</sup> **What is the feature you want to add?** - 支持 xds 的增量推送协议 - 为 polaris-sdk 也提供支持增量推送的 api **Why do you want to add this feature?** - 降低每次推送的数据包大小,降低server自身序列化所带来的CPU开销 - 降低整个集群的网络负担 **How to implement this feature?** **Additional context** Add any other context or screenshots about the feature request here. </div>
non_test
【function design】支持服务实例的增量推送 discussed in originally posted by chuntaojun april what is the feature you want to add 支持 xds 的增量推送协议 为 polaris sdk 也提供支持增量推送的 api why do you want to add this feature 降低每次推送的数据包大小,降低server自身序列化所带来的cpu开销 降低整个集群的网络负担 how to implement this feature additional context add any other context or screenshots about the feature request here
0
55,753
6,490,721,079
IssuesEvent
2017-08-21 08:09:43
Interaktivtechnology/Raimon-Web
https://api.github.com/repos/Interaktivtechnology/Raimon-Web
closed
Publish This Building --> Cannot click cancel
bug Need Testing
@benjaminl83 @purnomoeko Unable to click cancel for Publish this building ![image](https://user-images.githubusercontent.com/30112738/29508262-0c18f1b0-867e-11e7-8d30-e05966910554.png)
1.0
Publish This Building --> Cannot click cancel - @benjaminl83 @purnomoeko Unable to click cancel for Publish this building ![image](https://user-images.githubusercontent.com/30112738/29508262-0c18f1b0-867e-11e7-8d30-e05966910554.png)
test
publish this building cannot click cancel purnomoeko unable to click cancel for publish this building
1
246,946
26,660,536,942
IssuesEvent
2023-01-25 20:40:23
CMSgov/cms-carts-seds
https://api.github.com/repos/CMSgov/cms-carts-seds
closed
SHF - cms-carts-seds - main - LOW - APIGateway.3 API Gateway REST API stages should have AWS X-Ray tracing enabled
security-hub main
************************************************************** __This issue was generated from Security Hub data and is managed through automation.__ Please do not edit the title or body of this issue, or remove the security-hub tag. All other edits/comments are welcome. Finding Id: arn:aws:securityhub:us-east-1:519095364708:subscription/aws-foundational-security-best-practices/v/1.0.0/APIGateway.3/finding/20fcfdf5-71e4-431b-94af-eb131efff9e5 ************************************************************** ## Type of Issue: - [x] Security Hub Finding ## Title: APIGateway.3 API Gateway REST API stages should have AWS X-Ray tracing enabled ## Id: arn:aws:securityhub:us-east-1:519095364708:subscription/aws-foundational-security-best-practices/v/1.0.0/APIGateway.3/finding/20fcfdf5-71e4-431b-94af-eb131efff9e5 (You may use this ID to lookup this finding's details in Security Hub) ## Description This control checks whether AWS X-Ray active tracing is enabled for your Amazon API Gateway REST API stages. ## Remediation https://docs.aws.amazon.com/console/securityhub/APIGateway.3/remediation ## AC: - The security hub finding is resolved or suppressed, indicated by a Workflow Status of Resolved or Suppressed.
True
SHF - cms-carts-seds - main - LOW - APIGateway.3 API Gateway REST API stages should have AWS X-Ray tracing enabled - ************************************************************** __This issue was generated from Security Hub data and is managed through automation.__ Please do not edit the title or body of this issue, or remove the security-hub tag. All other edits/comments are welcome. Finding Id: arn:aws:securityhub:us-east-1:519095364708:subscription/aws-foundational-security-best-practices/v/1.0.0/APIGateway.3/finding/20fcfdf5-71e4-431b-94af-eb131efff9e5 ************************************************************** ## Type of Issue: - [x] Security Hub Finding ## Title: APIGateway.3 API Gateway REST API stages should have AWS X-Ray tracing enabled ## Id: arn:aws:securityhub:us-east-1:519095364708:subscription/aws-foundational-security-best-practices/v/1.0.0/APIGateway.3/finding/20fcfdf5-71e4-431b-94af-eb131efff9e5 (You may use this ID to lookup this finding's details in Security Hub) ## Description This control checks whether AWS X-Ray active tracing is enabled for your Amazon API Gateway REST API stages. ## Remediation https://docs.aws.amazon.com/console/securityhub/APIGateway.3/remediation ## AC: - The security hub finding is resolved or suppressed, indicated by a Workflow Status of Resolved or Suppressed.
non_test
shf cms carts seds main low apigateway api gateway rest api stages should have aws x ray tracing enabled this issue was generated from security hub data and is managed through automation please do not edit the title or body of this issue or remove the security hub tag all other edits comments are welcome finding id arn aws securityhub us east subscription aws foundational security best practices v apigateway finding type of issue security hub finding title apigateway api gateway rest api stages should have aws x ray tracing enabled id arn aws securityhub us east subscription aws foundational security best practices v apigateway finding you may use this id to lookup this finding s details in security hub description this control checks whether aws x ray active tracing is enabled for your amazon api gateway rest api stages remediation ac the security hub finding is resolved or suppressed indicated by a workflow status of resolved or suppressed
0
562,350
16,657,827,569
IssuesEvent
2021-06-05 21:13:48
zorkind/HRP-Light
https://api.github.com/repos/zorkind/HRP-Light
closed
Battery Recharger flips back to close with batteries attached.
Vanilla bug cosmetic low priority
**Vanilla / HRP Light** Vanilla. **Client Version** 0.5.2 (Vanilla) **Describe the bug** Battery Recharger only open/close its attach points on proximity detection. So if you step away from recharger and a battery is attached, it will flip close. Also the Icon on the charger should go Orange only when it is actually charging a battery. **To Reproduce** Steps to reproduce the behavior: 1. Go to a battery recharger, it will flip open 2. Plug a battery 3. Step away. 4. See error **Expected behavior** The attach point should remain open when batteries are charging. **Screenshots** None. **Additional context** None.
1.0
Battery Recharger flips back to close with batteries attached. - **Vanilla / HRP Light** Vanilla. **Client Version** 0.5.2 (Vanilla) **Describe the bug** Battery Recharger only open/close its attach points on proximity detection. So if you step away from recharger and a battery is attached, it will flip close. Also the Icon on the charger should go Orange only when it is actually charging a battery. **To Reproduce** Steps to reproduce the behavior: 1. Go to a battery recharger, it will flip open 2. Plug a battery 3. Step away. 4. See error **Expected behavior** The attach point should remain open when batteries are charging. **Screenshots** None. **Additional context** None.
non_test
battery recharger flips back to close with batteries attached vanilla hrp light vanilla client version vanilla describe the bug battery recharger only open close its attach points on proximity detection so if you step away from recharger and a battery is attached it will flip close also the icon on the charger should go orange only when it is actually charging a battery to reproduce steps to reproduce the behavior go to a battery recharger it will flip open plug a battery step away see error expected behavior the attach point should remain open when batteries are charging screenshots none additional context none
0
75,727
9,883,844,571
IssuesEvent
2019-06-24 20:28:45
intel/mkl-dnn
https://api.github.com/repos/intel/mkl-dnn
closed
Incorrect result of s8s8s32 gemm?
documentation
``` using MatrixInt32 = Eigen::Matrix<int, Dynamic, Dynamic, ColMajor>; using MatrixInt8 = Eigen::Matrix<int8_t, Dynamic, Dynamic, ColMajor>; MatrixInt8 lhs_int8(4, 4); lhs_int8 << -47, -123, -10, 92, -96, -38, 82, 40, 70, -60, -66, -127, -46, -91, 109, 101; MatrixInt8 rhs_int8(4, 4); rhs_int8 << 124, 8, -74, 32, -40, 4, -50, -84, 103, -84, 88, -66, 24, 115, 123, 54; std::cout << "LHS: \n" << lhs_int8.cast<int32_t>() << "\n"; std::cout << "RHS: \n" << rhs_int8.cast<int32_t>() << "\n"; MatrixInt32 out(4, 4); out.setZero(); char transposeA = 'N'; char transposeB = 'N'; const int m = 4; const int n = 4; const int k = 4; const int ldA = 4; const int ldB = 4; const int ldC = 4; const float beta = 1.0; // Symmetric quantization with zero point at 0. const int8_t ao = 0; const int8_t bo = 0; // Don't add any offset to the result C. const char offsetc = 'F'; const int32_t co = 0; const int8_t* A = lhs_int8.data(); const int8_t* B = rhs_int8.data(); int32_t* C = out.data(); const float alpha = 1.0; mkldnn_gemm_s8s8s32(&transposeA, &transposeB, &offsetc, // &m, &n, &k, // &alpha, // A, &ldA, &ao, // B, &ldB, &bo, // &beta, // C, &ldC, &co); std::cout << "OUT: \n" << out << "\n"; ``` Output: ``` LHS: -47 -123 -10 92 -96 -38 82 40 70 -60 -66 -127 -46 -91 109 101 RHS: 124 8 -74 32 -40 4 -50 -84 103 -84 88 -66 24 115 123 54 OUT: 270 10552 20064 14456 -978 -3208 21140 -3132 3016 -7744 -10244 4778 3823 1727 13841 4432 ``` But the correct MatMul result is: ``` 270 10552 20064 14456 -978 -3208 21140 -3132 1234 -8741 -23609 4778 11587 1727 29969 4432 ```
1.0
Incorrect result of s8s8s32 gemm? - ``` using MatrixInt32 = Eigen::Matrix<int, Dynamic, Dynamic, ColMajor>; using MatrixInt8 = Eigen::Matrix<int8_t, Dynamic, Dynamic, ColMajor>; MatrixInt8 lhs_int8(4, 4); lhs_int8 << -47, -123, -10, 92, -96, -38, 82, 40, 70, -60, -66, -127, -46, -91, 109, 101; MatrixInt8 rhs_int8(4, 4); rhs_int8 << 124, 8, -74, 32, -40, 4, -50, -84, 103, -84, 88, -66, 24, 115, 123, 54; std::cout << "LHS: \n" << lhs_int8.cast<int32_t>() << "\n"; std::cout << "RHS: \n" << rhs_int8.cast<int32_t>() << "\n"; MatrixInt32 out(4, 4); out.setZero(); char transposeA = 'N'; char transposeB = 'N'; const int m = 4; const int n = 4; const int k = 4; const int ldA = 4; const int ldB = 4; const int ldC = 4; const float beta = 1.0; // Symmetric quantization with zero point at 0. const int8_t ao = 0; const int8_t bo = 0; // Don't add any offset to the result C. const char offsetc = 'F'; const int32_t co = 0; const int8_t* A = lhs_int8.data(); const int8_t* B = rhs_int8.data(); int32_t* C = out.data(); const float alpha = 1.0; mkldnn_gemm_s8s8s32(&transposeA, &transposeB, &offsetc, // &m, &n, &k, // &alpha, // A, &ldA, &ao, // B, &ldB, &bo, // &beta, // C, &ldC, &co); std::cout << "OUT: \n" << out << "\n"; ``` Output: ``` LHS: -47 -123 -10 92 -96 -38 82 40 70 -60 -66 -127 -46 -91 109 101 RHS: 124 8 -74 32 -40 4 -50 -84 103 -84 88 -66 24 115 123 54 OUT: 270 10552 20064 14456 -978 -3208 21140 -3132 3016 -7744 -10244 4778 3823 1727 13841 4432 ``` But the correct MatMul result is: ``` 270 10552 20064 14456 -978 -3208 21140 -3132 1234 -8741 -23609 4778 11587 1727 29969 4432 ```
non_test
incorrect result of gemm using eigen matrix using eigen matrix lhs lhs rhs rhs std cout n std cout n out out setzero char transposea n char transposeb n const int m const int n const int k const int lda const int ldb const int ldc const float beta symmetric quantization with zero point at const t ao const t bo don t add any offset to the result c const char offsetc f const t co const t a lhs data const t b rhs data t c out data const float alpha mkldnn gemm transposea transposeb offsetc m n k alpha a lda ao b ldb bo beta c ldc co std cout out n out n output lhs rhs out but the correct matmul result is
0
531,136
15,441,630,424
IssuesEvent
2021-03-08 06:17:50
code4moldova/voluntar-web
https://api.github.com/repos/code4moldova/voluntar-web
opened
[demand] add information about cluster and volunteer for in_progress, canceled and resolved demands
Priority: Medium
- add link to the cluster to the tabs or in the edit forms - add volunteer information like name and phone number to the edit form or the tabs
1.0
[demand] add information about cluster and volunteer for in_progress, canceled and resolved demands - - add link to the cluster to the tabs or in the edit forms - add volunteer information like name and phone number to the edit form or the tabs
non_test
add information about cluster and volunteer for in progress canceled and resolved demands add link to the cluster to the tabs or in the edit forms add volunteer information like name and phone number to the edit form or the tabs
0
147,365
11,783,741,663
IssuesEvent
2020-03-17 06:34:44
longhorn/longhorn
https://api.github.com/repos/longhorn/longhorn
closed
[BUG] Nightly Upgrade Test: test_pvc_creation_with_default_sc_set failed
area/test bug
**Describe the bug** **To Reproduce** Steps to reproduce the behavior: 1. Install Longhorn v0.7.0 2. Upgrade Longhorn to v0.8.0 3. Run `test_pvc_creation_with_default_sc_set` test **Expected behavior** test should pass **Log** ``` client = <longhorn.Client object at 0x7f4ecbd21250> core_api = <kubernetes.client.apis.core_v1_api.CoreV1Api object at 0x7f4ecbff0190> storage_class = {'allowVolumeExpansion': True, 'apiVersion': 'storage.k8s.io/v1', 'kind': 'StorageClass', 'metadata': {'annotations': {'storageclass.kubernetes.io/is-default-class': 'true'}, 'name': 'longhorn-test'}, ...} pod = {'apiVersion': 'v1', 'kind': 'Pod', 'metadata': {'name': 'pod-test-pvc-creation-with-sc'}, 'spec': {'containers': [{'a...p', ...}], 'volumes': [{'name': 'pod-data', 'persistentVolumeClaim': {'claimName': 'pvc-test-pvc-creation-with-sc'}}]}} @pytest.mark.csi # NOQA def test_pvc_creation_with_default_sc_set( client, core_api, storage_class, pod): # NOQA # set default storage class storage_class['metadata']['annotations'] = \ {"storageclass.kubernetes.io/is-default-class": "true"} create_storage_class(storage_class) static_sc_name = "longhorn-static-test" setting = client.by_id_setting(SETTING_DEFAULT_LONGHORN_STATIC_SC) setting = client.update(setting, value=static_sc_name) assert setting.value == static_sc_name volume_name = "test-pvc-creation-with-sc" pod_name = "pod-" + volume_name client.create_volume(name=volume_name, size=SIZE, numberOfReplicas=2) volume = wait_for_volume_detached(client, volume_name) pv_name = "pv-" + volume_name pvc_name = "pvc-" + volume_name pvc_name_extra = "pvc-" + volume_name + "-extra" create_pv_for_volume(client, core_api, volume, pv_name) create_pvc_for_volume(client, core_api, volume, pvc_name) ret = core_api.list_namespaced_persistent_volume_claim( namespace='default') for item in ret.items: if item.metadata.name == pvc_name: pvc_found = item break assert pvc_found assert pvc_found.spec.storage_class_name == static_sc_name pod['metadata']['name'] = pod_name pod['spec']['volumes'] = [{ 'name': pod['spec']['containers'][0]['volumeMounts'][0]['name'], 'persistentVolumeClaim': { 'claimName': pvc_name, }, }] create_and_wait_pod(core_api, pod) ks = { 'pvName': pv_name, 'pvStatus': 'Bound', 'namespace': 'default', 'pvcName': pvc_name, 'lastPVCRefAt': '', 'lastPodRefAt': '', 'workloadsStatus': [{ 'podName': pod_name, 'podStatus': 'Running', 'workloadName': '', 'workloadType': '', }, ], } wait_volume_kubernetes_status(client, volume_name, ks) delete_and_wait_pod(core_api, pod_name) delete_and_wait_pvc(core_api, pvc_name) ks = { 'pvName': pv_name, 'pvStatus': 'Released', 'namespace': 'default', 'pvcName': pvc_name, 'lastPVCRefAt': 'not empty', 'lastPodRefAt': 'not empty', } wait_volume_kubernetes_status(client, volume_name, ks) # try to reuse the pv volume = wait_for_volume_detached(client, volume_name) > create_pvc_for_volume(client, core_api, volume, pvc_name_extra) test_kubernetes.py:355: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ common.py:2468: in create_pvc_for_volume volume.pvcCreate(namespace="default", pvcName=pvc_name) longhorn.py:263: in cb *args, **kw) longhorn.py:457: in action return self._post_and_retry(url, *args, **kw) longhorn.py:415: in _post_and_retry raise e longhorn.py:409: in _post_and_retry return self._post(url, data=self._to_dict(*args, **kw)) longhorn.py:74: in wrapped return fn(*args, **kw) longhorn.py:303: in _post self._error(r.text, r.status_code) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <longhorn.Client object at 0x7f4ecbd21250> text = '{"actions":{},"code":"Server Error","detail":"","links":{"self":"http://10.42.1.14:9500/v1/volumes/test-pvc-creation-...ated PVC pvc-test-pvc-creation-with-sc-extra doesn\'t bound PV, PV status is Available","status":500,"type":"error"}\n' status_code = 500 def _error(self, text, status_code): > raise ApiError(self._unmarshall(text), status_code) E longhorn.ApiError: (ApiError(...), '500 : unable to create PVC for volume test-pvc-creation-with-sc: created PVC pvc-test-pvc-creation-with-sc-extra doesn\'t bound PV, PV status is Available\n{\'code\': 500, \'detail\': \'\', \'message\': "unable to create PVC for volume test-pvc-creation-with-sc: created PVC pvc-test-pvc-creation-with-sc-extra doesn\'t bound PV, PV status is Available", \'status\': 500}') longhorn.py:283: ApiError ``` **Environment:** - Longhorn version: 0.8.0 - Kubernetes version: v1.17.2 - Node OS type and version: Ubuntu 18.04
1.0
[BUG] Nightly Upgrade Test: test_pvc_creation_with_default_sc_set failed - **Describe the bug** **To Reproduce** Steps to reproduce the behavior: 1. Install Longhorn v0.7.0 2. Upgrade Longhorn to v0.8.0 3. Run `test_pvc_creation_with_default_sc_set` test **Expected behavior** test should pass **Log** ``` client = <longhorn.Client object at 0x7f4ecbd21250> core_api = <kubernetes.client.apis.core_v1_api.CoreV1Api object at 0x7f4ecbff0190> storage_class = {'allowVolumeExpansion': True, 'apiVersion': 'storage.k8s.io/v1', 'kind': 'StorageClass', 'metadata': {'annotations': {'storageclass.kubernetes.io/is-default-class': 'true'}, 'name': 'longhorn-test'}, ...} pod = {'apiVersion': 'v1', 'kind': 'Pod', 'metadata': {'name': 'pod-test-pvc-creation-with-sc'}, 'spec': {'containers': [{'a...p', ...}], 'volumes': [{'name': 'pod-data', 'persistentVolumeClaim': {'claimName': 'pvc-test-pvc-creation-with-sc'}}]}} @pytest.mark.csi # NOQA def test_pvc_creation_with_default_sc_set( client, core_api, storage_class, pod): # NOQA # set default storage class storage_class['metadata']['annotations'] = \ {"storageclass.kubernetes.io/is-default-class": "true"} create_storage_class(storage_class) static_sc_name = "longhorn-static-test" setting = client.by_id_setting(SETTING_DEFAULT_LONGHORN_STATIC_SC) setting = client.update(setting, value=static_sc_name) assert setting.value == static_sc_name volume_name = "test-pvc-creation-with-sc" pod_name = "pod-" + volume_name client.create_volume(name=volume_name, size=SIZE, numberOfReplicas=2) volume = wait_for_volume_detached(client, volume_name) pv_name = "pv-" + volume_name pvc_name = "pvc-" + volume_name pvc_name_extra = "pvc-" + volume_name + "-extra" create_pv_for_volume(client, core_api, volume, pv_name) create_pvc_for_volume(client, core_api, volume, pvc_name) ret = core_api.list_namespaced_persistent_volume_claim( namespace='default') for item in ret.items: if item.metadata.name == pvc_name: pvc_found = item break assert pvc_found assert pvc_found.spec.storage_class_name == static_sc_name pod['metadata']['name'] = pod_name pod['spec']['volumes'] = [{ 'name': pod['spec']['containers'][0]['volumeMounts'][0]['name'], 'persistentVolumeClaim': { 'claimName': pvc_name, }, }] create_and_wait_pod(core_api, pod) ks = { 'pvName': pv_name, 'pvStatus': 'Bound', 'namespace': 'default', 'pvcName': pvc_name, 'lastPVCRefAt': '', 'lastPodRefAt': '', 'workloadsStatus': [{ 'podName': pod_name, 'podStatus': 'Running', 'workloadName': '', 'workloadType': '', }, ], } wait_volume_kubernetes_status(client, volume_name, ks) delete_and_wait_pod(core_api, pod_name) delete_and_wait_pvc(core_api, pvc_name) ks = { 'pvName': pv_name, 'pvStatus': 'Released', 'namespace': 'default', 'pvcName': pvc_name, 'lastPVCRefAt': 'not empty', 'lastPodRefAt': 'not empty', } wait_volume_kubernetes_status(client, volume_name, ks) # try to reuse the pv volume = wait_for_volume_detached(client, volume_name) > create_pvc_for_volume(client, core_api, volume, pvc_name_extra) test_kubernetes.py:355: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ common.py:2468: in create_pvc_for_volume volume.pvcCreate(namespace="default", pvcName=pvc_name) longhorn.py:263: in cb *args, **kw) longhorn.py:457: in action return self._post_and_retry(url, *args, **kw) longhorn.py:415: in _post_and_retry raise e longhorn.py:409: in _post_and_retry return self._post(url, data=self._to_dict(*args, **kw)) longhorn.py:74: in wrapped return fn(*args, **kw) longhorn.py:303: in _post self._error(r.text, r.status_code) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <longhorn.Client object at 0x7f4ecbd21250> text = '{"actions":{},"code":"Server Error","detail":"","links":{"self":"http://10.42.1.14:9500/v1/volumes/test-pvc-creation-...ated PVC pvc-test-pvc-creation-with-sc-extra doesn\'t bound PV, PV status is Available","status":500,"type":"error"}\n' status_code = 500 def _error(self, text, status_code): > raise ApiError(self._unmarshall(text), status_code) E longhorn.ApiError: (ApiError(...), '500 : unable to create PVC for volume test-pvc-creation-with-sc: created PVC pvc-test-pvc-creation-with-sc-extra doesn\'t bound PV, PV status is Available\n{\'code\': 500, \'detail\': \'\', \'message\': "unable to create PVC for volume test-pvc-creation-with-sc: created PVC pvc-test-pvc-creation-with-sc-extra doesn\'t bound PV, PV status is Available", \'status\': 500}') longhorn.py:283: ApiError ``` **Environment:** - Longhorn version: 0.8.0 - Kubernetes version: v1.17.2 - Node OS type and version: Ubuntu 18.04
test
nightly upgrade test test pvc creation with default sc set failed describe the bug to reproduce steps to reproduce the behavior install longhorn upgrade longhorn to run test pvc creation with default sc set test expected behavior test should pass log client core api storage class allowvolumeexpansion true apiversion storage io kind storageclass metadata annotations storageclass kubernetes io is default class true name longhorn test pod apiversion kind pod metadata name pod test pvc creation with sc spec containers volumes pytest mark csi noqa def test pvc creation with default sc set client core api storage class pod noqa set default storage class storage class storageclass kubernetes io is default class true create storage class storage class static sc name longhorn static test setting client by id setting setting default longhorn static sc setting client update setting value static sc name assert setting value static sc name volume name test pvc creation with sc pod name pod volume name client create volume name volume name size size numberofreplicas volume wait for volume detached client volume name pv name pv volume name pvc name pvc volume name pvc name extra pvc volume name extra create pv for volume client core api volume pv name create pvc for volume client core api volume pvc name ret core api list namespaced persistent volume claim namespace default for item in ret items if item metadata name pvc name pvc found item break assert pvc found assert pvc found spec storage class name static sc name pod pod name pod name pod persistentvolumeclaim claimname pvc name create and wait pod core api pod ks pvname pv name pvstatus bound namespace default pvcname pvc name lastpvcrefat lastpodrefat workloadsstatus podname pod name podstatus running workloadname workloadtype wait volume kubernetes status client volume name ks delete and wait pod core api pod name delete and wait pvc core api pvc name ks pvname pv name pvstatus released namespace default pvcname pvc name lastpvcrefat not empty lastpodrefat not empty wait volume kubernetes status client volume name ks try to reuse the pv volume wait for volume detached client volume name create pvc for volume client core api volume pvc name extra test kubernetes py common py in create pvc for volume volume pvccreate namespace default pvcname pvc name longhorn py in cb args kw longhorn py in action return self post and retry url args kw longhorn py in post and retry raise e longhorn py in post and retry return self post url data self to dict args kw longhorn py in wrapped return fn args kw longhorn py in post self error r text r status code self text actions code server error detail links self pvc pvc test pvc creation with sc extra doesn t bound pv pv status is available status type error n status code def error self text status code raise apierror self unmarshall text status code e longhorn apierror apierror unable to create pvc for volume test pvc creation with sc created pvc pvc test pvc creation with sc extra doesn t bound pv pv status is available n code detail message unable to create pvc for volume test pvc creation with sc created pvc pvc test pvc creation with sc extra doesn t bound pv pv status is available status longhorn py apierror environment longhorn version kubernetes version node os type and version ubuntu
1
67,853
17,088,753,460
IssuesEvent
2021-07-08 14:51:38
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
reopened
[SB] [PM] When logged in with default superadmin credentials should force the admin to change password
Feature request P2 Process: Reopened Study builder
Current behavior: When logged in with default superadmin(created from script) credentials for SB and PM, default password is being used unless admin manually changes the password Expected behaviour/approach: When logged in with default superadmin credentials for SB and PM, admins should be forced to change password and navigate to change password post logging in. Needs to be handled for Study Builder and Participant Manager.
1.0
[SB] [PM] When logged in with default superadmin credentials should force the admin to change password - Current behavior: When logged in with default superadmin(created from script) credentials for SB and PM, default password is being used unless admin manually changes the password Expected behaviour/approach: When logged in with default superadmin credentials for SB and PM, admins should be forced to change password and navigate to change password post logging in. Needs to be handled for Study Builder and Participant Manager.
non_test
when logged in with default superadmin credentials should force the admin to change password current behavior when logged in with default superadmin created from script credentials for sb and pm default password is being used unless admin manually changes the password expected behaviour approach when logged in with default superadmin credentials for sb and pm admins should be forced to change password and navigate to change password post logging in needs to be handled for study builder and participant manager
0
420,369
12,237,170,153
IssuesEvent
2020-05-04 17:34:02
balena-io/balena-supervisor
https://api.github.com/repos/balena-io/balena-supervisor
closed
Support api key in http header
Low-hanging fruit Medium Priority type/enhancement
Currently, the supervisor API only support `ApiKey` header (https://github.com/balena-io/balena-supervisor/blob/20e339317e53a592e17f9936bc70b36d8dd17a4b/src/supervisor-api.ts#L15-L27). Can we support `Bearer <API_KEY>` header (the same as Authorization header for the API) as well?
1.0
Support api key in http header - Currently, the supervisor API only support `ApiKey` header (https://github.com/balena-io/balena-supervisor/blob/20e339317e53a592e17f9936bc70b36d8dd17a4b/src/supervisor-api.ts#L15-L27). Can we support `Bearer <API_KEY>` header (the same as Authorization header for the API) as well?
non_test
support api key in http header currently the supervisor api only support apikey header can we support bearer header the same as authorization header for the api as well
0
615,647
19,271,508,825
IssuesEvent
2021-12-10 06:20:44
grpc/grpc
https://api.github.com/repos/grpc/grpc
opened
How to use OpenSSL Engine
kind/enhancement priority/P2
### Is your feature request related to a problem? Please describe. I check the code and add `engine:dynamic:` in the header of key file, but fail. I do not find an example or a doc for it. ### Describe the solution you'd like Provide an example or doc for it. ### Additional context Ref #20316
1.0
How to use OpenSSL Engine - ### Is your feature request related to a problem? Please describe. I check the code and add `engine:dynamic:` in the header of key file, but fail. I do not find an example or a doc for it. ### Describe the solution you'd like Provide an example or doc for it. ### Additional context Ref #20316
non_test
how to use openssl engine is your feature request related to a problem please describe i check the code and add engine dynamic in the header of key file but fail i do not find an example or a doc for it describe the solution you d like provide an example or doc for it additional context ref
0
442,857
30,860,429,018
IssuesEvent
2023-08-03 02:10:57
solidjs/solid-docs-next
https://api.github.com/repos/solidjs/solid-docs-next
closed
[docs-rewrite] Overview
improve documentation in progress add new content
**Objective**: This issue relates to the creation the overview page. This overview page serves as the main entry point for users, providing a high-level understanding of Solid and what it is. **What to Include**: 1. Define Solid. 2. "What is Solid?" Section - Include information about it being a JS framework, reactivity and what makes it unique. 3. Advantages of Solid - Include how it is performant, powerful, pragmatic, and productive. **Details**: - This page should avoid going into too much detail. It is intended to be the first page of the docs. - It should be concise and engaging.
1.0
[docs-rewrite] Overview - **Objective**: This issue relates to the creation the overview page. This overview page serves as the main entry point for users, providing a high-level understanding of Solid and what it is. **What to Include**: 1. Define Solid. 2. "What is Solid?" Section - Include information about it being a JS framework, reactivity and what makes it unique. 3. Advantages of Solid - Include how it is performant, powerful, pragmatic, and productive. **Details**: - This page should avoid going into too much detail. It is intended to be the first page of the docs. - It should be concise and engaging.
non_test
overview objective this issue relates to the creation the overview page this overview page serves as the main entry point for users providing a high level understanding of solid and what it is what to include define solid what is solid section include information about it being a js framework reactivity and what makes it unique advantages of solid include how it is performant powerful pragmatic and productive details this page should avoid going into too much detail it is intended to be the first page of the docs it should be concise and engaging
0
233,321
18,960,543,300
IssuesEvent
2021-11-19 03:49:59
HypothesisWorks/hypothesis
https://api.github.com/repos/HypothesisWorks/hypothesis
closed
Bad shrinking with complex strategy involving mutually_broadcastable_shapes
test-case-reduction
I have a test that generates several mutually broadcastable shapes and an `axes` keyword argument that generates a tuple of axes on the result shape. I have the strategies built using `shared` and `flatmap`. My test fails with certain axis arguments, but, the resulting failure case does not shrink properly at all. Here is a simplified test showing the problem: ```py >>> from hypothesis.strategies import integers, shared, tuples >>> from hypothesis import given >>> from hypothesis.extra.numpy import mutually_broadcastable_shapes >>> mbs = shared(integers(1, 32).flatmap(lambda i: mutually_broadcastable_shapes(num_shapes=i))) >>> axes = mbs.flatmap(lambda bs: tuples(*(integers(-i, max(0, i-1)) for i in range(len(bs.result_shape))))) >>> @given(mbs, axes) ... def test(broadcastable_shapes, axis): ... shapes, result_shape = broadcastable_shapes ... ... if len(shapes) > 1: ... assert set(result_shape) == {result_shape[i] for i in axis} >>> test() Falsifying example: test( broadcastable_shapes=BroadcastableShapes(input_shapes=((), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (2, 1)), result_shape=(2, 1)), axis=(0, 0), ) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "<stdin>", line 2, in test File "/Users/aaronmeurer/anaconda3/envs/ndindex/lib/python3.8/site-packages/hypothesis/core.py", line 1202, in wrapped_test raise the_error_hypothesis_found File "<stdin>", line 6, in test AssertionError ``` Here's proof that the multiple `()`s are unnecessary (the inputs could be shrunk in other dimensions as well). ```py >>> test.hypothesis.inner_test((((), (2, 1)), (2, 1)), (0, 0)) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "<stdin>", line 6, in test AssertionError ``` CC @honno
1.0
Bad shrinking with complex strategy involving mutually_broadcastable_shapes - I have a test that generates several mutually broadcastable shapes and an `axes` keyword argument that generates a tuple of axes on the result shape. I have the strategies built using `shared` and `flatmap`. My test fails with certain axis arguments, but, the resulting failure case does not shrink properly at all. Here is a simplified test showing the problem: ```py >>> from hypothesis.strategies import integers, shared, tuples >>> from hypothesis import given >>> from hypothesis.extra.numpy import mutually_broadcastable_shapes >>> mbs = shared(integers(1, 32).flatmap(lambda i: mutually_broadcastable_shapes(num_shapes=i))) >>> axes = mbs.flatmap(lambda bs: tuples(*(integers(-i, max(0, i-1)) for i in range(len(bs.result_shape))))) >>> @given(mbs, axes) ... def test(broadcastable_shapes, axis): ... shapes, result_shape = broadcastable_shapes ... ... if len(shapes) > 1: ... assert set(result_shape) == {result_shape[i] for i in axis} >>> test() Falsifying example: test( broadcastable_shapes=BroadcastableShapes(input_shapes=((), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (), (2, 1)), result_shape=(2, 1)), axis=(0, 0), ) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "<stdin>", line 2, in test File "/Users/aaronmeurer/anaconda3/envs/ndindex/lib/python3.8/site-packages/hypothesis/core.py", line 1202, in wrapped_test raise the_error_hypothesis_found File "<stdin>", line 6, in test AssertionError ``` Here's proof that the multiple `()`s are unnecessary (the inputs could be shrunk in other dimensions as well). ```py >>> test.hypothesis.inner_test((((), (2, 1)), (2, 1)), (0, 0)) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "<stdin>", line 6, in test AssertionError ``` CC @honno
test
bad shrinking with complex strategy involving mutually broadcastable shapes i have a test that generates several mutually broadcastable shapes and an axes keyword argument that generates a tuple of axes on the result shape i have the strategies built using shared and flatmap my test fails with certain axis arguments but the resulting failure case does not shrink properly at all here is a simplified test showing the problem py from hypothesis strategies import integers shared tuples from hypothesis import given from hypothesis extra numpy import mutually broadcastable shapes mbs shared integers flatmap lambda i mutually broadcastable shapes num shapes i axes mbs flatmap lambda bs tuples integers i max i for i in range len bs result shape given mbs axes def test broadcastable shapes axis shapes result shape broadcastable shapes if len shapes assert set result shape result shape for i in axis test falsifying example test broadcastable shapes broadcastableshapes input shapes result shape axis traceback most recent call last file line in file line in test file users aaronmeurer envs ndindex lib site packages hypothesis core py line in wrapped test raise the error hypothesis found file line in test assertionerror here s proof that the multiple s are unnecessary the inputs could be shrunk in other dimensions as well py test hypothesis inner test traceback most recent call last file line in file line in test assertionerror cc honno
1
33,078
7,022,504,240
IssuesEvent
2017-12-22 10:53:05
scalameta/language-server
https://api.github.com/repos/scalameta/language-server
closed
Key presentation compiler by unamanged source directories instead of sources
defect
Currently, you must re-run scalametaEnableCompletions every time a new file is created. If we key the compiler by sourceDirectories instead of sources then we can pick the right PC for new files.
1.0
Key presentation compiler by unamanged source directories instead of sources - Currently, you must re-run scalametaEnableCompletions every time a new file is created. If we key the compiler by sourceDirectories instead of sources then we can pick the right PC for new files.
non_test
key presentation compiler by unamanged source directories instead of sources currently you must re run scalametaenablecompletions every time a new file is created if we key the compiler by sourcedirectories instead of sources then we can pick the right pc for new files
0
314,644
27,015,271,740
IssuesEvent
2023-02-10 18:45:54
DevTraces/BackEnd
https://api.github.com/repos/DevTraces/BackEnd
closed
feat: JWT 로그인, 로그아웃 구현
☑️test ✨feature
## 📌 이슈 내용 > Spring Security와 JWT를 활용하여 일반 로그인, 로그아웃, 토큰 재발급을 구현한다. ### 로그인 흐름 1. 이메일, 비밀번호가 DB에 저장된 정보가 맞는지 확인한다. 2. Access Token과 Refresh Token을 발급한다. 3. Refresh Token은 Redis에 저장한다. 4. 클라이언트에게 헤더를 통해 토큰 정보들을 반환한다. ### 로그아웃 흐름 1. security filter를 통해 올바른 access token이 헤더에 담겼는지 확인한다. 2. JWT에 담긴 회원 정보를 통해 redis에 저장된 refresh token을 삭제한다. 3. 현재 사용 중인 access token을 블랙리스트에 추가한다. (블랙리스트는 Redis로 관리한다.) ### 토큰 재발급 흐름 1. nickname을 통해 사용자 계정을 조회한다. 5. access token이 만료되었음을 확인한다. (만료되지 않았을 경우 탈취 당한 것으로 간주하고 로그아웃을 진행한다.) 7. refresh token이 만료되지 않았음을 확인한다. 8. refresh token이 Redis에 저장되어 있는 정보와 일치하는지 확인한다. 9. 로그인의 2~4번 과정과 동일하게 토큰을 발급한다. ### 인증인가 흐름 1. 헤더에 "Authorization"이 있는지 확인한다. 2. 유효한 JWT 토큰이 담겼는지 확인한다. 3. 블랙리스트에 저장된 토큰인지 확인한다. (블랙리스트에 저장된 토큰일 경우 인증인가를 불허한다.) 5. JWT에 담긴 사용자 정보를 SecurityContextHolder에 저장한다. 6. @AuthenticationPrincipal 을 통해 JWT에 담긴 사용자 고유번호를 확인할 수 있다. <br> ## 📝 To-do - [x] Spring Security, JWT 라이브러리 추가 - [x] JWT 토큰 인증 - [x] JWT 토큰 생성 - [x] JWT 토큰 검증 - [x] Security Filter Chain에 추가 - [x] 로그인 컨트롤러, 서비스 구현 - [x] 비밀번호 해독화 - [x] 이메일, 비밀번호 일치 여부 확인 - [x] JWT 발급 및 Redis에 저장 - [x] 예외처리 - [x] 로그아웃 컨트롤러, 서비스 구현 - [x] 만료된 Access Token인지 확인 - [x] Access Token 블랙리스트에 추가 (Redis) - [x] Refresh Token Redis에서 삭제 - [x] 예외처리 - [x] 테스트코드 작성 <br> ### Reference
1.0
feat: JWT 로그인, 로그아웃 구현 - ## 📌 이슈 내용 > Spring Security와 JWT를 활용하여 일반 로그인, 로그아웃, 토큰 재발급을 구현한다. ### 로그인 흐름 1. 이메일, 비밀번호가 DB에 저장된 정보가 맞는지 확인한다. 2. Access Token과 Refresh Token을 발급한다. 3. Refresh Token은 Redis에 저장한다. 4. 클라이언트에게 헤더를 통해 토큰 정보들을 반환한다. ### 로그아웃 흐름 1. security filter를 통해 올바른 access token이 헤더에 담겼는지 확인한다. 2. JWT에 담긴 회원 정보를 통해 redis에 저장된 refresh token을 삭제한다. 3. 현재 사용 중인 access token을 블랙리스트에 추가한다. (블랙리스트는 Redis로 관리한다.) ### 토큰 재발급 흐름 1. nickname을 통해 사용자 계정을 조회한다. 5. access token이 만료되었음을 확인한다. (만료되지 않았을 경우 탈취 당한 것으로 간주하고 로그아웃을 진행한다.) 7. refresh token이 만료되지 않았음을 확인한다. 8. refresh token이 Redis에 저장되어 있는 정보와 일치하는지 확인한다. 9. 로그인의 2~4번 과정과 동일하게 토큰을 발급한다. ### 인증인가 흐름 1. 헤더에 "Authorization"이 있는지 확인한다. 2. 유효한 JWT 토큰이 담겼는지 확인한다. 3. 블랙리스트에 저장된 토큰인지 확인한다. (블랙리스트에 저장된 토큰일 경우 인증인가를 불허한다.) 5. JWT에 담긴 사용자 정보를 SecurityContextHolder에 저장한다. 6. @AuthenticationPrincipal 을 통해 JWT에 담긴 사용자 고유번호를 확인할 수 있다. <br> ## 📝 To-do - [x] Spring Security, JWT 라이브러리 추가 - [x] JWT 토큰 인증 - [x] JWT 토큰 생성 - [x] JWT 토큰 검증 - [x] Security Filter Chain에 추가 - [x] 로그인 컨트롤러, 서비스 구현 - [x] 비밀번호 해독화 - [x] 이메일, 비밀번호 일치 여부 확인 - [x] JWT 발급 및 Redis에 저장 - [x] 예외처리 - [x] 로그아웃 컨트롤러, 서비스 구현 - [x] 만료된 Access Token인지 확인 - [x] Access Token 블랙리스트에 추가 (Redis) - [x] Refresh Token Redis에서 삭제 - [x] 예외처리 - [x] 테스트코드 작성 <br> ### Reference
test
feat jwt 로그인 로그아웃 구현 📌 이슈 내용 spring security와 jwt를 활용하여 일반 로그인 로그아웃 토큰 재발급을 구현한다 로그인 흐름 이메일 비밀번호가 db에 저장된 정보가 맞는지 확인한다 access token과 refresh token을 발급한다 refresh token은 redis에 저장한다 클라이언트에게 헤더를 통해 토큰 정보들을 반환한다 로그아웃 흐름 security filter를 통해 올바른 access token이 헤더에 담겼는지 확인한다 jwt에 담긴 회원 정보를 통해 redis에 저장된 refresh token을 삭제한다 현재 사용 중인 access token을 블랙리스트에 추가한다 블랙리스트는 redis로 관리한다 토큰 재발급 흐름 nickname을 통해 사용자 계정을 조회한다 access token이 만료되었음을 확인한다 만료되지 않았을 경우 탈취 당한 것으로 간주하고 로그아웃을 진행한다 refresh token이 만료되지 않았음을 확인한다 refresh token이 redis에 저장되어 있는 정보와 일치하는지 확인한다 로그인의 과정과 동일하게 토큰을 발급한다 인증인가 흐름 헤더에 authorization 이 있는지 확인한다 유효한 jwt 토큰이 담겼는지 확인한다 블랙리스트에 저장된 토큰인지 확인한다 블랙리스트에 저장된 토큰일 경우 인증인가를 불허한다 jwt에 담긴 사용자 정보를 securitycontextholder에 저장한다 authenticationprincipal 을 통해 jwt에 담긴 사용자 고유번호를 확인할 수 있다 📝 to do spring security jwt 라이브러리 추가 jwt 토큰 인증 jwt 토큰 생성 jwt 토큰 검증 security filter chain에 추가 로그인 컨트롤러 서비스 구현 비밀번호 해독화 이메일 비밀번호 일치 여부 확인 jwt 발급 및 redis에 저장 예외처리 로그아웃 컨트롤러 서비스 구현 만료된 access token인지 확인 access token 블랙리스트에 추가 redis refresh token redis에서 삭제 예외처리 테스트코드 작성 reference
1
86,031
16,819,566,699
IssuesEvent
2021-06-17 11:29:48
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
Support search resolve for better repo: filter recognition in 1 click insight creation flow.
crud-ui team/code-insights
[Origin thread about this problem ](https://github.com/sourcegraph/sourcegraph/pull/21943#discussion_r649310260) Now we use regexp for processing values from `repo:` filters but this logic is not reliable in many cases - `repo:^github.com/sourcegraph/sourcegraph-` - `repo:^github.com/sourcegraph/.*\.js$` - `repo:^github.com/sourcegraph/sourcegraph-(sourcegraph|about)$` In order to support all possible cases, we should use a search resolver and use search results instead of runtime regexp processed values from URL. Now we support only `repo:` filter with simple straightforward value like - `repo:^github.com/sourcegraph/about$` - `repo:^github.com/sourcegraph/about$|^github.com/sourcegraph/about$` - `repo: github.com/sourcegraph/about`
1.0
Support search resolve for better repo: filter recognition in 1 click insight creation flow. - [Origin thread about this problem ](https://github.com/sourcegraph/sourcegraph/pull/21943#discussion_r649310260) Now we use regexp for processing values from `repo:` filters but this logic is not reliable in many cases - `repo:^github.com/sourcegraph/sourcegraph-` - `repo:^github.com/sourcegraph/.*\.js$` - `repo:^github.com/sourcegraph/sourcegraph-(sourcegraph|about)$` In order to support all possible cases, we should use a search resolver and use search results instead of runtime regexp processed values from URL. Now we support only `repo:` filter with simple straightforward value like - `repo:^github.com/sourcegraph/about$` - `repo:^github.com/sourcegraph/about$|^github.com/sourcegraph/about$` - `repo: github.com/sourcegraph/about`
non_test
support search resolve for better repo filter recognition in click insight creation flow now we use regexp for processing values from repo filters but this logic is not reliable in many cases repo github com sourcegraph sourcegraph repo github com sourcegraph js repo github com sourcegraph sourcegraph sourcegraph about in order to support all possible cases we should use a search resolver and use search results instead of runtime regexp processed values from url now we support only repo filter with simple straightforward value like repo github com sourcegraph about repo github com sourcegraph about github com sourcegraph about repo github com sourcegraph about
0
83,471
10,354,722,337
IssuesEvent
2019-09-05 14:17:04
nextcloud/server
https://api.github.com/repos/nextcloud/server
closed
Voice in Italian not centered in the button stores the credentials
1. to develop bug design good first issue
In version 16.0.4 localized in Italian, when you go to the setting menu -> basic setting -> mail server the voice saves credentials does not come out completely written in the button area can you fix it? I attach screenshot ![Immagine](https://user-images.githubusercontent.com/35575744/63754743-a8e92c00-c8b5-11e9-85aa-aac12dadfdf3.png)
1.0
Voice in Italian not centered in the button stores the credentials - In version 16.0.4 localized in Italian, when you go to the setting menu -> basic setting -> mail server the voice saves credentials does not come out completely written in the button area can you fix it? I attach screenshot ![Immagine](https://user-images.githubusercontent.com/35575744/63754743-a8e92c00-c8b5-11e9-85aa-aac12dadfdf3.png)
non_test
voice in italian not centered in the button stores the credentials in version localized in italian when you go to the setting menu basic setting mail server the voice saves credentials does not come out completely written in the button area can you fix it i attach screenshot
0
98,420
8,676,621,623
IssuesEvent
2018-11-30 14:38:44
italia/spid
https://api.github.com/repos/italia/spid
closed
Verifica metadata Provincia di Lucca
metadata nuovo md test
Buongiorno, Per conto della Provincia di Lucca abbiamo predisposto i metadata e pubblicati all'URL https://lucca.provincia-online.it/serviziSPID/metadata.xml cordiali saluti Facondini Stefano Maggioli spa [metadata_luccaprovincia-signed.zip](https://github.com/italia/spid/files/2623679/metadata_luccaprovincia-signed.zip)
1.0
Verifica metadata Provincia di Lucca - Buongiorno, Per conto della Provincia di Lucca abbiamo predisposto i metadata e pubblicati all'URL https://lucca.provincia-online.it/serviziSPID/metadata.xml cordiali saluti Facondini Stefano Maggioli spa [metadata_luccaprovincia-signed.zip](https://github.com/italia/spid/files/2623679/metadata_luccaprovincia-signed.zip)
test
verifica metadata provincia di lucca buongiorno per conto della provincia di lucca abbiamo predisposto i metadata e pubblicati all url cordiali saluti facondini stefano maggioli spa
1
230,757
18,685,529,468
IssuesEvent
2021-11-01 11:57:23
Tencent/bk-ci
https://api.github.com/repos/Tencent/bk-ci
closed
模板实例化页面,变量名超长没有作处理
kind/bug stage/uat stage/test area/ci/frontend test/passed uat/passed
### 现象: 如图 ![image](https://user-images.githubusercontent.com/10875468/121622197-305e5a00-caa0-11eb-806d-51094440fdb6.png) ### 期望: 如执行预览页面,变量名和变量值的排布采用上下结构,过长时,hover出浮窗展示变量名全称。 ![image](https://user-images.githubusercontent.com/10875468/121622470-a498fd80-caa0-11eb-972a-02cf4cf8a556.png)
2.0
模板实例化页面,变量名超长没有作处理 - ### 现象: 如图 ![image](https://user-images.githubusercontent.com/10875468/121622197-305e5a00-caa0-11eb-806d-51094440fdb6.png) ### 期望: 如执行预览页面,变量名和变量值的排布采用上下结构,过长时,hover出浮窗展示变量名全称。 ![image](https://user-images.githubusercontent.com/10875468/121622470-a498fd80-caa0-11eb-972a-02cf4cf8a556.png)
test
模板实例化页面,变量名超长没有作处理 现象: 如图 期望: 如执行预览页面,变量名和变量值的排布采用上下结构,过长时,hover出浮窗展示变量名全称。
1
342,513
10,318,398,670
IssuesEvent
2019-08-30 14:52:17
sjawhar/enchanted-brain
https://api.github.com/repos/sjawhar/enchanted-brain
opened
Sign Up Revisions
area/client kind/enhancement priority/q1 size/S
## Expected Behavior - [ ] Make agreement to scientific data use optional, record on user - [ ] Add country of residence, remove name - [ ] Move email and password to separate page - [ ] Explore a method to sign up without email and skip verification (maybe use device UUID as username?)
1.0
Sign Up Revisions - ## Expected Behavior - [ ] Make agreement to scientific data use optional, record on user - [ ] Add country of residence, remove name - [ ] Move email and password to separate page - [ ] Explore a method to sign up without email and skip verification (maybe use device UUID as username?)
non_test
sign up revisions expected behavior make agreement to scientific data use optional record on user add country of residence remove name move email and password to separate page explore a method to sign up without email and skip verification maybe use device uuid as username
0
62,787
6,813,399,325
IssuesEvent
2017-11-06 09:07:57
Kademi/kademi-dev
https://api.github.com/repos/Kademi/kademi-dev
reopened
classes needs @Docs
bug Ready to Test - Dev
Here in following pages : - @Docs needed . /hinda-settings/ - @Docs needed . /funnels/journey1/version1/#properties-tab
1.0
classes needs @Docs - Here in following pages : - @Docs needed . /hinda-settings/ - @Docs needed . /funnels/journey1/version1/#properties-tab
test
classes needs docs here in following pages docs needed hinda settings docs needed funnels properties tab
1
133,426
18,884,540,381
IssuesEvent
2021-11-15 05:35:05
AIESydneyStudents/DIGITAL_BASKET
https://api.github.com/repos/AIESydneyStudents/DIGITAL_BASKET
closed
Making a prototype
Programming Task Design Task
- [x] put in the player #19 - [x] put in collectable Asset - [x] put in the enemy #11 - [x] put in trap #2 - [x] add pickup and drop off to the player #21 #22 - [x] #58 - [x] add placeholder shelf #31 - [x] add placeholder cold shelf #7 - [x] add placeholder stands #10 - [x] win condition - [x] loss condition
1.0
Making a prototype - - [x] put in the player #19 - [x] put in collectable Asset - [x] put in the enemy #11 - [x] put in trap #2 - [x] add pickup and drop off to the player #21 #22 - [x] #58 - [x] add placeholder shelf #31 - [x] add placeholder cold shelf #7 - [x] add placeholder stands #10 - [x] win condition - [x] loss condition
non_test
making a prototype put in the player put in collectable asset put in the enemy put in trap add pickup and drop off to the player add placeholder shelf add placeholder cold shelf add placeholder stands win condition loss condition
0
77,532
14,877,533,855
IssuesEvent
2021-01-20 03:24:00
fprime-community/fpp
https://api.github.com/repos/fprime-community/fpp
closed
Add XML code gen for types defined in components
code generation
Mangle the names. For example, when enum `E` is defined inside component `C`, generate an XML array `C_EArrayAi.ml`. It would be better to define an inner class `C::E`, but the F Prime XML doesn't support this.
1.0
Add XML code gen for types defined in components - Mangle the names. For example, when enum `E` is defined inside component `C`, generate an XML array `C_EArrayAi.ml`. It would be better to define an inner class `C::E`, but the F Prime XML doesn't support this.
non_test
add xml code gen for types defined in components mangle the names for example when enum e is defined inside component c generate an xml array c earrayai ml it would be better to define an inner class c e but the f prime xml doesn t support this
0
155,695
12,267,509,650
IssuesEvent
2020-05-07 10:48:49
tendermint/tendermint
https://api.github.com/repos/tendermint/tendermint
closed
A consensus state test case may be wrong(?)
C:consensus T:test
Run consensus state test case [TestStateLockPOLRelock](https://github.com/tendermint/tendermint/blob/134fe2896275bb926b49743c1e25493f6b24cc31/consensus/state_test.go#L494) with a modification, changing line [571](https://github.com/tendermint/tendermint/blob/134fe2896275bb926b49743c1e25493f6b24cc31/consensus/state_test.go#L571) to: `signAddVotes(cs1, types.PrevoteType, propBlockHash, propBlockParts.Header(), vs3, vs4)` Test result: Pass OK Expected: Test fail because the modified code does not compose majority voting This indicates that either consensus state code or test code has bug(?) For more description, please see [here](https://as2.netlify.com/190901-tendermint-test-relock)
1.0
A consensus state test case may be wrong(?) - Run consensus state test case [TestStateLockPOLRelock](https://github.com/tendermint/tendermint/blob/134fe2896275bb926b49743c1e25493f6b24cc31/consensus/state_test.go#L494) with a modification, changing line [571](https://github.com/tendermint/tendermint/blob/134fe2896275bb926b49743c1e25493f6b24cc31/consensus/state_test.go#L571) to: `signAddVotes(cs1, types.PrevoteType, propBlockHash, propBlockParts.Header(), vs3, vs4)` Test result: Pass OK Expected: Test fail because the modified code does not compose majority voting This indicates that either consensus state code or test code has bug(?) For more description, please see [here](https://as2.netlify.com/190901-tendermint-test-relock)
test
a consensus state test case may be wrong run consensus state test case with a modification changing line to signaddvotes types prevotetype propblockhash propblockparts header test result pass ok expected test fail because the modified code does not compose majority voting this indicates that either consensus state code or test code has bug for more description please see
1
78,252
7,624,500,675
IssuesEvent
2018-05-03 18:13:09
vmware/vic
https://api.github.com/repos/vmware/vic
closed
Implement test for vmwarecna/nginx
component/test priority/p2
User statement: As a customer of VIC, I would like to know that VMware's version of nginx image with photon OS splash screen has been tested and works with VIC. Details: docker pull vmwarecna/nginx Acceptance Criteria: 1. New test is implemented and runs in either nightly or CI.
1.0
Implement test for vmwarecna/nginx - User statement: As a customer of VIC, I would like to know that VMware's version of nginx image with photon OS splash screen has been tested and works with VIC. Details: docker pull vmwarecna/nginx Acceptance Criteria: 1. New test is implemented and runs in either nightly or CI.
test
implement test for vmwarecna nginx user statement as a customer of vic i would like to know that vmware s version of nginx image with photon os splash screen has been tested and works with vic details docker pull vmwarecna nginx acceptance criteria new test is implemented and runs in either nightly or ci
1
155,119
12,239,365,246
IssuesEvent
2020-05-04 21:32:13
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Are there any plans to allow other testing frameworks to be plugged into DevOps?
Pri1 cba devops-test/tech devops/prod product-question
Since CodedUI is kaput --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7b6c92a2-2f5c-e34e-f285-3e1a2a02604b * Version Independent ID: 6f31f4c2-94ac-20eb-da01-e1d47f041945 * Content: [Run automated tests from test plans - Azure Test Plans](https://docs.microsoft.com/en-us/azure/devops/test/run-automated-tests-from-test-hub?view=azure-devops) * Content Source: [docs/test/run-automated-tests-from-test-hub.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/test/run-automated-tests-from-test-hub.md) * Product: **devops** * Technology: **devops-test** * GitHub Login: @steved0x * Microsoft Alias: **sdanie**
1.0
Are there any plans to allow other testing frameworks to be plugged into DevOps? - Since CodedUI is kaput --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7b6c92a2-2f5c-e34e-f285-3e1a2a02604b * Version Independent ID: 6f31f4c2-94ac-20eb-da01-e1d47f041945 * Content: [Run automated tests from test plans - Azure Test Plans](https://docs.microsoft.com/en-us/azure/devops/test/run-automated-tests-from-test-hub?view=azure-devops) * Content Source: [docs/test/run-automated-tests-from-test-hub.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/test/run-automated-tests-from-test-hub.md) * Product: **devops** * Technology: **devops-test** * GitHub Login: @steved0x * Microsoft Alias: **sdanie**
test
are there any plans to allow other testing frameworks to be plugged into devops since codedui is kaput document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops test github login microsoft alias sdanie
1
238,682
19,752,516,836
IssuesEvent
2022-01-15 07:27:27
sahib/rmlint
https://api.github.com/repos/sahib/rmlint
closed
MacPorts install option?
Feature Request In Progress Testers Wanted
Been working at purging our systems of Homebrew due to some various reasons and trouble that it causes. MacPorts has proven to be much more reliable and solid a package manager. Has anyone successfully built an `rmlint` MacPorts recipe as of yet or is this still something that needs to be done?
1.0
MacPorts install option? - Been working at purging our systems of Homebrew due to some various reasons and trouble that it causes. MacPorts has proven to be much more reliable and solid a package manager. Has anyone successfully built an `rmlint` MacPorts recipe as of yet or is this still something that needs to be done?
test
macports install option been working at purging our systems of homebrew due to some various reasons and trouble that it causes macports has proven to be much more reliable and solid a package manager has anyone successfully built an rmlint macports recipe as of yet or is this still something that needs to be done
1
292,466
25,215,824,311
IssuesEvent
2022-11-14 09:05:22
Test-Automation-Crash-Course-24-10-22/team_05
https://api.github.com/repos/Test-Automation-Crash-Course-24-10-22/team_05
opened
Choosing the alcohol product when the user is being under the age of 18
TestCase
**Descriptions:** Сhecking the impossibility of choosing the alcohol product when the user is being under the age of 18. **Preconditions:** The main page https://rozetka.com.ua/ua/ is opened. **Test steps** | Step | Test Data | Expected result | | ------------- | ------------- | ------------- | | 1 | Click the сatalog button on the footer. | The drop-down is displayed on the screen. | | 2 | Click on the“Alcohol and other drinks ” section on the left bottom corner of the drop-down menu. | The “Drinks” tab is opened. | | 3 | Click the “Alcohol” link in the opened tab. | The pop-up "Confirm your age" is displayed. | | 4 | Click the “I am under the age of 18” button. | The “Goods for kids” tab is opened. |
1.0
Choosing the alcohol product when the user is being under the age of 18 - **Descriptions:** Сhecking the impossibility of choosing the alcohol product when the user is being under the age of 18. **Preconditions:** The main page https://rozetka.com.ua/ua/ is opened. **Test steps** | Step | Test Data | Expected result | | ------------- | ------------- | ------------- | | 1 | Click the сatalog button on the footer. | The drop-down is displayed on the screen. | | 2 | Click on the“Alcohol and other drinks ” section on the left bottom corner of the drop-down menu. | The “Drinks” tab is opened. | | 3 | Click the “Alcohol” link in the opened tab. | The pop-up "Confirm your age" is displayed. | | 4 | Click the “I am under the age of 18” button. | The “Goods for kids” tab is opened. |
test
choosing the alcohol product when the user is being under the age of descriptions сhecking the impossibility of choosing the alcohol product when the user is being under the age of preconditions the main page is opened test steps step test data expected result click the сatalog button on the footer the drop down is displayed on the screen click on the“alcohol and other drinks ” section on the left bottom corner of the drop down menu the “drinks” tab is opened click the “alcohol” link in the opened tab the pop up confirm your age is displayed click the “i am under the age of ” button the “goods for kids” tab is opened
1
274,857
23,872,850,806
IssuesEvent
2022-09-07 16:10:48
numpy/numpy
https://api.github.com/repos/numpy/numpy
closed
BUG: np.equal.reduce no longer works with int and float
05 - Testing 57 - Close? Sprint
### Describe the issue: In `1.22.1` the below code fails. However, it used to work in `1.21.5`. It still works with `object` and `bool` dtypes. ### Reproduce the code example: ```python np.equal.reduce([0, 0]) ``` ### Error message: ```shell TypeError: No loop matching the specified signature and casting was found for ufunc equal ``` ### NumPy/Python version information: 1.22.1 3.10.2 (main, Jan 15 2022, 19:56:27) [GCC 11.1.0]
1.0
BUG: np.equal.reduce no longer works with int and float - ### Describe the issue: In `1.22.1` the below code fails. However, it used to work in `1.21.5`. It still works with `object` and `bool` dtypes. ### Reproduce the code example: ```python np.equal.reduce([0, 0]) ``` ### Error message: ```shell TypeError: No loop matching the specified signature and casting was found for ufunc equal ``` ### NumPy/Python version information: 1.22.1 3.10.2 (main, Jan 15 2022, 19:56:27) [GCC 11.1.0]
test
bug np equal reduce no longer works with int and float describe the issue in the below code fails however it used to work in it still works with object and bool dtypes reproduce the code example python np equal reduce error message shell typeerror no loop matching the specified signature and casting was found for ufunc equal numpy python version information main jan
1
328,820
28,137,530,534
IssuesEvent
2023-04-01 15:03:02
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix tensor.test_tensorflow_instance_sub
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4549431744/jobs/8021498736" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4549431744/jobs/8021498736" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4549431744/jobs/8021498736" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4549431744/jobs/8021498736" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_tensor.py::test_tensorflow_instance_sub[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-29T02:15:21.8863843Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8866929Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8870964Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8871417Z E Falsifying example: test_tensorflow_instance_sub( 2023-03-29T02:15:21.8871778Z E dtype_and_x=(['uint16', 'uint16'], 2023-03-29T02:15:21.8872075Z E [array(0, dtype=uint16), array(0, dtype=uint16)]), 2023-03-29T02:15:21.8873133Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.tensorflow' from '/ivy/ivy/functional/frontends/tensorflow/__init__.py'>, framework_init_module=<module 'tensorflow' from '/usr/local/lib/python3.8/dist-packages/tensorflow/__init__.py'>, init_name='constant', method_name='__sub__'), 2023-03-29T02:15:21.8873807Z E method_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8874081Z E num_positional_args=0, 2023-03-29T02:15:21.8874323Z E as_variable=[False], 2023-03-29T02:15:21.8874556Z E native_arrays=[False], 2023-03-29T02:15:21.8874770Z E ), 2023-03-29T02:15:21.8875013Z E init_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8875277Z E num_positional_args=0, 2023-03-29T02:15:21.8875515Z E as_variable=[False], 2023-03-29T02:15:21.8875740Z E native_arrays=[False], 2023-03-29T02:15:21.8875955Z E ), 2023-03-29T02:15:21.8876212Z E frontend='tensorflow', 2023-03-29T02:15:21.8876486Z E on_device='cpu', 2023-03-29T02:15:21.8876691Z E ) 2023-03-29T02:15:21.8876875Z E 2023-03-29T02:15:21.8877362Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2RABQAAIAAC') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_tensor.py::test_tensorflow_instance_sub[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-29T02:15:21.8863843Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8866929Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8870964Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8871417Z E Falsifying example: test_tensorflow_instance_sub( 2023-03-29T02:15:21.8871778Z E dtype_and_x=(['uint16', 'uint16'], 2023-03-29T02:15:21.8872075Z E [array(0, dtype=uint16), array(0, dtype=uint16)]), 2023-03-29T02:15:21.8873133Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.tensorflow' from '/ivy/ivy/functional/frontends/tensorflow/__init__.py'>, framework_init_module=<module 'tensorflow' from '/usr/local/lib/python3.8/dist-packages/tensorflow/__init__.py'>, init_name='constant', method_name='__sub__'), 2023-03-29T02:15:21.8873807Z E method_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8874081Z E num_positional_args=0, 2023-03-29T02:15:21.8874323Z E as_variable=[False], 2023-03-29T02:15:21.8874556Z E native_arrays=[False], 2023-03-29T02:15:21.8874770Z E ), 2023-03-29T02:15:21.8875013Z E init_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8875277Z E num_positional_args=0, 2023-03-29T02:15:21.8875515Z E as_variable=[False], 2023-03-29T02:15:21.8875740Z E native_arrays=[False], 2023-03-29T02:15:21.8875955Z E ), 2023-03-29T02:15:21.8876212Z E frontend='tensorflow', 2023-03-29T02:15:21.8876486Z E on_device='cpu', 2023-03-29T02:15:21.8876691Z E ) 2023-03-29T02:15:21.8876875Z E 2023-03-29T02:15:21.8877362Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2RABQAAIAAC') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_tensor.py::test_tensorflow_instance_sub[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-29T02:15:21.8863843Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8866929Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8870964Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8871417Z E Falsifying example: test_tensorflow_instance_sub( 2023-03-29T02:15:21.8871778Z E dtype_and_x=(['uint16', 'uint16'], 2023-03-29T02:15:21.8872075Z E [array(0, dtype=uint16), array(0, dtype=uint16)]), 2023-03-29T02:15:21.8873133Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.tensorflow' from '/ivy/ivy/functional/frontends/tensorflow/__init__.py'>, framework_init_module=<module 'tensorflow' from '/usr/local/lib/python3.8/dist-packages/tensorflow/__init__.py'>, init_name='constant', method_name='__sub__'), 2023-03-29T02:15:21.8873807Z E method_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8874081Z E num_positional_args=0, 2023-03-29T02:15:21.8874323Z E as_variable=[False], 2023-03-29T02:15:21.8874556Z E native_arrays=[False], 2023-03-29T02:15:21.8874770Z E ), 2023-03-29T02:15:21.8875013Z E init_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8875277Z E num_positional_args=0, 2023-03-29T02:15:21.8875515Z E as_variable=[False], 2023-03-29T02:15:21.8875740Z E native_arrays=[False], 2023-03-29T02:15:21.8875955Z E ), 2023-03-29T02:15:21.8876212Z E frontend='tensorflow', 2023-03-29T02:15:21.8876486Z E on_device='cpu', 2023-03-29T02:15:21.8876691Z E ) 2023-03-29T02:15:21.8876875Z E 2023-03-29T02:15:21.8877362Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2RABQAAIAAC') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_tensor.py::test_tensorflow_instance_sub[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-29T02:15:21.8863843Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8866929Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8870964Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8871417Z E Falsifying example: test_tensorflow_instance_sub( 2023-03-29T02:15:21.8871778Z E dtype_and_x=(['uint16', 'uint16'], 2023-03-29T02:15:21.8872075Z E [array(0, dtype=uint16), array(0, dtype=uint16)]), 2023-03-29T02:15:21.8873133Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.tensorflow' from '/ivy/ivy/functional/frontends/tensorflow/__init__.py'>, framework_init_module=<module 'tensorflow' from '/usr/local/lib/python3.8/dist-packages/tensorflow/__init__.py'>, init_name='constant', method_name='__sub__'), 2023-03-29T02:15:21.8873807Z E method_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8874081Z E num_positional_args=0, 2023-03-29T02:15:21.8874323Z E as_variable=[False], 2023-03-29T02:15:21.8874556Z E native_arrays=[False], 2023-03-29T02:15:21.8874770Z E ), 2023-03-29T02:15:21.8875013Z E init_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8875277Z E num_positional_args=0, 2023-03-29T02:15:21.8875515Z E as_variable=[False], 2023-03-29T02:15:21.8875740Z E native_arrays=[False], 2023-03-29T02:15:21.8875955Z E ), 2023-03-29T02:15:21.8876212Z E frontend='tensorflow', 2023-03-29T02:15:21.8876486Z E on_device='cpu', 2023-03-29T02:15:21.8876691Z E ) 2023-03-29T02:15:21.8876875Z E 2023-03-29T02:15:21.8877362Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2RABQAAIAAC') as a decorator on your test case </details>
1.0
Fix tensor.test_tensorflow_instance_sub - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4549431744/jobs/8021498736" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4549431744/jobs/8021498736" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4549431744/jobs/8021498736" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4549431744/jobs/8021498736" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_tensor.py::test_tensorflow_instance_sub[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-29T02:15:21.8863843Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8866929Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8870964Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8871417Z E Falsifying example: test_tensorflow_instance_sub( 2023-03-29T02:15:21.8871778Z E dtype_and_x=(['uint16', 'uint16'], 2023-03-29T02:15:21.8872075Z E [array(0, dtype=uint16), array(0, dtype=uint16)]), 2023-03-29T02:15:21.8873133Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.tensorflow' from '/ivy/ivy/functional/frontends/tensorflow/__init__.py'>, framework_init_module=<module 'tensorflow' from '/usr/local/lib/python3.8/dist-packages/tensorflow/__init__.py'>, init_name='constant', method_name='__sub__'), 2023-03-29T02:15:21.8873807Z E method_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8874081Z E num_positional_args=0, 2023-03-29T02:15:21.8874323Z E as_variable=[False], 2023-03-29T02:15:21.8874556Z E native_arrays=[False], 2023-03-29T02:15:21.8874770Z E ), 2023-03-29T02:15:21.8875013Z E init_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8875277Z E num_positional_args=0, 2023-03-29T02:15:21.8875515Z E as_variable=[False], 2023-03-29T02:15:21.8875740Z E native_arrays=[False], 2023-03-29T02:15:21.8875955Z E ), 2023-03-29T02:15:21.8876212Z E frontend='tensorflow', 2023-03-29T02:15:21.8876486Z E on_device='cpu', 2023-03-29T02:15:21.8876691Z E ) 2023-03-29T02:15:21.8876875Z E 2023-03-29T02:15:21.8877362Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2RABQAAIAAC') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_tensor.py::test_tensorflow_instance_sub[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-29T02:15:21.8863843Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8866929Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8870964Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8871417Z E Falsifying example: test_tensorflow_instance_sub( 2023-03-29T02:15:21.8871778Z E dtype_and_x=(['uint16', 'uint16'], 2023-03-29T02:15:21.8872075Z E [array(0, dtype=uint16), array(0, dtype=uint16)]), 2023-03-29T02:15:21.8873133Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.tensorflow' from '/ivy/ivy/functional/frontends/tensorflow/__init__.py'>, framework_init_module=<module 'tensorflow' from '/usr/local/lib/python3.8/dist-packages/tensorflow/__init__.py'>, init_name='constant', method_name='__sub__'), 2023-03-29T02:15:21.8873807Z E method_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8874081Z E num_positional_args=0, 2023-03-29T02:15:21.8874323Z E as_variable=[False], 2023-03-29T02:15:21.8874556Z E native_arrays=[False], 2023-03-29T02:15:21.8874770Z E ), 2023-03-29T02:15:21.8875013Z E init_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8875277Z E num_positional_args=0, 2023-03-29T02:15:21.8875515Z E as_variable=[False], 2023-03-29T02:15:21.8875740Z E native_arrays=[False], 2023-03-29T02:15:21.8875955Z E ), 2023-03-29T02:15:21.8876212Z E frontend='tensorflow', 2023-03-29T02:15:21.8876486Z E on_device='cpu', 2023-03-29T02:15:21.8876691Z E ) 2023-03-29T02:15:21.8876875Z E 2023-03-29T02:15:21.8877362Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2RABQAAIAAC') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_tensor.py::test_tensorflow_instance_sub[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-29T02:15:21.8863843Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8866929Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8870964Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8871417Z E Falsifying example: test_tensorflow_instance_sub( 2023-03-29T02:15:21.8871778Z E dtype_and_x=(['uint16', 'uint16'], 2023-03-29T02:15:21.8872075Z E [array(0, dtype=uint16), array(0, dtype=uint16)]), 2023-03-29T02:15:21.8873133Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.tensorflow' from '/ivy/ivy/functional/frontends/tensorflow/__init__.py'>, framework_init_module=<module 'tensorflow' from '/usr/local/lib/python3.8/dist-packages/tensorflow/__init__.py'>, init_name='constant', method_name='__sub__'), 2023-03-29T02:15:21.8873807Z E method_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8874081Z E num_positional_args=0, 2023-03-29T02:15:21.8874323Z E as_variable=[False], 2023-03-29T02:15:21.8874556Z E native_arrays=[False], 2023-03-29T02:15:21.8874770Z E ), 2023-03-29T02:15:21.8875013Z E init_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8875277Z E num_positional_args=0, 2023-03-29T02:15:21.8875515Z E as_variable=[False], 2023-03-29T02:15:21.8875740Z E native_arrays=[False], 2023-03-29T02:15:21.8875955Z E ), 2023-03-29T02:15:21.8876212Z E frontend='tensorflow', 2023-03-29T02:15:21.8876486Z E on_device='cpu', 2023-03-29T02:15:21.8876691Z E ) 2023-03-29T02:15:21.8876875Z E 2023-03-29T02:15:21.8877362Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2RABQAAIAAC') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_tensor.py::test_tensorflow_instance_sub[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-29T02:15:21.8863843Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8866929Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8870964Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-29T02:15:21.8871417Z E Falsifying example: test_tensorflow_instance_sub( 2023-03-29T02:15:21.8871778Z E dtype_and_x=(['uint16', 'uint16'], 2023-03-29T02:15:21.8872075Z E [array(0, dtype=uint16), array(0, dtype=uint16)]), 2023-03-29T02:15:21.8873133Z E frontend_method_data=FrontendMethodData(ivy_init_module=<module 'ivy.functional.frontends.tensorflow' from '/ivy/ivy/functional/frontends/tensorflow/__init__.py'>, framework_init_module=<module 'tensorflow' from '/usr/local/lib/python3.8/dist-packages/tensorflow/__init__.py'>, init_name='constant', method_name='__sub__'), 2023-03-29T02:15:21.8873807Z E method_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8874081Z E num_positional_args=0, 2023-03-29T02:15:21.8874323Z E as_variable=[False], 2023-03-29T02:15:21.8874556Z E native_arrays=[False], 2023-03-29T02:15:21.8874770Z E ), 2023-03-29T02:15:21.8875013Z E init_flags=FrontendMethodTestFlags( 2023-03-29T02:15:21.8875277Z E num_positional_args=0, 2023-03-29T02:15:21.8875515Z E as_variable=[False], 2023-03-29T02:15:21.8875740Z E native_arrays=[False], 2023-03-29T02:15:21.8875955Z E ), 2023-03-29T02:15:21.8876212Z E frontend='tensorflow', 2023-03-29T02:15:21.8876486Z E on_device='cpu', 2023-03-29T02:15:21.8876691Z E ) 2023-03-29T02:15:21.8876875Z E 2023-03-29T02:15:21.8877362Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2RABQAAIAAC') as a decorator on your test case </details>
test
fix tensor test tensorflow instance sub tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test tensorflow test tensor py test tensorflow instance sub e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow instance sub e dtype and x e e frontend method data frontendmethoddata ivy init module framework init module init name constant method name sub e method flags frontendmethodtestflags e num positional args e as variable e native arrays e e init flags frontendmethodtestflags e num positional args e as variable e native arrays e e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test tensor py test tensorflow instance sub e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow instance sub e dtype and x e e frontend method data frontendmethoddata ivy init module framework init module init name constant method name sub e method flags frontendmethodtestflags e num positional args e as variable e native arrays e e init flags frontendmethodtestflags e num positional args e as variable e native arrays e e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test tensor py test tensorflow instance sub e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow instance sub e dtype and x e e frontend method data frontendmethoddata ivy init module framework init module init name constant method name sub e method flags frontendmethodtestflags e num positional args e as variable e native arrays e e init flags frontendmethodtestflags e num positional args e as variable e native arrays e e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test tensor py test tensorflow instance sub e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow instance sub e dtype and x e e frontend method data frontendmethoddata ivy init module framework init module init name constant method name sub e method flags frontendmethodtestflags e num positional args e as variable e native arrays e e init flags frontendmethodtestflags e num positional args e as variable e native arrays e e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
294,638
9,037,919,882
IssuesEvent
2019-02-09 15:27:51
esteemapp/esteem-surfer
https://api.github.com/repos/esteemapp/esteem-surfer
closed
issues/suggestions 1.0.9
bug enhancement high priority
Firstly, the scheduler is showing the current date and time - not the date and time that the post is scheduled to be posted. Secondly, although I like the notifications thing (and just turned it off), it would be nice to have some sort of in between mode. I don't mind it telling me that people have upvoted me, but I can just imagine it next time I get @curied... 550 votes the other day! My poor computer will be a flood of blue notifications! Oh, one more thing - any chance of getting templates so I don't have to keep copy/pasting from my text file? (And I'd love an integrated spell-check - even though I often have arguments (and win!) with it, it's still a very useful tool.) Keep up the great work!
1.0
issues/suggestions 1.0.9 - Firstly, the scheduler is showing the current date and time - not the date and time that the post is scheduled to be posted. Secondly, although I like the notifications thing (and just turned it off), it would be nice to have some sort of in between mode. I don't mind it telling me that people have upvoted me, but I can just imagine it next time I get @curied... 550 votes the other day! My poor computer will be a flood of blue notifications! Oh, one more thing - any chance of getting templates so I don't have to keep copy/pasting from my text file? (And I'd love an integrated spell-check - even though I often have arguments (and win!) with it, it's still a very useful tool.) Keep up the great work!
non_test
issues suggestions firstly the scheduler is showing the current date and time not the date and time that the post is scheduled to be posted secondly although i like the notifications thing and just turned it off it would be nice to have some sort of in between mode i don t mind it telling me that people have upvoted me but i can just imagine it next time i get curied votes the other day my poor computer will be a flood of blue notifications oh one more thing any chance of getting templates so i don t have to keep copy pasting from my text file and i d love an integrated spell check even though i often have arguments and win with it it s still a very useful tool keep up the great work
0
254,065
21,726,196,117
IssuesEvent
2022-05-11 07:53:53
owncloud/client
https://api.github.com/repos/owncloud/client
closed
[Linux] Move + overwrite does upload + delete instead of server MOVE
bug ReadyToTest Linux
### Steps 1. Create a file "b.txt" with contents 1. Wait for sync 1. Create a file "a.txt" with contents 1. Wait for sync 1. `mv a.txt b.txt` 1. Wait for sync 1. Check activity ### Expected result Server MOVE ### Actual result ~~The file "b.txt" is first uploaded, and then "a.txt" is deleted instead of MOVE.~~ First "a.txt" is deleted and then "b.txt" is uploaded, instead of MOVE. ### Version Observed on 2.5.1 by me and on 2.5.4 daily by @jnweiger ### Env My env is openSUSE Tumbleweed 20190126 with owncloud-client-2.5.1.10973-1.1.x86_64 Note: I don't think it's a problem with inotify as I only have about 1300 entries in my local folder. @guruz
1.0
[Linux] Move + overwrite does upload + delete instead of server MOVE - ### Steps 1. Create a file "b.txt" with contents 1. Wait for sync 1. Create a file "a.txt" with contents 1. Wait for sync 1. `mv a.txt b.txt` 1. Wait for sync 1. Check activity ### Expected result Server MOVE ### Actual result ~~The file "b.txt" is first uploaded, and then "a.txt" is deleted instead of MOVE.~~ First "a.txt" is deleted and then "b.txt" is uploaded, instead of MOVE. ### Version Observed on 2.5.1 by me and on 2.5.4 daily by @jnweiger ### Env My env is openSUSE Tumbleweed 20190126 with owncloud-client-2.5.1.10973-1.1.x86_64 Note: I don't think it's a problem with inotify as I only have about 1300 entries in my local folder. @guruz
test
move overwrite does upload delete instead of server move steps create a file b txt with contents wait for sync create a file a txt with contents wait for sync mv a txt b txt wait for sync check activity expected result server move actual result the file b txt is first uploaded and then a txt is deleted instead of move first a txt is deleted and then b txt is uploaded instead of move version observed on by me and on daily by jnweiger env my env is opensuse tumbleweed with owncloud client note i don t think it s a problem with inotify as i only have about entries in my local folder guruz
1
93,090
8,393,783,895
IssuesEvent
2018-10-09 21:37:29
Microsoft/openenclave
https://api.github.com/repos/Microsoft/openenclave
closed
Samples tests do not work for gcc/g++ builds
testing
The samples tests are also not enabled for gcc/g++ CI builds.
1.0
Samples tests do not work for gcc/g++ builds - The samples tests are also not enabled for gcc/g++ CI builds.
test
samples tests do not work for gcc g builds the samples tests are also not enabled for gcc g ci builds
1
302,431
26,143,742,649
IssuesEvent
2022-12-29 23:18:50
pytorch/torchdynamo
https://api.github.com/repos/pytorch/torchdynamo
closed
TorchTidy Fixes in TorchBench
test/benchmark
Torch Tidy recommends which backends to use, and those backend calls are added in TorchBench.
1.0
TorchTidy Fixes in TorchBench - Torch Tidy recommends which backends to use, and those backend calls are added in TorchBench.
test
torchtidy fixes in torchbench torch tidy recommends which backends to use and those backend calls are added in torchbench
1
180,933
21,625,859,024
IssuesEvent
2022-05-05 01:59:21
AlexRogalskiy/java-patterns
https://api.github.com/repos/AlexRogalskiy/java-patterns
opened
CVE-2020-8244 (Medium) detected in bl-1.2.1.tgz, bl-1.1.2.tgz
security vulnerability
## CVE-2020-8244 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bl-1.2.1.tgz</b>, <b>bl-1.1.2.tgz</b></p></summary> <p> <details><summary><b>bl-1.2.1.tgz</b></p></summary> <p>Buffer List: collect buffers and access with a standard readable Buffer interface, streamable too!</p> <p>Library home page: <a href="https://registry.npmjs.org/bl/-/bl-1.2.1.tgz">https://registry.npmjs.org/bl/-/bl-1.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/pacote/node_modules/tar-stream/node_modules/bl/package.json</p> <p> Dependency Hierarchy: - gitbook-cli-2.3.2.tgz (Root Library) - npm-5.1.0.tgz - pacote-2.7.38.tgz - tar-stream-1.5.4.tgz - :x: **bl-1.2.1.tgz** (Vulnerable Library) </details> <details><summary><b>bl-1.1.2.tgz</b></p></summary> <p>Buffer List: collect buffers and access with a standard readable Buffer interface, streamable too!</p> <p>Library home page: <a href="https://registry.npmjs.org/bl/-/bl-1.1.2.tgz">https://registry.npmjs.org/bl/-/bl-1.1.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npmi/node_modules/npm/node_modules/request/node_modules/bl/package.json</p> <p> Dependency Hierarchy: - gitbook-cli-2.3.2.tgz (Root Library) - npmi-1.0.1.tgz - npm-2.15.12.tgz - request-2.74.0.tgz - :x: **bl-1.1.2.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A buffer over-read vulnerability exists in bl <4.0.3, <3.0.1, <2.2.1, and <1.2.3 which could allow an attacker to supply user input (even typed) that if it ends up in consume() argument and can become negative, the BufferList state can be corrupted, tricking it into exposing uninitialized memory via regular .slice() calls. <p>Publish Date: 2020-08-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244>CVE-2020-8244</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-pp7h-53gx-mx7r">https://github.com/advisories/GHSA-pp7h-53gx-mx7r</a></p> <p>Release Date: 2020-08-30</p> <p>Fix Resolution: bl - 1.2.3,2.2.1,3.0.1,4.0.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8244 (Medium) detected in bl-1.2.1.tgz, bl-1.1.2.tgz - ## CVE-2020-8244 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bl-1.2.1.tgz</b>, <b>bl-1.1.2.tgz</b></p></summary> <p> <details><summary><b>bl-1.2.1.tgz</b></p></summary> <p>Buffer List: collect buffers and access with a standard readable Buffer interface, streamable too!</p> <p>Library home page: <a href="https://registry.npmjs.org/bl/-/bl-1.2.1.tgz">https://registry.npmjs.org/bl/-/bl-1.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/pacote/node_modules/tar-stream/node_modules/bl/package.json</p> <p> Dependency Hierarchy: - gitbook-cli-2.3.2.tgz (Root Library) - npm-5.1.0.tgz - pacote-2.7.38.tgz - tar-stream-1.5.4.tgz - :x: **bl-1.2.1.tgz** (Vulnerable Library) </details> <details><summary><b>bl-1.1.2.tgz</b></p></summary> <p>Buffer List: collect buffers and access with a standard readable Buffer interface, streamable too!</p> <p>Library home page: <a href="https://registry.npmjs.org/bl/-/bl-1.1.2.tgz">https://registry.npmjs.org/bl/-/bl-1.1.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npmi/node_modules/npm/node_modules/request/node_modules/bl/package.json</p> <p> Dependency Hierarchy: - gitbook-cli-2.3.2.tgz (Root Library) - npmi-1.0.1.tgz - npm-2.15.12.tgz - request-2.74.0.tgz - :x: **bl-1.1.2.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A buffer over-read vulnerability exists in bl <4.0.3, <3.0.1, <2.2.1, and <1.2.3 which could allow an attacker to supply user input (even typed) that if it ends up in consume() argument and can become negative, the BufferList state can be corrupted, tricking it into exposing uninitialized memory via regular .slice() calls. <p>Publish Date: 2020-08-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8244>CVE-2020-8244</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-pp7h-53gx-mx7r">https://github.com/advisories/GHSA-pp7h-53gx-mx7r</a></p> <p>Release Date: 2020-08-30</p> <p>Fix Resolution: bl - 1.2.3,2.2.1,3.0.1,4.0.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in bl tgz bl tgz cve medium severity vulnerability vulnerable libraries bl tgz bl tgz bl tgz buffer list collect buffers and access with a standard readable buffer interface streamable too library home page a href path to dependency file package json path to vulnerable library node modules npm node modules pacote node modules tar stream node modules bl package json dependency hierarchy gitbook cli tgz root library npm tgz pacote tgz tar stream tgz x bl tgz vulnerable library bl tgz buffer list collect buffers and access with a standard readable buffer interface streamable too library home page a href path to dependency file package json path to vulnerable library node modules npmi node modules npm node modules request node modules bl package json dependency hierarchy gitbook cli tgz root library npmi tgz npm tgz request tgz x bl tgz vulnerable library found in base branch master vulnerability details a buffer over read vulnerability exists in bl and which could allow an attacker to supply user input even typed that if it ends up in consume argument and can become negative the bufferlist state can be corrupted tricking it into exposing uninitialized memory via regular slice calls publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bl step up your open source security game with whitesource
0
179,480
13,883,476,783
IssuesEvent
2020-10-18 12:10:50
Exiv2/exiv2
https://api.github.com/repos/Exiv2/exiv2
closed
Test Localisation
testing
Localisation is not tested in the test suite. It was never added because I didn't know how to use localisation. Furthermore, I found it almost impossible to build because of dependencies. Setting up the platform dependencies is now well documented in README.md, or you can use conan. I documented localisation in README.md for v0.27. You build localisation with the cmake option `-DEXIV2_ENABLE_NLS=On`. To use localisation from the command line: ```bash 680 rmills@rmillsmbp:~/gnu/github/exiv2/0.27-maintenance/build $ env LANG=fr_FR exiv2 exiv2: Une action doit être spécifié exiv2: Au moins un fichier est nécessaire Utilisation : exiv2 [ options ] [ action ] fichier ... Manipulation des métadonnées EXIF issues des images. 681 rmills@rmillsmbp:~/gnu/github/exiv2/0.27-maintenance/build $ ``` We should skip the test if NLS (natural language support) has not been built. We detect this with: ```bash 700 rmills@rmillsmbp:~/gnu/github/exiv2/0.27-maintenance $ exiv2 --version --verbose --grep nls exiv2 0.27.3 enable_nls=1 701 rmills@rmillsmbp:~/gnu/github/exiv2/0.27-maintenance $ ``` There are two additional issues. 1. Currently, you must do a `$ sudo make install` and run the installed version of exiv2. This is simply due to the code in src/exiv2.cpp main() which uses a EXV_ to find the localisation files. When running in the build directory, this should be `bin/../build/po` (I think. I've never tried this!). 2. Localisation is not supported on msvc builds. I don't think this will have any impact as you will detect enable_nls=0. I was surprised to discover that vcpkg builds with localisation and I have to investigate that: #1255 We won't delay the PR for this issue for msvc.
1.0
Test Localisation - Localisation is not tested in the test suite. It was never added because I didn't know how to use localisation. Furthermore, I found it almost impossible to build because of dependencies. Setting up the platform dependencies is now well documented in README.md, or you can use conan. I documented localisation in README.md for v0.27. You build localisation with the cmake option `-DEXIV2_ENABLE_NLS=On`. To use localisation from the command line: ```bash 680 rmills@rmillsmbp:~/gnu/github/exiv2/0.27-maintenance/build $ env LANG=fr_FR exiv2 exiv2: Une action doit être spécifié exiv2: Au moins un fichier est nécessaire Utilisation : exiv2 [ options ] [ action ] fichier ... Manipulation des métadonnées EXIF issues des images. 681 rmills@rmillsmbp:~/gnu/github/exiv2/0.27-maintenance/build $ ``` We should skip the test if NLS (natural language support) has not been built. We detect this with: ```bash 700 rmills@rmillsmbp:~/gnu/github/exiv2/0.27-maintenance $ exiv2 --version --verbose --grep nls exiv2 0.27.3 enable_nls=1 701 rmills@rmillsmbp:~/gnu/github/exiv2/0.27-maintenance $ ``` There are two additional issues. 1. Currently, you must do a `$ sudo make install` and run the installed version of exiv2. This is simply due to the code in src/exiv2.cpp main() which uses a EXV_ to find the localisation files. When running in the build directory, this should be `bin/../build/po` (I think. I've never tried this!). 2. Localisation is not supported on msvc builds. I don't think this will have any impact as you will detect enable_nls=0. I was surprised to discover that vcpkg builds with localisation and I have to investigate that: #1255 We won't delay the PR for this issue for msvc.
test
test localisation localisation is not tested in the test suite it was never added because i didn t know how to use localisation furthermore i found it almost impossible to build because of dependencies setting up the platform dependencies is now well documented in readme md or you can use conan i documented localisation in readme md for you build localisation with the cmake option enable nls on to use localisation from the command line bash rmills rmillsmbp gnu github maintenance build env lang fr fr une action doit être spécifié au moins un fichier est nécessaire utilisation fichier manipulation des métadonnées exif issues des images rmills rmillsmbp gnu github maintenance build we should skip the test if nls natural language support has not been built we detect this with bash rmills rmillsmbp gnu github maintenance version verbose grep nls enable nls rmills rmillsmbp gnu github maintenance there are two additional issues currently you must do a sudo make install and run the installed version of this is simply due to the code in src cpp main which uses a exv to find the localisation files when running in the build directory this should be bin build po i think i ve never tried this localisation is not supported on msvc builds i don t think this will have any impact as you will detect enable nls i was surprised to discover that vcpkg builds with localisation and i have to investigate that we won t delay the pr for this issue for msvc
1
62,809
14,656,621,613
IssuesEvent
2020-12-28 13:50:08
fu1771695yongxie/react-beautiful-dnd
https://api.github.com/repos/fu1771695yongxie/react-beautiful-dnd
opened
CVE-2020-11023 (Medium) detected in multiple libraries
security vulnerability
## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.1.min.js</b>, <b>jquery-1.9.0.min.js</b>, <b>jquery-1.11.3.js</b>, <b>jquery-3.1.0.js</b></p></summary> <p> <details><summary><b>jquery-2.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js</a></p> <p>Path to dependency file: react-beautiful-dnd/node_modules/lighthouse/lighthouse-cli/test/fixtures/dobetterweb/dbw_tester.html</p> <p>Path to vulnerable library: react-beautiful-dnd/node_modules/lighthouse/lighthouse-cli/test/fixtures/dobetterweb/dbw_tester.html</p> <p> Dependency Hierarchy: - :x: **jquery-2.1.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.9.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js</a></p> <p>Path to dependency file: react-beautiful-dnd/node_modules/es6-shim/test-sham/index.html</p> <p>Path to vulnerable library: react-beautiful-dnd/node_modules/es6-shim/test-sham/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.0.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.3.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.js</a></p> <p>Path to dependency file: react-beautiful-dnd/node_modules/es6-shim/test/native.html</p> <p>Path to vulnerable library: react-beautiful-dnd/node_modules/es6-shim/test/native.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.3.js** (Vulnerable Library) </details> <details><summary><b>jquery-3.1.0.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.0/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.0/jquery.js</a></p> <p>Path to dependency file: react-beautiful-dnd/node_modules/lighthouse/docs/recipes/gulp/public/index.html</p> <p>Path to vulnerable library: react-beautiful-dnd/node_modules/lighthouse/docs/recipes/gulp/public/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.1.0.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/react-beautiful-dnd/commit/1b8fdade31513b58c89ac3536c50f4e9f6976477">1b8fdade31513b58c89ac3536c50f4e9f6976477</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11023 (Medium) detected in multiple libraries - ## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.1.min.js</b>, <b>jquery-1.9.0.min.js</b>, <b>jquery-1.11.3.js</b>, <b>jquery-3.1.0.js</b></p></summary> <p> <details><summary><b>jquery-2.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js</a></p> <p>Path to dependency file: react-beautiful-dnd/node_modules/lighthouse/lighthouse-cli/test/fixtures/dobetterweb/dbw_tester.html</p> <p>Path to vulnerable library: react-beautiful-dnd/node_modules/lighthouse/lighthouse-cli/test/fixtures/dobetterweb/dbw_tester.html</p> <p> Dependency Hierarchy: - :x: **jquery-2.1.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.9.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.0/jquery.min.js</a></p> <p>Path to dependency file: react-beautiful-dnd/node_modules/es6-shim/test-sham/index.html</p> <p>Path to vulnerable library: react-beautiful-dnd/node_modules/es6-shim/test-sham/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.0.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.3.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.js</a></p> <p>Path to dependency file: react-beautiful-dnd/node_modules/es6-shim/test/native.html</p> <p>Path to vulnerable library: react-beautiful-dnd/node_modules/es6-shim/test/native.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.3.js** (Vulnerable Library) </details> <details><summary><b>jquery-3.1.0.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.0/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.0/jquery.js</a></p> <p>Path to dependency file: react-beautiful-dnd/node_modules/lighthouse/docs/recipes/gulp/public/index.html</p> <p>Path to vulnerable library: react-beautiful-dnd/node_modules/lighthouse/docs/recipes/gulp/public/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.1.0.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/react-beautiful-dnd/commit/1b8fdade31513b58c89ac3536c50f4e9f6976477">1b8fdade31513b58c89ac3536c50f4e9f6976477</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery js jquery js jquery min js javascript library for dom operations library home page a href path to dependency file react beautiful dnd node modules lighthouse lighthouse cli test fixtures dobetterweb dbw tester html path to vulnerable library react beautiful dnd node modules lighthouse lighthouse cli test fixtures dobetterweb dbw tester html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file react beautiful dnd node modules shim test sham index html path to vulnerable library react beautiful dnd node modules shim test sham index html dependency hierarchy x jquery min js vulnerable library jquery js javascript library for dom operations library home page a href path to dependency file react beautiful dnd node modules shim test native html path to vulnerable library react beautiful dnd node modules shim test native html dependency hierarchy x jquery js vulnerable library jquery js javascript library for dom operations library home page a href path to dependency file react beautiful dnd node modules lighthouse docs recipes gulp public index html path to vulnerable library react beautiful dnd node modules lighthouse docs recipes gulp public index html dependency hierarchy x jquery js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
73,172
7,328,776,574
IssuesEvent
2018-03-04 23:57:58
EyeSeeTea/dataset-configuration-blessed
https://api.github.com/repos/EyeSeeTea/dataset-configuration-blessed
closed
Slow loading/drawing of forms
enhancement help wanted testing
Implemented attribute "Group" on the output data elements now and it considerably slows down the loading/drawing of the form. Just the outputs now takes about 20 seconds to load on fast laptops. Any way we can look to speed this up? Haven't tested with outcomes yet but it's likely not to help.. Test on: ![slow loading test.PNG](https://waffleio-direct-uploads-production.s3.amazonaws.com/uploads/590c634ad13d7b00206a7721/125516c66e82c728ace21e0d46fb989534c28fa68eedb64da31de4913a52653f6b14b1247a80e03cfd183954570c4ba4100e584dada88f2bf7b63c72d95e0fbcd5695e66da2f53fde6d640e258f075c95d8fa4bd8d46c1ed3c20e38989473c18041c80ed10.PNG)
1.0
Slow loading/drawing of forms - Implemented attribute "Group" on the output data elements now and it considerably slows down the loading/drawing of the form. Just the outputs now takes about 20 seconds to load on fast laptops. Any way we can look to speed this up? Haven't tested with outcomes yet but it's likely not to help.. Test on: ![slow loading test.PNG](https://waffleio-direct-uploads-production.s3.amazonaws.com/uploads/590c634ad13d7b00206a7721/125516c66e82c728ace21e0d46fb989534c28fa68eedb64da31de4913a52653f6b14b1247a80e03cfd183954570c4ba4100e584dada88f2bf7b63c72d95e0fbcd5695e66da2f53fde6d640e258f075c95d8fa4bd8d46c1ed3c20e38989473c18041c80ed10.PNG)
test
slow loading drawing of forms implemented attribute group on the output data elements now and it considerably slows down the loading drawing of the form just the outputs now takes about seconds to load on fast laptops any way we can look to speed this up haven t tested with outcomes yet but it s likely not to help test on
1
190,117
6,808,917,336
IssuesEvent
2017-11-04 10:30:39
openml/OpenML
https://api.github.com/repos/openml/OpenML
opened
API returns nothing. Stochastic bug?
bug priority: highest
Hm this only seems to appear sometimes. The listing calls give me the right results but sometimes I a completely empty page is returned. For example: If I open this url https://www.openml.org/api/v1/json/data/list/limit/100/status/active multiple times (e.g. 10 times), it gives me a completely empty page with a 50% chance. This seems to happen for almost all listing calls? And causes many failing unit tests.
1.0
API returns nothing. Stochastic bug? - Hm this only seems to appear sometimes. The listing calls give me the right results but sometimes I a completely empty page is returned. For example: If I open this url https://www.openml.org/api/v1/json/data/list/limit/100/status/active multiple times (e.g. 10 times), it gives me a completely empty page with a 50% chance. This seems to happen for almost all listing calls? And causes many failing unit tests.
non_test
api returns nothing stochastic bug hm this only seems to appear sometimes the listing calls give me the right results but sometimes i a completely empty page is returned for example if i open this url multiple times e g times it gives me a completely empty page with a chance this seems to happen for almost all listing calls and causes many failing unit tests
0
353,173
25,104,923,163
IssuesEvent
2022-11-08 15:59:29
ICEI-PUC-Minas-PMV-ADS/pmv-ads-2022-2-e1-proj-web-t2-seguranca_informacao
https://api.github.com/repos/ICEI-PUC-Minas-PMV-ADS/pmv-ads-2022-2-e1-proj-web-t2-seguranca_informacao
closed
Segunda Etapa - 19/09/22 a 09/10/22
documentation
Projeto da Solução - [x] Metodologia; - [x] Projeto de Interface; - [x] Arquitetura da Solução.
1.0
Segunda Etapa - 19/09/22 a 09/10/22 - Projeto da Solução - [x] Metodologia; - [x] Projeto de Interface; - [x] Arquitetura da Solução.
non_test
segunda etapa a projeto da solução metodologia projeto de interface arquitetura da solução
0
564,085
16,707,214,656
IssuesEvent
2021-06-09 11:28:23
googleapis/google-api-ruby-client
https://api.github.com/repos/googleapis/google-api-ruby-client
closed
Synthesis failed for sheets-v4
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate sheets-v4. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the sheets-v4 API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` 2021-06-08 03:33:21,077 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client 2021-06-08 03:33:21,874 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2021-06-08 03:33:21,876 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2021-06-08 03:33:21,878 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2021-06-08 03:33:21,881 autosynth [DEBUG] > Running: git config push.default simple 2021-06-08 03:33:21,883 autosynth [DEBUG] > Running: git branch -f autosynth-sheets-v4 2021-06-08 03:33:21,886 autosynth [DEBUG] > Running: git checkout autosynth-sheets-v4 Switched to branch 'autosynth-sheets-v4' 2021-06-08 03:33:22,098 autosynth [INFO] > Running synthtool 2021-06-08 03:33:22,098 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-sheets_v4/synth.metadata', 'synth.py', '--'] 2021-06-08 03:33:22,098 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/sheets/v4/sponge_log.log 2021-06-08 03:33:22,100 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-sheets_v4/synth.metadata synth.py -- sheets v4 2021-06-08 03:33:22,294 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py. On branch autosynth-sheets-v4 nothing to commit, working tree clean 2021-06-08 03:33:22,353 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth sheets v4 DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth sheets v4 git clean -df bundle install Don't run Bundler as root. Bundler can ask for sudo if it is needed, and installing your bundle as root will break this application for all non-root users on this machine. Fetching gem metadata from https://rubygems.org/.. Retrying dependency api due to error (2/4): Bundler::HTTPError Net::HTTPServiceUnavailable: <?xml version="1.0" encoding="utf-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html> <head> <title>503 Service Unavailable</title> </head> <body> <h1>Error 503 Service Unavailable</h1> <p>Service Unavailable</p> <h3>Guru Mediation:</h3> <p>Details: cache-sea4456-SEA 1623148404 987424819</p> <hr> <p>Varnish cache server</p> </body> </html> Retrying dependency api due to error (3/4): Bundler::HTTPError Net::HTTPServiceUnavailable: <?xml version="1.0" encoding="utf-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html> <head> <title>503 Service Unavailable</title> </head> <body> <h1>Error 503 Service Unavailable</h1> <p>Service Unavailable</p> <h3>Guru Mediation:</h3> <p>Details: cache-sea4472-SEA 1623148404 1502646647</p> <hr> <p>Varnish cache server</p> </body> </html> Retrying dependency api due to error (4/4): Bundler::HTTPError Net::HTTPBadGateway: connection failure Retrying fetcher due to error (2/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Gateway Error 502 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (3/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (4/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Net::HTTPServiceUnavailable: <?xml version="1.0" encoding="utf-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html> <head> <title>503 Service Unavailable</title> </head> <body> <h1>Error 503 Service Unavailable</h1> <p>Service Unavailable</p> <h3>Guru Mediation:</h3> <p>Details: cache-sea4422-SEA 1623148412 860850770</p> <hr> <p>Varnish cache server</p> </body> </html> chown -R 1000:1000 /workspace/generated 2021-06-08 03:33:31,680 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth sheets v4: None ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth sheets v4: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module> shell.run(command, hide_output=False) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'sheets', 'v4']' returned non-zero exit status 1. 2021-06-08 03:33:31,705 autosynth [ERROR] > Synthesis failed 2021-06-08 03:33:31,706 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-sheets_v4/synth.metadata', 'synth.py', '--', 'sheets', 'v4']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
1.0
Synthesis failed for sheets-v4 - Hello! Autosynth couldn't regenerate sheets-v4. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the sheets-v4 API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` 2021-06-08 03:33:21,077 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client 2021-06-08 03:33:21,874 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2021-06-08 03:33:21,876 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2021-06-08 03:33:21,878 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2021-06-08 03:33:21,881 autosynth [DEBUG] > Running: git config push.default simple 2021-06-08 03:33:21,883 autosynth [DEBUG] > Running: git branch -f autosynth-sheets-v4 2021-06-08 03:33:21,886 autosynth [DEBUG] > Running: git checkout autosynth-sheets-v4 Switched to branch 'autosynth-sheets-v4' 2021-06-08 03:33:22,098 autosynth [INFO] > Running synthtool 2021-06-08 03:33:22,098 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-sheets_v4/synth.metadata', 'synth.py', '--'] 2021-06-08 03:33:22,098 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/sheets/v4/sponge_log.log 2021-06-08 03:33:22,100 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-sheets_v4/synth.metadata synth.py -- sheets v4 2021-06-08 03:33:22,294 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py. On branch autosynth-sheets-v4 nothing to commit, working tree clean 2021-06-08 03:33:22,353 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth sheets v4 DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth sheets v4 git clean -df bundle install Don't run Bundler as root. Bundler can ask for sudo if it is needed, and installing your bundle as root will break this application for all non-root users on this machine. Fetching gem metadata from https://rubygems.org/.. Retrying dependency api due to error (2/4): Bundler::HTTPError Net::HTTPServiceUnavailable: <?xml version="1.0" encoding="utf-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html> <head> <title>503 Service Unavailable</title> </head> <body> <h1>Error 503 Service Unavailable</h1> <p>Service Unavailable</p> <h3>Guru Mediation:</h3> <p>Details: cache-sea4456-SEA 1623148404 987424819</p> <hr> <p>Varnish cache server</p> </body> </html> Retrying dependency api due to error (3/4): Bundler::HTTPError Net::HTTPServiceUnavailable: <?xml version="1.0" encoding="utf-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html> <head> <title>503 Service Unavailable</title> </head> <body> <h1>Error 503 Service Unavailable</h1> <p>Service Unavailable</p> <h3>Guru Mediation:</h3> <p>Details: cache-sea4472-SEA 1623148404 1502646647</p> <hr> <p>Varnish cache server</p> </body> </html> Retrying dependency api due to error (4/4): Bundler::HTTPError Net::HTTPBadGateway: connection failure Retrying fetcher due to error (2/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Gateway Error 502 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (3/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (4/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Net::HTTPServiceUnavailable: <?xml version="1.0" encoding="utf-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html> <head> <title>503 Service Unavailable</title> </head> <body> <h1>Error 503 Service Unavailable</h1> <p>Service Unavailable</p> <h3>Guru Mediation:</h3> <p>Details: cache-sea4422-SEA 1623148412 860850770</p> <hr> <p>Varnish cache server</p> </body> </html> chown -R 1000:1000 /workspace/generated 2021-06-08 03:33:31,680 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth sheets v4: None ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth sheets v4: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module> shell.run(command, hide_output=False) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'sheets', 'v4']' returned non-zero exit status 1. 2021-06-08 03:33:31,705 autosynth [ERROR] > Synthesis failed 2021-06-08 03:33:31,706 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-sheets_v4/synth.metadata', 'synth.py', '--', 'sheets', 'v4']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
non_test
synthesis failed for sheets hello autosynth couldn t regenerate sheets broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the sheets api and the library grows stale see for trouble shooting tips here s the output from running synth py autosynth logs will be written to tmpfs src logs google api ruby client autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth sheets autosynth running git checkout autosynth sheets switched to branch autosynth sheets autosynth running synthtool autosynth autosynth log file path tmpfs src logs google api ruby client sheets sponge log log autosynth running tmpfs src github synthtool env bin m synthtool metadata generated google apis sheets synth metadata synth py sheets synthtool executing home kbuilder cache synthtool google api ruby client synth py on branch autosynth sheets nothing to commit working tree clean synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth sheets debug synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth sheets git clean df bundle install don t run bundler as root bundler can ask for sudo if it is needed and installing your bundle as root will break this application for all non root users on this machine fetching gem metadata from retrying dependency api due to error bundler httperror net httpserviceunavailable doctype html public dtd xhtml strict en service unavailable error service unavailable service unavailable guru mediation details cache sea varnish cache server retrying dependency api due to error bundler httperror net httpserviceunavailable doctype html public dtd xhtml strict en service unavailable error service unavailable service unavailable guru mediation details cache sea varnish cache server retrying dependency api due to error bundler httperror net httpbadgateway connection failure retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response gateway error retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable net httpserviceunavailable doctype html public dtd xhtml strict en service unavailable error service unavailable service unavailable guru mediation details cache sea varnish cache server chown r workspace generated synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth sheets none error synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth sheets none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google api ruby client synth py line in shell run command hide output false file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
113,263
9,634,123,912
IssuesEvent
2019-05-15 20:25:15
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
opened
Investigate security test flakiness
Security Test
The test `test_security_alerts` is behaving flaky on ADO. Possibly related to #9385. Occasional ADO stack trace: ``` ====================================================================== FAIL: test_security_alerts (azure.cli.command_modules.security.tests.latest.test_alerts_scenario.SecurityCenterAlertsTests) ---------------------------------------------------------------------- Traceback (most recent call last): File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/base.py", line 251, in _in_process_execute self.exit_code = cli_ctx.invoke(shlex.split(command), out_file=stdout_buf) or 0 File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/knack/cli.py", line 218, in invoke exit_code = self.exception_handler(ex) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/core/__init__.py", line 98, in exception_handler return handle_exception(ex) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/patches.py", line 33, in _handle_main_exception raise ex File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/knack/cli.py", line 206, in invoke cmd_result = self.invocation.execute(args) vcr.errors.CannotOverwriteExistingCassetteException: No match for the request (<Request (GET) https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/cli_role_assign000001/providers/Microsoft.Security/alerts?api-version=2015-06-01-preview>) was found. Can't overwrite existing cassette ('/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/command_modules/security/tests/latest/recordings/test_security_alerts.yaml') in your current record mode ('once'). During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure_devtools/scenario_tests/decorators.py", line 40, in _preparer_wrapper fn(test_class_instance, **kwargs) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/command_modules/security/tests/latest/test_alerts_scenario.py", line 16, in test_security_alerts alerts = self.cmd('az security alert list').get_output_in_json() File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/base.py", line 161, in cmd return execute(self.cli_ctx, command, expect_failure=expect_failure).assert_with_checks(checks) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/base.py", line 200, in __init__ self._in_process_execute(cli_ctx, command, expect_failure=expect_failure) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/base.py", line 256, in _in_process_execute raise AssertionError(ex) AssertionError: No match for the request (<Request (GET) https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/cli_role_assign000001/providers/Microsoft.Security/alerts?api-version=2015-06-01-preview>) was found. Can't overwrite existing cassette ('/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/command_modules/security/tests/latest/recordings/test_security_alerts.yaml') in your current record mode ('once'). ``` When run locally, these tests pass consistently.
1.0
Investigate security test flakiness - The test `test_security_alerts` is behaving flaky on ADO. Possibly related to #9385. Occasional ADO stack trace: ``` ====================================================================== FAIL: test_security_alerts (azure.cli.command_modules.security.tests.latest.test_alerts_scenario.SecurityCenterAlertsTests) ---------------------------------------------------------------------- Traceback (most recent call last): File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/base.py", line 251, in _in_process_execute self.exit_code = cli_ctx.invoke(shlex.split(command), out_file=stdout_buf) or 0 File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/knack/cli.py", line 218, in invoke exit_code = self.exception_handler(ex) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/core/__init__.py", line 98, in exception_handler return handle_exception(ex) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/patches.py", line 33, in _handle_main_exception raise ex File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/knack/cli.py", line 206, in invoke cmd_result = self.invocation.execute(args) vcr.errors.CannotOverwriteExistingCassetteException: No match for the request (<Request (GET) https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/cli_role_assign000001/providers/Microsoft.Security/alerts?api-version=2015-06-01-preview>) was found. Can't overwrite existing cassette ('/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/command_modules/security/tests/latest/recordings/test_security_alerts.yaml') in your current record mode ('once'). During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure_devtools/scenario_tests/decorators.py", line 40, in _preparer_wrapper fn(test_class_instance, **kwargs) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/command_modules/security/tests/latest/test_alerts_scenario.py", line 16, in test_security_alerts alerts = self.cmd('az security alert list').get_output_in_json() File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/base.py", line 161, in cmd return execute(self.cli_ctx, command, expect_failure=expect_failure).assert_with_checks(checks) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/base.py", line 200, in __init__ self._in_process_execute(cli_ctx, command, expect_failure=expect_failure) File "/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/testsdk/base.py", line 256, in _in_process_execute raise AssertionError(ex) AssertionError: No match for the request (<Request (GET) https://management.azure.com/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/cli_role_assign000001/providers/Microsoft.Security/alerts?api-version=2015-06-01-preview>) was found. Can't overwrite existing cassette ('/opt/hostedtoolcache/Python/3.6.8/x64/lib/python3.6/site-packages/azure/cli/command_modules/security/tests/latest/recordings/test_security_alerts.yaml') in your current record mode ('once'). ``` When run locally, these tests pass consistently.
test
investigate security test flakiness the test test security alerts is behaving flaky on ado possibly related to occasional ado stack trace fail test security alerts azure cli command modules security tests latest test alerts scenario securitycenteralertstests traceback most recent call last file opt hostedtoolcache python lib site packages azure cli testsdk base py line in in process execute self exit code cli ctx invoke shlex split command out file stdout buf or file opt hostedtoolcache python lib site packages knack cli py line in invoke exit code self exception handler ex file opt hostedtoolcache python lib site packages azure cli core init py line in exception handler return handle exception ex file opt hostedtoolcache python lib site packages azure cli testsdk patches py line in handle main exception raise ex file opt hostedtoolcache python lib site packages knack cli py line in invoke cmd result self invocation execute args vcr errors cannotoverwriteexistingcassetteexception no match for the request request get was found can t overwrite existing cassette opt hostedtoolcache python lib site packages azure cli command modules security tests latest recordings test security alerts yaml in your current record mode once during handling of the above exception another exception occurred traceback most recent call last file opt hostedtoolcache python lib site packages azure devtools scenario tests decorators py line in preparer wrapper fn test class instance kwargs file opt hostedtoolcache python lib site packages azure cli command modules security tests latest test alerts scenario py line in test security alerts alerts self cmd az security alert list get output in json file opt hostedtoolcache python lib site packages azure cli testsdk base py line in cmd return execute self cli ctx command expect failure expect failure assert with checks checks file opt hostedtoolcache python lib site packages azure cli testsdk base py line in init self in process execute cli ctx command expect failure expect failure file opt hostedtoolcache python lib site packages azure cli testsdk base py line in in process execute raise assertionerror ex assertionerror no match for the request request get was found can t overwrite existing cassette opt hostedtoolcache python lib site packages azure cli command modules security tests latest recordings test security alerts yaml in your current record mode once when run locally these tests pass consistently
1
305,851
26,416,392,927
IssuesEvent
2023-01-13 16:19:15
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix solving_equations_and_inverting_matrices.test_numpy_pinv
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3911376725/jobs/6684697686" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3911376725/jobs/6684697686" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3911376725/jobs/6684697686" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3911376725/jobs/6684697686" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_linear_algebra/test_solving_equations_and_inverting_matrices.py::test_numpy_pinv[cpu-ivy.functional.backends.tensorflow-False-False]</summary> 2023-01-13T13:07:14.4751595Z E AssertionError: [[-1.0058957e+00 3.2188606e+01] 2023-01-13T13:07:14.4751972Z E [ 5.8956528e-03 -3.2188606e+01]] != [[-1.0058939e+00 3.2188606e+01] 2023-01-13T13:07:14.4752282Z E [ 5.8939760e-03 -3.2188606e+01]] 2023-01-13T13:07:14.4752533Z E Falsifying example: test_numpy_pinv( 2023-01-13T13:07:14.4753220Z E dtype_and_x=(['float32'], [array([[-1.0000000e+00, -1.0000000e+00], 2023-01-13T13:07:14.4753618Z E [-1.8310547e-04, -3.1250000e-02]], dtype=float32)]), 2023-01-13T13:07:14.4753887Z E native_array=[False], 2023-01-13T13:07:14.4754119Z E as_variable=[False], 2023-01-13T13:07:14.4754353Z E num_positional_args=0, 2023-01-13T13:07:14.4754737Z E fn_tree='ivy.functional.frontends.numpy.linalg.pinv', 2023-01-13T13:07:14.4755067Z E frontend='numpy', 2023-01-13T13:07:14.4755321Z E on_device='cpu', 2023-01-13T13:07:14.4755520Z E ) 2023-01-13T13:07:14.4755691Z E 2023-01-13T13:07:14.4756268Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BgBEIGMGaA0ijsAweR2A4McAAAMI8Ciw==') as a decorator on your test case </details>
1.0
Fix solving_equations_and_inverting_matrices.test_numpy_pinv - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3911376725/jobs/6684697686" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3911376725/jobs/6684697686" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3911376725/jobs/6684697686" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3911376725/jobs/6684697686" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_linear_algebra/test_solving_equations_and_inverting_matrices.py::test_numpy_pinv[cpu-ivy.functional.backends.tensorflow-False-False]</summary> 2023-01-13T13:07:14.4751595Z E AssertionError: [[-1.0058957e+00 3.2188606e+01] 2023-01-13T13:07:14.4751972Z E [ 5.8956528e-03 -3.2188606e+01]] != [[-1.0058939e+00 3.2188606e+01] 2023-01-13T13:07:14.4752282Z E [ 5.8939760e-03 -3.2188606e+01]] 2023-01-13T13:07:14.4752533Z E Falsifying example: test_numpy_pinv( 2023-01-13T13:07:14.4753220Z E dtype_and_x=(['float32'], [array([[-1.0000000e+00, -1.0000000e+00], 2023-01-13T13:07:14.4753618Z E [-1.8310547e-04, -3.1250000e-02]], dtype=float32)]), 2023-01-13T13:07:14.4753887Z E native_array=[False], 2023-01-13T13:07:14.4754119Z E as_variable=[False], 2023-01-13T13:07:14.4754353Z E num_positional_args=0, 2023-01-13T13:07:14.4754737Z E fn_tree='ivy.functional.frontends.numpy.linalg.pinv', 2023-01-13T13:07:14.4755067Z E frontend='numpy', 2023-01-13T13:07:14.4755321Z E on_device='cpu', 2023-01-13T13:07:14.4755520Z E ) 2023-01-13T13:07:14.4755691Z E 2023-01-13T13:07:14.4756268Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2BgBEIGMGaA0ijsAweR2A4McAAAMI8Ciw==') as a decorator on your test case </details>
test
fix solving equations and inverting matrices test numpy pinv tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test numpy test linear algebra test solving equations and inverting matrices py test numpy pinv e assertionerror e e e falsifying example test numpy pinv e dtype and x e dtype e native array e as variable e num positional args e fn tree ivy functional frontends numpy linalg pinv e frontend numpy e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
800,581
28,371,847,534
IssuesEvent
2023-04-12 17:35:25
project-pareto/project-pareto
https://api.github.com/repos/project-pareto/project-pareto
closed
Missing step in installation guide on GitHub main page
Priority:Normal
The GitHub main page installation guide for developers seems to be missing a step. Once everything else is installed, users will need to run `pip install -e .` to actually install the PARETO module and have it available from outside its own directory. This instruction is included in the documentation, but not on the main page installation guide.
1.0
Missing step in installation guide on GitHub main page - The GitHub main page installation guide for developers seems to be missing a step. Once everything else is installed, users will need to run `pip install -e .` to actually install the PARETO module and have it available from outside its own directory. This instruction is included in the documentation, but not on the main page installation guide.
non_test
missing step in installation guide on github main page the github main page installation guide for developers seems to be missing a step once everything else is installed users will need to run pip install e to actually install the pareto module and have it available from outside its own directory this instruction is included in the documentation but not on the main page installation guide
0
143,554
11,569,448,903
IssuesEvent
2020-02-20 17:34:18
recp/cglm
https://api.github.com/repos/recp/cglm
closed
cglm rotate differs from glm rotate
feedback wanted needs tests question
GLM: ``` glm::mat4 trans = glm::mat4(1.0f); trans = glm::rotate(trans, glm::radians(90.0f), glm::vec3(0.0, 0.0, 1.0)); ``` CGLM: ``` mat4 trans = GLM_MAT4_IDENTITY_INIT; glm_rotate(trans, glm_rad(90.0f), (vec3){0.0f, 0.0f, 1.0f}); ``` Output of GLM: `mat4x4((-0.000000, 1.000000, 0.000000, 0.000000), (-1.000000, -0.000000, 0.000000, 0.000000), (0.000000, 0.000000, 1.000000, 0.000000), (0.000000, 0.000000, 0.000000, 1.000000))` Output of CGLM: ``` Matrix (float4x4): |-0.0000 -1.0000 0.0000 0.0000| |1.0000 -0.0000 0.0000 0.0000| |0.0000 0.0000 1.0000 0.0000| |0.0000 0.0000 0.0000 1.0000| ``` As you can see (1,2) and (2,1) are different. Did I used it wrong?
1.0
cglm rotate differs from glm rotate - GLM: ``` glm::mat4 trans = glm::mat4(1.0f); trans = glm::rotate(trans, glm::radians(90.0f), glm::vec3(0.0, 0.0, 1.0)); ``` CGLM: ``` mat4 trans = GLM_MAT4_IDENTITY_INIT; glm_rotate(trans, glm_rad(90.0f), (vec3){0.0f, 0.0f, 1.0f}); ``` Output of GLM: `mat4x4((-0.000000, 1.000000, 0.000000, 0.000000), (-1.000000, -0.000000, 0.000000, 0.000000), (0.000000, 0.000000, 1.000000, 0.000000), (0.000000, 0.000000, 0.000000, 1.000000))` Output of CGLM: ``` Matrix (float4x4): |-0.0000 -1.0000 0.0000 0.0000| |1.0000 -0.0000 0.0000 0.0000| |0.0000 0.0000 1.0000 0.0000| |0.0000 0.0000 0.0000 1.0000| ``` As you can see (1,2) and (2,1) are different. Did I used it wrong?
test
cglm rotate differs from glm rotate glm glm trans glm trans glm rotate trans glm radians glm cglm trans glm identity init glm rotate trans glm rad output of glm output of cglm matrix as you can see and are different did i used it wrong
1
275,006
23,887,971,422
IssuesEvent
2022-09-08 09:11:51
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: liquibase failed
C-test-failure O-robot O-roachtest release-blocker branch-release-22.2
roachtest.liquibase [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6383281?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6383281?buildTab=artifacts#/liquibase) on release-22.2 @ [f8e04f32b84a9727c9e813d0096bde2de64d5675](https://github.com/cockroachdb/cockroach/commits/f8e04f32b84a9727c9e813d0096bde2de64d5675): ``` test artifacts and logs in: /artifacts/liquibase/run_1 orm_helpers.go:191,orm_helpers.go:117,java_helpers.go:220,liquibase.go:123,liquibase.go:140,test_runner.go:908: Tests run on Cockroach v22.2.0-alpha.2-283-gf8e04f32b8 Tests run against liquibase 1790ddef2d0339c5c96839ac60ac424c130dadd8 47 Total Tests Run 38 tests passed 9 tests failed 17 tests skipped 0 tests ignored 0 tests passed unexpectedly 6 tests failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- FAIL: liquibase.harness.change.ChangeObjectTests.apply renameSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply dropSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply addDefaultValueSequenceNext against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply alterSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply addAutoIncrement against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply createSequence against cockroachdb 20.2 - unknown (unexpected) For a full summary look at the liquibase artifacts An updated blocklist (liquibaseBlocklist22_2) is available in the artifacts' liquibase log ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #83313 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.2] - #83050 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-master] - #83048 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1] </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*liquibase.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: liquibase failed - roachtest.liquibase [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6383281?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6383281?buildTab=artifacts#/liquibase) on release-22.2 @ [f8e04f32b84a9727c9e813d0096bde2de64d5675](https://github.com/cockroachdb/cockroach/commits/f8e04f32b84a9727c9e813d0096bde2de64d5675): ``` test artifacts and logs in: /artifacts/liquibase/run_1 orm_helpers.go:191,orm_helpers.go:117,java_helpers.go:220,liquibase.go:123,liquibase.go:140,test_runner.go:908: Tests run on Cockroach v22.2.0-alpha.2-283-gf8e04f32b8 Tests run against liquibase 1790ddef2d0339c5c96839ac60ac424c130dadd8 47 Total Tests Run 38 tests passed 9 tests failed 17 tests skipped 0 tests ignored 0 tests passed unexpectedly 6 tests failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- FAIL: liquibase.harness.change.ChangeObjectTests.apply renameSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply dropSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply addDefaultValueSequenceNext against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply alterSequence against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply addAutoIncrement against cockroachdb 20.2 - unknown (unexpected) --- FAIL: liquibase.harness.change.ChangeObjectTests.apply createSequence against cockroachdb 20.2 - unknown (unexpected) For a full summary look at the liquibase artifacts An updated blocklist (liquibaseBlocklist22_2) is available in the artifacts' liquibase log ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #83313 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.2] - #83050 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-master] - #83048 roachtest: liquibase failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1] </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*liquibase.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest liquibase failed roachtest liquibase with on release test artifacts and logs in artifacts liquibase run orm helpers go orm helpers go java helpers go liquibase go liquibase go test runner go tests run on cockroach alpha tests run against liquibase total tests run tests passed tests failed tests skipped tests ignored tests passed unexpectedly tests failed unexpectedly tests expected failed but skipped tests expected failed but not run fail liquibase harness change changeobjecttests apply renamesequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply dropsequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply adddefaultvaluesequencenext against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply altersequence against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply addautoincrement against cockroachdb unknown unexpected fail liquibase harness change changeobjecttests apply createsequence against cockroachdb unknown unexpected for a full summary look at the liquibase artifacts an updated blocklist is available in the artifacts liquibase log parameters roachtest cloud gce roachtest cpu roachtest ssd help see see same failure on other branches roachtest liquibase failed roachtest liquibase failed roachtest liquibase failed cc cockroachdb sql experience
1
222,784
17,483,429,494
IssuesEvent
2021-08-09 07:45:51
Open-CAS/open-cas-linux
https://api.github.com/repos/Open-CAS/open-cas-linux
closed
LBAs outside of a LBA range configured in IO class are cached
P2-medium tests
## Description <!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug --> test_ioclass_lba ## Expected Behavior <!--- Tell us what should happen --> Only LBAs inside the configured range should be cached ## Actual Behavior <!--- Tell us what happens instead --> LBAs outside the configured range are cached ## Steps to Reproduce <!--- Provide a repeatable set of steps to reproduce this bug --> 1. Start cache and add core 2. Configure IO class ``` IO class id,IO class name,Eviction priority,Allocation 0,unclassified,255,0.00 1,lba:ge:56&lba:le:200&done,1,1.00 ``` 3. Write to LBAs inside and outside of configured LBA range ```dd if=/dev/zero of=/dev/cas1-1 count=1 bs=512 seek=72``` ## Logs <!--- Error logs in any form are really helpful --> ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * OpenCAS version (commit hash or tag): 21.06.0.0534.master b57a8bf19575cc3d2707f5953ad6f9ac9ab59fd9 * Operating System: Centos 7.9 * Kernel version: 5.10 * Cache device type (NAND/Optane/other): Optane * Core device type (HDD/SSD/other): TLC NVMe
1.0
LBAs outside of a LBA range configured in IO class are cached - ## Description <!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug --> test_ioclass_lba ## Expected Behavior <!--- Tell us what should happen --> Only LBAs inside the configured range should be cached ## Actual Behavior <!--- Tell us what happens instead --> LBAs outside the configured range are cached ## Steps to Reproduce <!--- Provide a repeatable set of steps to reproduce this bug --> 1. Start cache and add core 2. Configure IO class ``` IO class id,IO class name,Eviction priority,Allocation 0,unclassified,255,0.00 1,lba:ge:56&lba:le:200&done,1,1.00 ``` 3. Write to LBAs inside and outside of configured LBA range ```dd if=/dev/zero of=/dev/cas1-1 count=1 bs=512 seek=72``` ## Logs <!--- Error logs in any form are really helpful --> ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * OpenCAS version (commit hash or tag): 21.06.0.0534.master b57a8bf19575cc3d2707f5953ad6f9ac9ab59fd9 * Operating System: Centos 7.9 * Kernel version: 5.10 * Cache device type (NAND/Optane/other): Optane * Core device type (HDD/SSD/other): TLC NVMe
test
lbas outside of a lba range configured in io class are cached description test ioclass lba expected behavior only lbas inside the configured range should be cached actual behavior lbas outside the configured range are cached steps to reproduce start cache and add core configure io class io class id io class name eviction priority allocation unclassified lba ge lba le done write to lbas inside and outside of configured lba range dd if dev zero of dev count bs seek logs your environment opencas version commit hash or tag master operating system centos kernel version cache device type nand optane other optane core device type hdd ssd other tlc nvme
1
95,245
8,553,519,186
IssuesEvent
2018-11-08 01:15:25
zfsonlinux/zfs
https://api.github.com/repos/zfsonlinux/zfs
closed
ztest: dmu_read() ASSERT on EIO
Type: Test Suite
Version Information ================================================================= Distribution Name Centos 7.2 Distribution Version 7.2 Linux Kernel 3.10.0-327.36.3.el7.x86_64 Architecture x86_64 ZFS Version master Dec 1st RPM/modinfo say: 0.7.0-rc2 SPL Version master Dec 1st RPM/modinfo say: 0.7.0-rc2 Git Build Data Revision: 5f20c145df3ca45f86cd1892137edb117d4c8a82 refs/remotes/origin/master Built Branches refs/remotes/origin/master: Build #183 of Revision 5f20c145df3ca45f86cd1892137edb117d4c8a82 (refs/remotes/origin/master) Problem ================================================================= Running zloop we hit an assert: dmu_read(zd->zd_os, object, offset, blocksize, data, 1) == 0 (0x5 == 0x0) ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] Reproduce ================================================================= /sbin/ztest -VVVVV -m 2 -r 0 -R 1 -v 2 -a 9 -T 7 -P 12 -s 128m -f /var/tmp /sbin/ztest -VVVVV -m 2 -r 0 -R 1 -v 2 -a 9 -T 72 -P 25 -s 128m -f /var/tmp Log detail ================================================================= ztest.out ``` starting main threads... 0.00 sec in ztest_verify_dnode_bt 0.01 sec in ztest_dmu_object_alloc_free 0.00 sec in ztest_split_pool 0.00 sec in ztest_zil_commit 0.00 sec in ztest_zil_commit 0.00 sec in ztest_dmu_commit_callbacks 0.01 sec in ztest_zap 0.11 sec in ztest_vdev_attach_detach 0.11 sec in ztest_fault_inject 0.13 sec in ztest_spa_prop_get_set 0.11 sec in ztest_dmu_read_write_zcopy 0.20 sec in ztest_vdev_attach_detach 0.21 sec in ztest_zil_commit 0.23 sec in ztest_spa_prop_get_set 0.22 sec in ztest_zap_parallel 0.00 sec in ztest_dmu_commit_callbacks 0.23 sec in ztest_fault_inject 0.07 sec in ztest_dmu_read_write 0.02 sec in ztest_dmu_read_write 0.95 sec in ztest_zil_remount 1.05 sec in ztest_dmu_object_alloc_free 0.00 sec in ztest_dmu_commit_callbacks 1.13 sec in ztest_spa_create_destroy 0.00 sec in ztest_zil_commit Setting dataset ztest/temp_0 to sync always 1.17 sec in ztest_dmu_object_alloc_free 1.32 sec in ztest_dmu_snapshot_create_destroy 1.52 sec in ztest_dmu_objset_create_destroy 0.01 sec in ztest_zap 0.00 sec in ztest_split_pool Setting dataset ztest/temp_22 to sync always 1.91 sec in ztest_fzap 0.85 sec in ztest_dmu_read_write_zcopy 0.01 sec in ztest_dmu_read_write 0.01 sec in ztest_fault_inject Setting dataset ztest/temp_20 to sync always 1.62 sec in ztest_dmu_object_alloc_free 3.00 sec in ztest_dmu_read_write_zcopy 1.21 sec in ztest_spa_create_destroy 3.24 sec in ztest_zap_parallel 1.50 sec in ztest_dmu_snapshot_create_destroy 0.31 sec in ztest_zap_parallel 0.79 sec in ztest_spa_create_destroy 3.79 sec in ztest_dsl_prop_get_set 3.79 sec in ztest_dsl_prop_get_set 3.80 sec in ztest_dsl_prop_get_set 0.00 sec in ztest_dmu_commit_callbacks 3.79 sec in ztest_dsl_prop_get_set 0.00 sec in ztest_split_pool 0.00 sec in ztest_dmu_read_write 0.00 sec in ztest_dmu_object_alloc_free 0.69 sec in ztest_spa_create_destroy 4.48 sec in ztest_dmu_objset_create_destroy 0.69 sec in ztest_dmu_read_write_zcopy 0.00 sec in ztest_zap 0.00 sec in ztest_dmu_read_write 3.56 sec in ztest_dsl_prop_get_set 1.01 sec in ztest_dmu_snapshot_create_destroy 1.46 sec in ztest_dmu_snapshot_create_destroy 0.00 sec in ztest_zil_commit 3.19 sec in ztest_dsl_prop_get_set 1.42 sec in ztest_zil_remount 4.12 sec in ztest_dmu_object_alloc_free 1.27 sec in ztest_zap 1.43 sec in ztest_dmu_read_write 1.53 sec in ztest_dmu_read_write_zcopy 5.87 sec in ztest_dmu_objset_create_destroy 6.09 sec in ztest_dmu_snapshot_hold 0.00 sec in ztest_zil_commit Unable to expand vdev, state 4, error 0 Could not expand LUN because the vdev configuration changed. 0.91 sec in ztest_vdev_LUN_growth 2.61 sec in ztest_dsl_prop_get_set 2.98 sec in ztest_dsl_prop_get_set 0.33 sec in ztest_zap 0.00 sec in ztest_split_pool 0.00 sec in ztest_dmu_commit_callbacks 0.32 sec in ztest_zil_commit 0.00 sec in ztest_zil_commit 0.17 sec in ztest_zil_commit 6.73 sec in ztest_vdev_attach_detach 13.17 sec in ztest_dmu_snapshot_hold 0.00 sec in ztest_verify_dnode_bt 8.93 sec in ztest_zap_parallel 9.10 sec in ztest_dsl_prop_get_set 0.71 sec in ztest_spa_prop_get_set 0.00 sec in ztest_zil_commit 14.52 sec in ztest_dsl_dataset_promote_busy 7.94 sec in ztest_dmu_read_write_zcopy 9.91 sec in ztest_dmu_read_write_zcopy 9.49 sec in ztest_dsl_prop_get_set 9.48 sec in ztest_dsl_prop_get_set 0.00 sec in ztest_split_pool 0.00 sec in ztest_split_pool 0.00 sec in ztest_dmu_commit_callbacks 8.84 sec in ztest_zap 0.01 sec in ztest_fault_inject 0.70 sec in ztest_zil_commit 1.68 sec in ztest_dmu_read_write_zcopy 1.70 sec in ztest_dmu_commit_callbacks 10.04 sec in ztest_dsl_prop_get_set 13.44 sec in ztest_dmu_snapshot_hold 13.48 sec in ztest_dmu_snapshot_hold 0.00 sec in ztest_dmu_commit_callbacks 11.46 sec in ztest_zap 0.24 sec in ztest_zil_commit 0.03 sec in ztest_fzap 1.43 sec in ztest_dmu_snapshot_create_destroy 4.48 sec in ztest_dmu_read_write_zcopy 17.55 sec in ztest_dmu_object_alloc_free 0.26 sec in ztest_zil_commit 2.71 sec in ztest_dmu_object_alloc_free 1.32 sec in ztest_spa_create_destroy 2.55 sec in ztest_zil_remount 0.00 sec in ztest_dmu_commit_callbacks 3.44 sec in ztest_dmu_read_write_zcopy 6.95 sec in ztest_zap 0.20 sec in ztest_vdev_attach_detach 14.30 sec in ztest_dmu_object_alloc_free 0.00 sec in ztest_split_pool 2.22 sec in ztest_dmu_read_write dmu_read(zd->zd_os, object, offset, blocksize, data, 1) == 0 (0x5 == 0x0) ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] /lib64/libpthread.so.0(+0xf100)[0x7fd37ed94100] /lib64/libc.so.6(gsignal+0x37)[0x7fd37e9f85f7] /lib64/libc.so.6(abort+0x148)[0x7fd37e9f9ce8] /usr/sbin/ztest[0x4089dd] /usr/sbin/ztest[0x41045d] /usr/sbin/ztest[0x41323f] /usr/sbin/ztest[0x40cda3] /lib64/libzpool.so.2(zk_thread_helper+0x1ec)[0x7fd37ff887ac] /lib64/libpthread.so.0(+0x7dc5)[0x7fd37ed8cdc5] /lib64/libc.so.6(clone+0x6d)[0x7fd37eab9ced] 20.91 sec in ztest_dmu_object_alloc_free child died with signal 6 ``` status ``` [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib64/libthread_db.so.1". Core was generated by `/usr/sbin/ztest'. Program terminated with signal 6, Aborted. #0 0x00007fd37e9f85f7 in raise () from /lib64/libc.so.6 * * Backtrace * #0 0x00007fd37e9f85f7 in raise () from /lib64/libc.so.6 #1 0x00007fd37e9f9ce8 in abort () from /lib64/libc.so.6 #2 0x00000000004089dd in libspl_assert () #3 0x000000000041045d in ztest_io () #4 0x000000000041323f in ztest_dmu_write_parallel () #5 0x000000000040cda3 in ztest_thread () #6 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #7 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #8 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 749 (Thread 0x7fd362aec700 (LWP 116099)): #0 0x00007fd37ed9003e in pthread_rwlock_wrlock () from /lib64/libpthread.so.0 #1 0x000000000040a40c in ztest_spa_rename () #2 0x000000000040cda3 in ztest_thread () #3 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #4 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #5 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 538 (Thread 0x7fd37d335700 (LWP 114824)): #0 0x00007fd37ed93af3 in pwrite64 () from /lib64/libpthread.so.0 #1 0x00007fd37ff8a8b7 in vn_rdwr () from /lib64/libzpool.so.2 #2 0x00007fd380006df3 in vdev_file_io_strategy () from /lib64/libzpool.so.2 #3 0x00007fd37ff8bd94 in taskq_thread () from /lib64/libzpool.so.2 #4 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #5 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #6 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 284 (Thread 0x7fd380ae57c0 (LWP 114793)): #0 0x00007fd37ed8def7 in pthread_join () from /lib64/libpthread.so.0 #1 0x00007fd37ff892f1 in zk_thread_join () from /lib64/libzpool.so.2 #2 0x0000000000407805 in ztest_run () #3 0x00000000004087e4 in main () Thread 7 (Thread 0x7fd37d4bb700 (LWP 114818)): #0 0x00007fd37ed93af3 in pwrite64 () from /lib64/libpthread.so.0 #1 0x00007fd37ff8a892 in vn_rdwr () from /lib64/libzpool.so.2 #2 0x00007fd380006df3 in vdev_file_io_strategy () from /lib64/libzpool.so.2 #3 0x00007fd37ff8bd94 in taskq_thread () from /lib64/libzpool.so.2 #4 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #5 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #6 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 2 (Thread 0x7fd362a6a700 (LWP 116098)): #0 0x00007fd3804d283c in atomic_inc_ulong_nv () from /lib64/libuutil.so.1 #1 0x00007fd37fff300d in spa_open_common () from /lib64/libzpool.so.2 #2 0x00007fd37ffafbc5 in dmu_objset_find () from /lib64/libzpool.so.2 #3 0x0000000000412ceb in ztest_dmu_objset_create_destroy () #4 0x000000000040cda3 in ztest_thread () #5 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #6 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #7 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 1 (Thread 0x7fd362aab700 (LWP 116082)): #0 0x00007fd37e9f85f7 in raise () from /lib64/libc.so.6 #1 0x00007fd37e9f9ce8 in abort () from /lib64/libc.so.6 #2 0x00000000004089dd in libspl_assert () #3 0x000000000041045d in ztest_io () #4 0x000000000041323f in ztest_dmu_write_parallel () #5 0x000000000040cda3 in ztest_thread () #6 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #7 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #8 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 ``` zdb ``` Dataset mos [META], ID 0, cr_txg 4, 562K, 150 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 2 16K 16K 71.5K 512 80.0K 93.75 DMU dnode 1 1 16K 16K 24.0K 512 32K 100.00 object directory 2 1 16K 512 0 512 512 0.00 DSL directory 3 1 16K 512 0 512 512 100.00 DSL props 4 1 16K 512 0 512 512 100.00 DSL directory child map 5 1 16K 512 0 512 512 100.00 zap 6 1 16K 512 0 512 512 100.00 DSL dataset snap map 7 1 16K 512 0 512 512 100.00 DSL deadlist map 8 1 16K 128K 0 512 128K 0.00 bpobj 9 1 16K 4K 12.0K 512 4K 100.00 SPA space map 10 1 16K 4K 12.0K 512 4K 100.00 SPA space map 11 1 16K 4K 12.0K 512 4K 100.00 SPA space map 12 1 16K 4K 12.0K 512 4K 100.00 SPA space map 13 1 16K 4K 12.0K 512 4K 100.00 SPA space map 14 1 16K 4K 12.0K 512 4K 100.00 SPA space map 15 1 16K 512 0 512 512 0.00 DSL directory 16 1 16K 512 0 512 512 100.00 DSL props 17 1 16K 512 0 512 512 100.00 DSL directory child map 18 1 16K 512 0 512 512 100.00 zap 19 1 16K 512 0 512 512 100.00 DSL dataset snap map 20 1 16K 512 0 512 512 100.00 DSL deadlist map 21 2 16K 4K 36.0K 512 8K 100.00 SPA space map 22 2 16K 4K 36.0K 512 8K 100.00 SPA space map 23 2 16K 4K 24.0K 512 4K 100.00 SPA space map 24 1 16K 512 0 512 512 0.00 DSL directory 25 1 16K 512 0 512 512 100.00 DSL props 26 1 16K 512 0 512 512 100.00 DSL directory child map 27 1 16K 512 0 512 512 100.00 zap 28 1 16K 512 0 512 512 100.00 DSL dataset snap map 29 1 16K 512 0 512 512 100.00 DSL deadlist map 30 1 16K 512 0 512 512 0.00 DSL directory 31 1 16K 512 12.0K 512 512 100.00 DSL props 32 1 16K 512 0 512 512 0.00 DSL directory 33 1 16K 512 0 512 512 100.00 DSL props 34 1 16K 1.50K 12.0K 512 1.50K 100.00 DSL directory child map 35 1 16K 512 0 512 512 0.00 DSL directory 36 1 16K 512 0 512 512 100.00 DSL props 37 1 16K 512 0 512 512 100.00 DSL directory child map 38 1 16K 512 0 512 512 0.00 DSL directory 39 1 16K 512 0 512 512 100.00 DSL props 40 1 16K 512 0 512 512 100.00 DSL directory child map 41 1 16K 128K 0 512 128K 0.00 bpobj 42 1 16K 512 0 512 512 0.00 DSL directory 43 1 16K 512 0 512 512 100.00 DSL props 44 1 16K 512 0 512 512 100.00 DSL directory child map 45 1 16K 512 0 512 512 0.00 DSL dataset 46 1 16K 512 0 512 512 100.00 DSL dataset snap map 47 1 16K 512 0 512 512 100.00 DSL deadlist map 48 1 16K 512 0 512 512 0.00 DSL dataset 49 1 16K 512 0 512 512 100.00 DSL deadlist map 50 1 16K 128K 0 512 128K 0.00 bpobj 51 1 16K 512 0 512 512 0.00 DSL dataset 52 1 16K 512 0 512 512 100.00 DSL dataset snap map 53 1 16K 512 0 512 512 100.00 DSL deadlist map 54 1 16K 128K 0 512 128K 0.00 bpobj 55 1 16K 1K 12.0K 512 1K 100.00 DSL dataset next clones 56 1 16K 1K 12.0K 512 1K 100.00 DSL dir clones 57 1 16K 16K 12.0K 512 16K 100.00 packed nvlist 58 1 16K 1K 12.0K 512 1K 100.00 zap 59 1 16K 1K 12.0K 512 1K 100.00 zap 60 1 16K 16K 36.0K 512 32K 100.00 zap 61 1 16K 16K 12.0K 512 16K 100.00 bpobj (Z=uncompressed) 62 1 16K 128K 12.0K 512 128K 100.00 SPA history 63 1 16K 512 0 512 512 100.00 Pool properties 64 1 16K 1K 12.0K 512 1K 100.00 zap 65 1 16K 1K 12.0K 512 1K 100.00 zap 66 1 16K 512 0 512 512 100.00 zap 67 1 16K 512 0 512 512 100.00 zap 68 1 16K 512 0 512 512 100.00 zap 69 1 16K 512 0 512 512 100.00 zap 70 1 16K 512 0 512 512 100.00 object array 71 1 16K 4K 12.0K 512 4K 100.00 SPA space map 72 1 16K 4K 12.0K 512 4K 100.00 SPA space map 73 1 16K 4K 12.0K 512 4K 100.00 SPA space map 74 1 16K 512 0 512 512 100.00 DSL directory child map 75 1 16K 512 0 512 512 100.00 zap 76 1 16K 512 0 512 512 100.00 DSL dataset snap map 77 1 16K 512 0 512 512 100.00 DSL deadlist map 78 1 16K 512 0 512 512 0.00 DSL directory 79 1 16K 512 0 512 512 100.00 DSL props 80 1 16K 512 0 512 512 100.00 DSL directory child map 81 1 16K 512 0 512 512 100.00 zap 82 1 16K 512 0 512 512 100.00 DSL dataset snap map 83 1 16K 512 0 512 512 100.00 DSL deadlist map 84 1 16K 512 0 512 512 0.00 DSL directory 85 1 16K 512 0 512 512 100.00 DSL props 86 1 16K 512 0 512 512 100.00 DSL directory child map 87 1 16K 512 0 512 512 100.00 zap 88 1 16K 512 0 512 512 100.00 DSL dataset snap map 89 1 16K 512 0 512 512 100.00 DSL deadlist map 90 1 16K 512 0 512 512 0.00 DSL directory 91 1 16K 512 0 512 512 100.00 DSL props 92 1 16K 512 0 512 512 100.00 DSL directory child map 93 1 16K 512 0 512 512 100.00 zap 94 1 16K 512 0 512 512 100.00 DSL dataset snap map 95 1 16K 512 0 512 512 100.00 DSL deadlist map 96 1 16K 512 0 512 512 100.00 zap 97 1 16K 512 0 512 512 100.00 DSL deadlist map 98 1 16K 512 0 512 512 0.00 DSL directory 99 1 16K 512 0 512 512 100.00 DSL props 100 1 16K 512 0 512 512 100.00 DSL directory child map 101 1 16K 512 0 512 512 100.00 zap 102 1 16K 512 0 512 512 100.00 DSL dataset snap map 103 1 16K 512 0 512 512 100.00 DSL deadlist map 104 1 16K 512 0 512 512 0.00 DSL directory 105 1 16K 512 0 512 512 100.00 DSL props 106 1 16K 512 0 512 512 100.00 DSL directory child map 107 1 16K 512 0 512 512 100.00 zap 108 1 16K 512 0 512 512 100.00 DSL dataset snap map 109 1 16K 512 0 512 512 100.00 DSL deadlist map 110 1 16K 16K 24.0K 512 32K 100.00 DDT statistics 111 1 4K 4K 12.0K 512 8K 100.00 DDT ZAP algorithm 112 1 16K 512 0 512 512 0.00 DSL directory 113 1 16K 512 0 512 512 100.00 DSL props 114 1 16K 512 0 512 512 100.00 DSL directory child map 115 1 16K 512 0 512 512 100.00 zap 116 1 16K 512 0 512 512 100.00 DSL dataset snap map 117 1 16K 512 0 512 512 100.00 DSL deadlist map 118 1 16K 512 0 512 512 0.00 DSL directory 119 1 16K 512 0 512 512 100.00 DSL props 120 1 16K 512 0 512 512 100.00 DSL directory child map 121 1 16K 512 0 512 512 100.00 zap 122 1 16K 512 0 512 512 100.00 DSL dataset snap map 123 1 16K 512 0 512 512 100.00 DSL deadlist map 124 1 16K 512 0 512 512 100.00 zap 125 1 16K 512 0 512 512 100.00 zap 126 1 16K 512 0 512 512 100.00 zap 127 1 16K 512 0 512 512 100.00 zap 128 1 16K 512 0 512 512 100.00 object array 129 1 16K 4K 12.0K 512 4K 100.00 SPA space map 130 1 16K 4K 12.0K 512 4K 100.00 SPA space map 131 1 16K 4K 12.0K 512 4K 100.00 SPA space map 132 1 16K 512 0 512 512 100.00 zap 133 1 16K 512 0 512 512 100.00 zap 134 1 16K 512 0 512 512 100.00 zap 135 1 16K 512 0 512 512 100.00 zap 136 1 16K 512 0 512 512 0.00 DSL directory 137 1 16K 512 0 512 512 100.00 DSL props 138 1 16K 512 0 512 512 100.00 DSL directory child map 139 1 16K 512 0 512 512 0.00 DSL dataset 140 1 16K 512 0 512 512 100.00 DSL dataset snap map 141 1 16K 512 0 512 512 100.00 DSL deadlist map 142 1 16K 512 0 512 512 100.00 object array 143 1 16K 4K 0 512 4K 100.00 SPA space map 144 1 16K 4K 12.0K 512 4K 100.00 SPA space map 145 1 16K 512 0 512 512 0.00 DSL directory 146 1 16K 512 0 512 512 100.00 DSL props 147 1 16K 512 0 512 512 100.00 DSL directory child map 148 1 16K 512 0 512 512 100.00 zap 149 1 16K 512 0 512 512 100.00 DSL dataset snap map 150 1 16K 512 0 512 512 100.00 DSL deadlist map Dataset ztest/temp_12 [OTHER], ID 115, cr_txg 85, 566K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 4 8K 4K 87.5K 1K 928M 0.00 other uint64[] 4 3 16K 64K 223K 1K 960M 0.02 other uint64[] 6 4 4K 128K 152K 1.50K 304M 0.04 other uint64[] 9 4 8K 8K 32K 512 400M 0.00 other uint64[] Dataset ztest/temp_7 [OTHER], ID 121, cr_txg 87, 195K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 5 8K 512 32K 512 400M 0.00 other uint64[] 3 3 16K 8K 24.0K 2K 32.0M 0.02 other uint64[] 7 4 16K 16K 28.0K 512 976M 0.00 other uint64[] 8 4 16K 16K 40.0K 512 432M 0.00 other uint64[] Dataset ztest/ds_2 [OTHER], ID 27, cr_txg 65, 1000K, 13 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 108K 512 48.0K 13.54 DMU dnode 1 1 16K 1K 12.0K 512 1K 100.00 other ZAP 2 7 8K 8K 271K 512 343T 0.00 other uint64[] 3 6 8K 32K 219K 512 1.01T 0.00 other uint64[] 4 1 16K 1.50K 8K 512 1.50K 100.00 other ZAP 5 2 16K 16K 24.0K 512 32K 100.00 other ZAP 6 5 8K 1K 91.5K 512 784M 0.00 other uint64[] 7 6 4K 512 104K 2K 1008M 0.00 other uint64[] 11 5 4K 2K 36.0K 512 896M 0.00 other uint64[] 12 4 16K 4K 0 1K 544M 0.00 other uint64[] 72 1 16K 512 0 9.00K 512 100.00 other ZAP 90 8 8K 64K 64K 1K 4.32P 0.00 other uint64[] 92 7 4K 16K 52.0K 512 1.62T 0.00 other uint64[] 93 6 16K 2K 0 1K 16.0T 0.00 other uint64[] Dataset ztest/ds_5 [OTHER], ID 87, cr_txg 75, 566K, 6 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 79.5K 512 48.0K 6.25 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 8 16K 128K 439K 1K 4.00E 0.00 other uint64[] 4 1 16K 2K 8K 512 2K 100.00 other ZAP 5 2 16K 16K 24.0K 1K 32K 100.00 other ZAP 32 1 16K 512 0 7.50K 512 100.00 other ZAP 64 1 4K 32K 0 14.5K 32K 0.00 other uint64[] Dataset ztest/temp_4@0 [OTHER], ID 96, cr_txg 77, 71.7K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 1 8K 2K 0 2K 2K 0.00 other uint64[] 6 1 4K 128K 0 2K 128K 0.00 other uint64[] 10 1 16K 512 0 1K 512 0.00 other uint64[] 12 1 16K 32K 0 512 32K 0.00 other uint64[] Dataset ztest/temp_4 [OTHER], ID 93, cr_txg 75, 71.7K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 1 8K 2K 0 2K 2K 0.00 other uint64[] 6 1 4K 128K 0 2K 128K 0.00 other uint64[] 10 1 16K 512 0 1K 512 0.00 other uint64[] 12 1 16K 32K 0 512 32K 0.00 other uint64[] Dataset ztest/ds_1 [OTHER], ID 18, cr_txg 63, 928K, 9 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 87.5K 512 112K 4.02 DMU dnode 1 1 16K 1K 8K 512 1K 100.00 other ZAP 2 2 16K 16K 24.0K 1.50K 32K 100.00 other ZAP 5 1 16K 2.50K 8K 1.50K 2.50K 100.00 other ZAP 8 8 4K 128K 64K 512 834T 0.00 other uint64[] (K=sha256) (Z=gzip-5) 9 7 4K 16K 64K 512 2.24T 0.00 other uint64[] 10 1 4K 128K 4K 1K 128K 100.00 other uint64[] 96 9 8K 64K 638K 11.0K 2.00E 0.00 other uint64[] 139 2 16K 16K 24.0K 6.00K 32K 100.00 other ZAP 208 1 16K 512 0 512 512 100.00 other ZAP Dataset ztest/ds_6 [OTHER], ID 101, cr_txg 77, 295K, 6 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 79.5K 512 64K 4.69 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 3 1 4K 16K 16K 13.0K 16K 100.00 other uint64[] 47 6 16K 128K 95.5K 1K 871T 0.00 other uint64[] (K=fletcher4) (Z=on) 49 6 8K 8K 87.5K 512 1.85T 0.00 other uint64[] 79 1 16K 512 0 2K 512 100.00 other ZAP 97 1 16K 512 0 2K 512 100.00 other ZAP Dataset ztest/temp_9 [OTHER], ID 107, cr_txg 80, 618K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 6 4K 512 136K 8K 912M 0.00 other uint64[] 18 5 8K 2K 83.5K 512 576M 0.00 other uint64[] 19 4 16K 4K 48.0K 1.50K 720M 0.00 other uint64[] 22 3 16K 128K 279K 1K 848M 0.03 other uint64[] Dataset ztest/ds_4 [OTHER], ID 81, cr_txg 73, 506K, 8 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 87.5K 512 64K 6.25 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 2 16K 16K 24.0K 2K 32K 100.00 other ZAP 32 1 16K 2.50K 8K 14.5K 2.50K 100.00 other ZAP 61 1 8K 512 0 512 512 0.00 other uint64[] 62 8 16K 64K 255K 512 1.00E 0.00 other uint64[] 96 1 16K 512 0 11.5K 512 100.00 other ZAP 119 9 4K 512 64K 512 48.7T 0.00 other uint64[] 120 7 4K 8K 52.0K 2K 2.28T 0.00 other uint64[] Dataset ztest/ds_3 [OTHER], ID 75, cr_txg 69, 1.26M, 17 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 87.5K 512 128K 6.64 DMU dnode 1 1 16K 1.50K 8K 512 1.50K 100.00 other ZAP 2 9 4K 2K 71.5K 2K 81.4T 0.00 other uint64[] 6 5 16K 8K 36.0K 12.0K 977G 0.00 other uint64[] 30 1 4K 64K 79.5K 512 64K 100.00 other uint64[] 31 1 16K 512 0 512 512 100.00 other uint64[] 65 1 16K 512 0 512 512 100.00 other ZAP 66 9 16K 512 415K 1.50K 2.00E 0.00 other uint64[] 224 1 16K 512 0 1K 512 100.00 other ZAP 226 7 8K 32K 295K 512 635T 0.00 other uint64[] (K=fletcher2) (Z=zle) 227 8 4K 1K 239K 512 1.86T 0.00 other uint64[] 228 1 4K 64K 4K 1.50K 64K 100.00 other uint64[] 231 1 4K 4K 0 1K 4K 0.00 other uint64[] 233 1 8K 512 0 2K 512 0.00 other uint64[] 237 1 4K 32K 0 512 32K 0.00 other uint64[] 238 1 16K 64K 0 1.50K 64K 0.00 other uint64[] 241 1 16K 1.50K 12.0K 1K 1.50K 100.00 other ZAP 243 2 16K 16K 36.0K 2K 32K 100.00 other ZAP Dataset ztest/temp_2 [OTHER], ID 139, cr_txg 109, 63.8K, 1 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 3.12 DMU dnode 1 1 16K 512 0 512 512 100.00 other ZAP Dataset ztest/temp_5 [OTHER], ID 148, cr_txg 116, 586K, 5 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 71.5K 512 32K 7.81 DMU dnode 1 1 16K 512 0 512 512 100.00 other ZAP 8 5 4K 1K 91.5K 9.5K 960M 0.00 other uint64[] 27 5 4K 8K 40.0K 512 528M 0.00 other uint64[] 28 4 4K 64K 104K 2K 480M 0.01 other uint64[] 32 3 16K 64K 271K 512 928M 0.02 other uint64[] Dataset ztest/ds_0 [OTHER], ID 5, cr_txg 7, 1.45M, 14 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 71.5K 512 32K 21.88 DMU dnode 1 1 16K 1K 8K 512 1K 100.00 other ZAP 2 5 8K 512 108K 1K 912M 0.00 other uint64[] 4 4 16K 4K 28.0K 3.00K 384M 0.00 other uint64[] 10 4 4K 64K 108K 1K 864M 0.01 other uint64[] 12 4 4K 64K 0 2K 464M 0.00 other uint64[] 16 4 4K 32K 383K 1K 1008M 0.02 other uint64[] 18 9 16K 8K 622K 512 4.00E 0.00 other uint64[] 19 2 16K 16K 24.0K 2K 32K 100.00 other ZAP 23 8 8K 8K 56.0K 512 595T 0.00 other uint64[] (K=fletcher2) (Z=gzip-7) 24 8 4K 1K 60.0K 512 1.19T 0.00 other uint64[] 25 1 16K 1K 8K 1K 1K 100.00 other ZAP 27 1 4K 16K 0 2K 16K 100.00 other uint64[] 32 1 4K 8K 0 2K 8K 0.00 other uint64[] (K=on) (Z=zle) 36 6 4K 64K 0 512 486G 0.00 other uint64[] Dataset ztest [ZPL], ID 51, cr_txg 1, 7.97K, 0 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 0 512 16K 0.00 DMU dnode Verified large_blocks feature refcount of 0 is correct Verified large_dnode feature refcount of 13 is correct Verified sha512 feature refcount of 0 is correct Verified skein feature refcount of 2 is correct Verified edonr feature refcount of 1 is correct Verified userobj_accounting feature refcount of 0 is correct Traversing all blocks to verify checksums and verify nothing leaked ... loading space map for vdev 2 of 3, metaslab 0 of 23 .... No leaks (block sum matches space maps exactly) bp count: 1062 ganged count: 10 bp logical: 14778368 avg: 13915 bp physical: 5126144 avg: 4826 compression: 2.88 bp allocated: 27193344 avg: 25605 compression: 0.54 bp deduped: 0 ref>1: 0 deduplication: 1.00 SPA allocated: 27193344 used: 2.35% additional, non-pointer bps of type 0: 173 Dittoed blocks on same vdev: 610 Blocks LSIZE PSIZE ASIZE avg comp %Total Type - - - - - - - unallocated 2 32K 8K 72.0K 36.0K 4.00 0.27 object directory - - - - - - - object array 1 16K 4K 36.0K 36.0K 4.00 0.14 packed nvlist - - - - - - - packed nvlist size 1 16K 4K 36.0K 36.0K 4.00 0.14 bpobj - - - - - - - bpobj header - - - - - - - SPA space map header 22 124K 84.0K 756K 34.4K 1.48 2.85 SPA space map 5 20.0K 20.0K 60.0K 12.0K 1.00 0.23 ZIL intent log 112 1.75M 480K 3.01M 27.5K 3.73 11.61 DMU dnode 15 30.0K 30.0K 384K 25.6K 1.00 1.45 DMU objset - - - - - - - DSL directory 17 9.5K 1.50K 36.0K 2.12K 6.33 0.14 DSL directory child map - - - - - - - DSL dataset snap map 17 8.50K 512 36.0K 2.12K 17.00 0.14 DSL props - - - - - - - DSL dataset - - - - - - - ZFS znode - - - - - - - ZFS V0 ACL - - - - - - - ZFS plain file - - - - - - - ZFS directory - - - - - - - ZFS master node - - - - - - - ZFS delete queue - - - - - - - zvol object - - - - - - - zvol prop - - - - - - - other uint8[] 743 11.5M 4.12M 20.0M 27.5K 2.78 76.95 other uint64[] 48 360K 103K 972K 20.2K 3.50 3.66 other ZAP - - - - - - - persistent error log 1 128K 4K 36.0K 36.0K 32.00 0.14 SPA history - - - - - - - SPA history offsets - - - - - - - Pool properties - - - - - - - DSL permissions - - - - - - - ZFS ACL - - - - - - - ZFS SYSACL - - - - - - - FUID table - - - - - - - FUID table size 1 1K 1K 36.0K 36.0K 1.00 0.14 DSL dataset next clones - - - - - - - scan work queue - - - - - - - ZFS user/group used - - - - - - - ZFS user/group quota - - - - - - - snapshot refcount tags 2 8K 4K 36.0K 18.0K 2.00 0.14 DDT ZAP algorithm 2 32K 8K 72.0K 36.0K 4.00 0.27 DDT statistics - - - - - - - System attributes - - - - - - - SA master node - - - - - - - SA attr registration - - - - - - - SA attr layouts - - - - - - - scan translations - - - - - - - deduplicated block - - - - - - - DSL deadlist map - - - - - - - DSL deadlist map hdr 1 1K 1K 36.0K 36.0K 1.00 0.14 DSL dir clones - - - - - - - bpobj subobj 5 44.0K 20.0K 180K 36.0K 2.20 0.68 deferred free - - - - - - - dedup ditto 31 48.5K 16K 252K 8.13K 3.03 0.95 other 1.04K 14.1M 4.89M 25.9M 25.0K 2.88 100.00 Total capacity operations bandwidth ---- errors ---- description used avail read write read write read write cksum ztest 25.9M 1.05G 6.58K 0 41.4M 0 0 0 0 raidz2 22.3M 346M 5.49K 0 30.7M 0 0 0 0 /var/tmp/ztest.0a 1.67K 0 9.43M 0 0 0 0 /var/tmp/ztest.1a 1.59K 0 8.94M 0 0 0 34 /var/tmp/ztest.2a 2.23K 0 12.3M 0 0 0 0 raidz2 1.63M 366M 649 0 5.21M 0 0 0 0 /var/tmp/ztest.3a 195 0 1.65M 0 0 0 0 /var/tmp/ztest.4a 195 0 1.65M 0 0 0 0 /var/tmp/ztest.5a 259 0 1.91M 0 0 0 0 raidz2 1.96M 366M 471 0 5.44M 0 0 0 0 /var/tmp/ztest.6a 139 0 1.66M 0 0 0 0 /var/tmp/ztest.7a 139 0 1.66M 0 0 0 0 /var/tmp/ztest.8a 193 0 2.11M 0 0 0 0 ``` We have hit this 4 times recently: zloop-161216-162658/ztest.out:ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] zloop-161217-173015/ztest.out:ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] zloop-161218-091534/ztest.out:ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] zloop-161217-001943/ztest.out:ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf]
1.0
ztest: dmu_read() ASSERT on EIO - Version Information ================================================================= Distribution Name Centos 7.2 Distribution Version 7.2 Linux Kernel 3.10.0-327.36.3.el7.x86_64 Architecture x86_64 ZFS Version master Dec 1st RPM/modinfo say: 0.7.0-rc2 SPL Version master Dec 1st RPM/modinfo say: 0.7.0-rc2 Git Build Data Revision: 5f20c145df3ca45f86cd1892137edb117d4c8a82 refs/remotes/origin/master Built Branches refs/remotes/origin/master: Build #183 of Revision 5f20c145df3ca45f86cd1892137edb117d4c8a82 (refs/remotes/origin/master) Problem ================================================================= Running zloop we hit an assert: dmu_read(zd->zd_os, object, offset, blocksize, data, 1) == 0 (0x5 == 0x0) ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] Reproduce ================================================================= /sbin/ztest -VVVVV -m 2 -r 0 -R 1 -v 2 -a 9 -T 7 -P 12 -s 128m -f /var/tmp /sbin/ztest -VVVVV -m 2 -r 0 -R 1 -v 2 -a 9 -T 72 -P 25 -s 128m -f /var/tmp Log detail ================================================================= ztest.out ``` starting main threads... 0.00 sec in ztest_verify_dnode_bt 0.01 sec in ztest_dmu_object_alloc_free 0.00 sec in ztest_split_pool 0.00 sec in ztest_zil_commit 0.00 sec in ztest_zil_commit 0.00 sec in ztest_dmu_commit_callbacks 0.01 sec in ztest_zap 0.11 sec in ztest_vdev_attach_detach 0.11 sec in ztest_fault_inject 0.13 sec in ztest_spa_prop_get_set 0.11 sec in ztest_dmu_read_write_zcopy 0.20 sec in ztest_vdev_attach_detach 0.21 sec in ztest_zil_commit 0.23 sec in ztest_spa_prop_get_set 0.22 sec in ztest_zap_parallel 0.00 sec in ztest_dmu_commit_callbacks 0.23 sec in ztest_fault_inject 0.07 sec in ztest_dmu_read_write 0.02 sec in ztest_dmu_read_write 0.95 sec in ztest_zil_remount 1.05 sec in ztest_dmu_object_alloc_free 0.00 sec in ztest_dmu_commit_callbacks 1.13 sec in ztest_spa_create_destroy 0.00 sec in ztest_zil_commit Setting dataset ztest/temp_0 to sync always 1.17 sec in ztest_dmu_object_alloc_free 1.32 sec in ztest_dmu_snapshot_create_destroy 1.52 sec in ztest_dmu_objset_create_destroy 0.01 sec in ztest_zap 0.00 sec in ztest_split_pool Setting dataset ztest/temp_22 to sync always 1.91 sec in ztest_fzap 0.85 sec in ztest_dmu_read_write_zcopy 0.01 sec in ztest_dmu_read_write 0.01 sec in ztest_fault_inject Setting dataset ztest/temp_20 to sync always 1.62 sec in ztest_dmu_object_alloc_free 3.00 sec in ztest_dmu_read_write_zcopy 1.21 sec in ztest_spa_create_destroy 3.24 sec in ztest_zap_parallel 1.50 sec in ztest_dmu_snapshot_create_destroy 0.31 sec in ztest_zap_parallel 0.79 sec in ztest_spa_create_destroy 3.79 sec in ztest_dsl_prop_get_set 3.79 sec in ztest_dsl_prop_get_set 3.80 sec in ztest_dsl_prop_get_set 0.00 sec in ztest_dmu_commit_callbacks 3.79 sec in ztest_dsl_prop_get_set 0.00 sec in ztest_split_pool 0.00 sec in ztest_dmu_read_write 0.00 sec in ztest_dmu_object_alloc_free 0.69 sec in ztest_spa_create_destroy 4.48 sec in ztest_dmu_objset_create_destroy 0.69 sec in ztest_dmu_read_write_zcopy 0.00 sec in ztest_zap 0.00 sec in ztest_dmu_read_write 3.56 sec in ztest_dsl_prop_get_set 1.01 sec in ztest_dmu_snapshot_create_destroy 1.46 sec in ztest_dmu_snapshot_create_destroy 0.00 sec in ztest_zil_commit 3.19 sec in ztest_dsl_prop_get_set 1.42 sec in ztest_zil_remount 4.12 sec in ztest_dmu_object_alloc_free 1.27 sec in ztest_zap 1.43 sec in ztest_dmu_read_write 1.53 sec in ztest_dmu_read_write_zcopy 5.87 sec in ztest_dmu_objset_create_destroy 6.09 sec in ztest_dmu_snapshot_hold 0.00 sec in ztest_zil_commit Unable to expand vdev, state 4, error 0 Could not expand LUN because the vdev configuration changed. 0.91 sec in ztest_vdev_LUN_growth 2.61 sec in ztest_dsl_prop_get_set 2.98 sec in ztest_dsl_prop_get_set 0.33 sec in ztest_zap 0.00 sec in ztest_split_pool 0.00 sec in ztest_dmu_commit_callbacks 0.32 sec in ztest_zil_commit 0.00 sec in ztest_zil_commit 0.17 sec in ztest_zil_commit 6.73 sec in ztest_vdev_attach_detach 13.17 sec in ztest_dmu_snapshot_hold 0.00 sec in ztest_verify_dnode_bt 8.93 sec in ztest_zap_parallel 9.10 sec in ztest_dsl_prop_get_set 0.71 sec in ztest_spa_prop_get_set 0.00 sec in ztest_zil_commit 14.52 sec in ztest_dsl_dataset_promote_busy 7.94 sec in ztest_dmu_read_write_zcopy 9.91 sec in ztest_dmu_read_write_zcopy 9.49 sec in ztest_dsl_prop_get_set 9.48 sec in ztest_dsl_prop_get_set 0.00 sec in ztest_split_pool 0.00 sec in ztest_split_pool 0.00 sec in ztest_dmu_commit_callbacks 8.84 sec in ztest_zap 0.01 sec in ztest_fault_inject 0.70 sec in ztest_zil_commit 1.68 sec in ztest_dmu_read_write_zcopy 1.70 sec in ztest_dmu_commit_callbacks 10.04 sec in ztest_dsl_prop_get_set 13.44 sec in ztest_dmu_snapshot_hold 13.48 sec in ztest_dmu_snapshot_hold 0.00 sec in ztest_dmu_commit_callbacks 11.46 sec in ztest_zap 0.24 sec in ztest_zil_commit 0.03 sec in ztest_fzap 1.43 sec in ztest_dmu_snapshot_create_destroy 4.48 sec in ztest_dmu_read_write_zcopy 17.55 sec in ztest_dmu_object_alloc_free 0.26 sec in ztest_zil_commit 2.71 sec in ztest_dmu_object_alloc_free 1.32 sec in ztest_spa_create_destroy 2.55 sec in ztest_zil_remount 0.00 sec in ztest_dmu_commit_callbacks 3.44 sec in ztest_dmu_read_write_zcopy 6.95 sec in ztest_zap 0.20 sec in ztest_vdev_attach_detach 14.30 sec in ztest_dmu_object_alloc_free 0.00 sec in ztest_split_pool 2.22 sec in ztest_dmu_read_write dmu_read(zd->zd_os, object, offset, blocksize, data, 1) == 0 (0x5 == 0x0) ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] /lib64/libpthread.so.0(+0xf100)[0x7fd37ed94100] /lib64/libc.so.6(gsignal+0x37)[0x7fd37e9f85f7] /lib64/libc.so.6(abort+0x148)[0x7fd37e9f9ce8] /usr/sbin/ztest[0x4089dd] /usr/sbin/ztest[0x41045d] /usr/sbin/ztest[0x41323f] /usr/sbin/ztest[0x40cda3] /lib64/libzpool.so.2(zk_thread_helper+0x1ec)[0x7fd37ff887ac] /lib64/libpthread.so.0(+0x7dc5)[0x7fd37ed8cdc5] /lib64/libc.so.6(clone+0x6d)[0x7fd37eab9ced] 20.91 sec in ztest_dmu_object_alloc_free child died with signal 6 ``` status ``` [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib64/libthread_db.so.1". Core was generated by `/usr/sbin/ztest'. Program terminated with signal 6, Aborted. #0 0x00007fd37e9f85f7 in raise () from /lib64/libc.so.6 * * Backtrace * #0 0x00007fd37e9f85f7 in raise () from /lib64/libc.so.6 #1 0x00007fd37e9f9ce8 in abort () from /lib64/libc.so.6 #2 0x00000000004089dd in libspl_assert () #3 0x000000000041045d in ztest_io () #4 0x000000000041323f in ztest_dmu_write_parallel () #5 0x000000000040cda3 in ztest_thread () #6 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #7 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #8 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 749 (Thread 0x7fd362aec700 (LWP 116099)): #0 0x00007fd37ed9003e in pthread_rwlock_wrlock () from /lib64/libpthread.so.0 #1 0x000000000040a40c in ztest_spa_rename () #2 0x000000000040cda3 in ztest_thread () #3 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #4 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #5 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 538 (Thread 0x7fd37d335700 (LWP 114824)): #0 0x00007fd37ed93af3 in pwrite64 () from /lib64/libpthread.so.0 #1 0x00007fd37ff8a8b7 in vn_rdwr () from /lib64/libzpool.so.2 #2 0x00007fd380006df3 in vdev_file_io_strategy () from /lib64/libzpool.so.2 #3 0x00007fd37ff8bd94 in taskq_thread () from /lib64/libzpool.so.2 #4 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #5 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #6 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 284 (Thread 0x7fd380ae57c0 (LWP 114793)): #0 0x00007fd37ed8def7 in pthread_join () from /lib64/libpthread.so.0 #1 0x00007fd37ff892f1 in zk_thread_join () from /lib64/libzpool.so.2 #2 0x0000000000407805 in ztest_run () #3 0x00000000004087e4 in main () Thread 7 (Thread 0x7fd37d4bb700 (LWP 114818)): #0 0x00007fd37ed93af3 in pwrite64 () from /lib64/libpthread.so.0 #1 0x00007fd37ff8a892 in vn_rdwr () from /lib64/libzpool.so.2 #2 0x00007fd380006df3 in vdev_file_io_strategy () from /lib64/libzpool.so.2 #3 0x00007fd37ff8bd94 in taskq_thread () from /lib64/libzpool.so.2 #4 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #5 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #6 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 2 (Thread 0x7fd362a6a700 (LWP 116098)): #0 0x00007fd3804d283c in atomic_inc_ulong_nv () from /lib64/libuutil.so.1 #1 0x00007fd37fff300d in spa_open_common () from /lib64/libzpool.so.2 #2 0x00007fd37ffafbc5 in dmu_objset_find () from /lib64/libzpool.so.2 #3 0x0000000000412ceb in ztest_dmu_objset_create_destroy () #4 0x000000000040cda3 in ztest_thread () #5 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #6 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #7 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 Thread 1 (Thread 0x7fd362aab700 (LWP 116082)): #0 0x00007fd37e9f85f7 in raise () from /lib64/libc.so.6 #1 0x00007fd37e9f9ce8 in abort () from /lib64/libc.so.6 #2 0x00000000004089dd in libspl_assert () #3 0x000000000041045d in ztest_io () #4 0x000000000041323f in ztest_dmu_write_parallel () #5 0x000000000040cda3 in ztest_thread () #6 0x00007fd37ff887ac in zk_thread_helper () from /lib64/libzpool.so.2 #7 0x00007fd37ed8cdc5 in start_thread () from /lib64/libpthread.so.0 #8 0x00007fd37eab9ced in clone () from /lib64/libc.so.6 ``` zdb ``` Dataset mos [META], ID 0, cr_txg 4, 562K, 150 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 2 16K 16K 71.5K 512 80.0K 93.75 DMU dnode 1 1 16K 16K 24.0K 512 32K 100.00 object directory 2 1 16K 512 0 512 512 0.00 DSL directory 3 1 16K 512 0 512 512 100.00 DSL props 4 1 16K 512 0 512 512 100.00 DSL directory child map 5 1 16K 512 0 512 512 100.00 zap 6 1 16K 512 0 512 512 100.00 DSL dataset snap map 7 1 16K 512 0 512 512 100.00 DSL deadlist map 8 1 16K 128K 0 512 128K 0.00 bpobj 9 1 16K 4K 12.0K 512 4K 100.00 SPA space map 10 1 16K 4K 12.0K 512 4K 100.00 SPA space map 11 1 16K 4K 12.0K 512 4K 100.00 SPA space map 12 1 16K 4K 12.0K 512 4K 100.00 SPA space map 13 1 16K 4K 12.0K 512 4K 100.00 SPA space map 14 1 16K 4K 12.0K 512 4K 100.00 SPA space map 15 1 16K 512 0 512 512 0.00 DSL directory 16 1 16K 512 0 512 512 100.00 DSL props 17 1 16K 512 0 512 512 100.00 DSL directory child map 18 1 16K 512 0 512 512 100.00 zap 19 1 16K 512 0 512 512 100.00 DSL dataset snap map 20 1 16K 512 0 512 512 100.00 DSL deadlist map 21 2 16K 4K 36.0K 512 8K 100.00 SPA space map 22 2 16K 4K 36.0K 512 8K 100.00 SPA space map 23 2 16K 4K 24.0K 512 4K 100.00 SPA space map 24 1 16K 512 0 512 512 0.00 DSL directory 25 1 16K 512 0 512 512 100.00 DSL props 26 1 16K 512 0 512 512 100.00 DSL directory child map 27 1 16K 512 0 512 512 100.00 zap 28 1 16K 512 0 512 512 100.00 DSL dataset snap map 29 1 16K 512 0 512 512 100.00 DSL deadlist map 30 1 16K 512 0 512 512 0.00 DSL directory 31 1 16K 512 12.0K 512 512 100.00 DSL props 32 1 16K 512 0 512 512 0.00 DSL directory 33 1 16K 512 0 512 512 100.00 DSL props 34 1 16K 1.50K 12.0K 512 1.50K 100.00 DSL directory child map 35 1 16K 512 0 512 512 0.00 DSL directory 36 1 16K 512 0 512 512 100.00 DSL props 37 1 16K 512 0 512 512 100.00 DSL directory child map 38 1 16K 512 0 512 512 0.00 DSL directory 39 1 16K 512 0 512 512 100.00 DSL props 40 1 16K 512 0 512 512 100.00 DSL directory child map 41 1 16K 128K 0 512 128K 0.00 bpobj 42 1 16K 512 0 512 512 0.00 DSL directory 43 1 16K 512 0 512 512 100.00 DSL props 44 1 16K 512 0 512 512 100.00 DSL directory child map 45 1 16K 512 0 512 512 0.00 DSL dataset 46 1 16K 512 0 512 512 100.00 DSL dataset snap map 47 1 16K 512 0 512 512 100.00 DSL deadlist map 48 1 16K 512 0 512 512 0.00 DSL dataset 49 1 16K 512 0 512 512 100.00 DSL deadlist map 50 1 16K 128K 0 512 128K 0.00 bpobj 51 1 16K 512 0 512 512 0.00 DSL dataset 52 1 16K 512 0 512 512 100.00 DSL dataset snap map 53 1 16K 512 0 512 512 100.00 DSL deadlist map 54 1 16K 128K 0 512 128K 0.00 bpobj 55 1 16K 1K 12.0K 512 1K 100.00 DSL dataset next clones 56 1 16K 1K 12.0K 512 1K 100.00 DSL dir clones 57 1 16K 16K 12.0K 512 16K 100.00 packed nvlist 58 1 16K 1K 12.0K 512 1K 100.00 zap 59 1 16K 1K 12.0K 512 1K 100.00 zap 60 1 16K 16K 36.0K 512 32K 100.00 zap 61 1 16K 16K 12.0K 512 16K 100.00 bpobj (Z=uncompressed) 62 1 16K 128K 12.0K 512 128K 100.00 SPA history 63 1 16K 512 0 512 512 100.00 Pool properties 64 1 16K 1K 12.0K 512 1K 100.00 zap 65 1 16K 1K 12.0K 512 1K 100.00 zap 66 1 16K 512 0 512 512 100.00 zap 67 1 16K 512 0 512 512 100.00 zap 68 1 16K 512 0 512 512 100.00 zap 69 1 16K 512 0 512 512 100.00 zap 70 1 16K 512 0 512 512 100.00 object array 71 1 16K 4K 12.0K 512 4K 100.00 SPA space map 72 1 16K 4K 12.0K 512 4K 100.00 SPA space map 73 1 16K 4K 12.0K 512 4K 100.00 SPA space map 74 1 16K 512 0 512 512 100.00 DSL directory child map 75 1 16K 512 0 512 512 100.00 zap 76 1 16K 512 0 512 512 100.00 DSL dataset snap map 77 1 16K 512 0 512 512 100.00 DSL deadlist map 78 1 16K 512 0 512 512 0.00 DSL directory 79 1 16K 512 0 512 512 100.00 DSL props 80 1 16K 512 0 512 512 100.00 DSL directory child map 81 1 16K 512 0 512 512 100.00 zap 82 1 16K 512 0 512 512 100.00 DSL dataset snap map 83 1 16K 512 0 512 512 100.00 DSL deadlist map 84 1 16K 512 0 512 512 0.00 DSL directory 85 1 16K 512 0 512 512 100.00 DSL props 86 1 16K 512 0 512 512 100.00 DSL directory child map 87 1 16K 512 0 512 512 100.00 zap 88 1 16K 512 0 512 512 100.00 DSL dataset snap map 89 1 16K 512 0 512 512 100.00 DSL deadlist map 90 1 16K 512 0 512 512 0.00 DSL directory 91 1 16K 512 0 512 512 100.00 DSL props 92 1 16K 512 0 512 512 100.00 DSL directory child map 93 1 16K 512 0 512 512 100.00 zap 94 1 16K 512 0 512 512 100.00 DSL dataset snap map 95 1 16K 512 0 512 512 100.00 DSL deadlist map 96 1 16K 512 0 512 512 100.00 zap 97 1 16K 512 0 512 512 100.00 DSL deadlist map 98 1 16K 512 0 512 512 0.00 DSL directory 99 1 16K 512 0 512 512 100.00 DSL props 100 1 16K 512 0 512 512 100.00 DSL directory child map 101 1 16K 512 0 512 512 100.00 zap 102 1 16K 512 0 512 512 100.00 DSL dataset snap map 103 1 16K 512 0 512 512 100.00 DSL deadlist map 104 1 16K 512 0 512 512 0.00 DSL directory 105 1 16K 512 0 512 512 100.00 DSL props 106 1 16K 512 0 512 512 100.00 DSL directory child map 107 1 16K 512 0 512 512 100.00 zap 108 1 16K 512 0 512 512 100.00 DSL dataset snap map 109 1 16K 512 0 512 512 100.00 DSL deadlist map 110 1 16K 16K 24.0K 512 32K 100.00 DDT statistics 111 1 4K 4K 12.0K 512 8K 100.00 DDT ZAP algorithm 112 1 16K 512 0 512 512 0.00 DSL directory 113 1 16K 512 0 512 512 100.00 DSL props 114 1 16K 512 0 512 512 100.00 DSL directory child map 115 1 16K 512 0 512 512 100.00 zap 116 1 16K 512 0 512 512 100.00 DSL dataset snap map 117 1 16K 512 0 512 512 100.00 DSL deadlist map 118 1 16K 512 0 512 512 0.00 DSL directory 119 1 16K 512 0 512 512 100.00 DSL props 120 1 16K 512 0 512 512 100.00 DSL directory child map 121 1 16K 512 0 512 512 100.00 zap 122 1 16K 512 0 512 512 100.00 DSL dataset snap map 123 1 16K 512 0 512 512 100.00 DSL deadlist map 124 1 16K 512 0 512 512 100.00 zap 125 1 16K 512 0 512 512 100.00 zap 126 1 16K 512 0 512 512 100.00 zap 127 1 16K 512 0 512 512 100.00 zap 128 1 16K 512 0 512 512 100.00 object array 129 1 16K 4K 12.0K 512 4K 100.00 SPA space map 130 1 16K 4K 12.0K 512 4K 100.00 SPA space map 131 1 16K 4K 12.0K 512 4K 100.00 SPA space map 132 1 16K 512 0 512 512 100.00 zap 133 1 16K 512 0 512 512 100.00 zap 134 1 16K 512 0 512 512 100.00 zap 135 1 16K 512 0 512 512 100.00 zap 136 1 16K 512 0 512 512 0.00 DSL directory 137 1 16K 512 0 512 512 100.00 DSL props 138 1 16K 512 0 512 512 100.00 DSL directory child map 139 1 16K 512 0 512 512 0.00 DSL dataset 140 1 16K 512 0 512 512 100.00 DSL dataset snap map 141 1 16K 512 0 512 512 100.00 DSL deadlist map 142 1 16K 512 0 512 512 100.00 object array 143 1 16K 4K 0 512 4K 100.00 SPA space map 144 1 16K 4K 12.0K 512 4K 100.00 SPA space map 145 1 16K 512 0 512 512 0.00 DSL directory 146 1 16K 512 0 512 512 100.00 DSL props 147 1 16K 512 0 512 512 100.00 DSL directory child map 148 1 16K 512 0 512 512 100.00 zap 149 1 16K 512 0 512 512 100.00 DSL dataset snap map 150 1 16K 512 0 512 512 100.00 DSL deadlist map Dataset ztest/temp_12 [OTHER], ID 115, cr_txg 85, 566K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 4 8K 4K 87.5K 1K 928M 0.00 other uint64[] 4 3 16K 64K 223K 1K 960M 0.02 other uint64[] 6 4 4K 128K 152K 1.50K 304M 0.04 other uint64[] 9 4 8K 8K 32K 512 400M 0.00 other uint64[] Dataset ztest/temp_7 [OTHER], ID 121, cr_txg 87, 195K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 5 8K 512 32K 512 400M 0.00 other uint64[] 3 3 16K 8K 24.0K 2K 32.0M 0.02 other uint64[] 7 4 16K 16K 28.0K 512 976M 0.00 other uint64[] 8 4 16K 16K 40.0K 512 432M 0.00 other uint64[] Dataset ztest/ds_2 [OTHER], ID 27, cr_txg 65, 1000K, 13 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 108K 512 48.0K 13.54 DMU dnode 1 1 16K 1K 12.0K 512 1K 100.00 other ZAP 2 7 8K 8K 271K 512 343T 0.00 other uint64[] 3 6 8K 32K 219K 512 1.01T 0.00 other uint64[] 4 1 16K 1.50K 8K 512 1.50K 100.00 other ZAP 5 2 16K 16K 24.0K 512 32K 100.00 other ZAP 6 5 8K 1K 91.5K 512 784M 0.00 other uint64[] 7 6 4K 512 104K 2K 1008M 0.00 other uint64[] 11 5 4K 2K 36.0K 512 896M 0.00 other uint64[] 12 4 16K 4K 0 1K 544M 0.00 other uint64[] 72 1 16K 512 0 9.00K 512 100.00 other ZAP 90 8 8K 64K 64K 1K 4.32P 0.00 other uint64[] 92 7 4K 16K 52.0K 512 1.62T 0.00 other uint64[] 93 6 16K 2K 0 1K 16.0T 0.00 other uint64[] Dataset ztest/ds_5 [OTHER], ID 87, cr_txg 75, 566K, 6 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 79.5K 512 48.0K 6.25 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 8 16K 128K 439K 1K 4.00E 0.00 other uint64[] 4 1 16K 2K 8K 512 2K 100.00 other ZAP 5 2 16K 16K 24.0K 1K 32K 100.00 other ZAP 32 1 16K 512 0 7.50K 512 100.00 other ZAP 64 1 4K 32K 0 14.5K 32K 0.00 other uint64[] Dataset ztest/temp_4@0 [OTHER], ID 96, cr_txg 77, 71.7K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 1 8K 2K 0 2K 2K 0.00 other uint64[] 6 1 4K 128K 0 2K 128K 0.00 other uint64[] 10 1 16K 512 0 1K 512 0.00 other uint64[] 12 1 16K 32K 0 512 32K 0.00 other uint64[] Dataset ztest/temp_4 [OTHER], ID 93, cr_txg 75, 71.7K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 1 8K 2K 0 2K 2K 0.00 other uint64[] 6 1 4K 128K 0 2K 128K 0.00 other uint64[] 10 1 16K 512 0 1K 512 0.00 other uint64[] 12 1 16K 32K 0 512 32K 0.00 other uint64[] Dataset ztest/ds_1 [OTHER], ID 18, cr_txg 63, 928K, 9 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 87.5K 512 112K 4.02 DMU dnode 1 1 16K 1K 8K 512 1K 100.00 other ZAP 2 2 16K 16K 24.0K 1.50K 32K 100.00 other ZAP 5 1 16K 2.50K 8K 1.50K 2.50K 100.00 other ZAP 8 8 4K 128K 64K 512 834T 0.00 other uint64[] (K=sha256) (Z=gzip-5) 9 7 4K 16K 64K 512 2.24T 0.00 other uint64[] 10 1 4K 128K 4K 1K 128K 100.00 other uint64[] 96 9 8K 64K 638K 11.0K 2.00E 0.00 other uint64[] 139 2 16K 16K 24.0K 6.00K 32K 100.00 other ZAP 208 1 16K 512 0 512 512 100.00 other ZAP Dataset ztest/ds_6 [OTHER], ID 101, cr_txg 77, 295K, 6 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 79.5K 512 64K 4.69 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 3 1 4K 16K 16K 13.0K 16K 100.00 other uint64[] 47 6 16K 128K 95.5K 1K 871T 0.00 other uint64[] (K=fletcher4) (Z=on) 49 6 8K 8K 87.5K 512 1.85T 0.00 other uint64[] 79 1 16K 512 0 2K 512 100.00 other ZAP 97 1 16K 512 0 2K 512 100.00 other ZAP Dataset ztest/temp_9 [OTHER], ID 107, cr_txg 80, 618K, 5 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 15.62 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 6 4K 512 136K 8K 912M 0.00 other uint64[] 18 5 8K 2K 83.5K 512 576M 0.00 other uint64[] 19 4 16K 4K 48.0K 1.50K 720M 0.00 other uint64[] 22 3 16K 128K 279K 1K 848M 0.03 other uint64[] Dataset ztest/ds_4 [OTHER], ID 81, cr_txg 73, 506K, 8 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 87.5K 512 64K 6.25 DMU dnode 1 1 16K 512 8K 512 512 100.00 other ZAP 2 2 16K 16K 24.0K 2K 32K 100.00 other ZAP 32 1 16K 2.50K 8K 14.5K 2.50K 100.00 other ZAP 61 1 8K 512 0 512 512 0.00 other uint64[] 62 8 16K 64K 255K 512 1.00E 0.00 other uint64[] 96 1 16K 512 0 11.5K 512 100.00 other ZAP 119 9 4K 512 64K 512 48.7T 0.00 other uint64[] 120 7 4K 8K 52.0K 2K 2.28T 0.00 other uint64[] Dataset ztest/ds_3 [OTHER], ID 75, cr_txg 69, 1.26M, 17 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 87.5K 512 128K 6.64 DMU dnode 1 1 16K 1.50K 8K 512 1.50K 100.00 other ZAP 2 9 4K 2K 71.5K 2K 81.4T 0.00 other uint64[] 6 5 16K 8K 36.0K 12.0K 977G 0.00 other uint64[] 30 1 4K 64K 79.5K 512 64K 100.00 other uint64[] 31 1 16K 512 0 512 512 100.00 other uint64[] 65 1 16K 512 0 512 512 100.00 other ZAP 66 9 16K 512 415K 1.50K 2.00E 0.00 other uint64[] 224 1 16K 512 0 1K 512 100.00 other ZAP 226 7 8K 32K 295K 512 635T 0.00 other uint64[] (K=fletcher2) (Z=zle) 227 8 4K 1K 239K 512 1.86T 0.00 other uint64[] 228 1 4K 64K 4K 1.50K 64K 100.00 other uint64[] 231 1 4K 4K 0 1K 4K 0.00 other uint64[] 233 1 8K 512 0 2K 512 0.00 other uint64[] 237 1 4K 32K 0 512 32K 0.00 other uint64[] 238 1 16K 64K 0 1.50K 64K 0.00 other uint64[] 241 1 16K 1.50K 12.0K 1K 1.50K 100.00 other ZAP 243 2 16K 16K 36.0K 2K 32K 100.00 other ZAP Dataset ztest/temp_2 [OTHER], ID 139, cr_txg 109, 63.8K, 1 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 56.0K 512 16K 3.12 DMU dnode 1 1 16K 512 0 512 512 100.00 other ZAP Dataset ztest/temp_5 [OTHER], ID 148, cr_txg 116, 586K, 5 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 71.5K 512 32K 7.81 DMU dnode 1 1 16K 512 0 512 512 100.00 other ZAP 8 5 4K 1K 91.5K 9.5K 960M 0.00 other uint64[] 27 5 4K 8K 40.0K 512 528M 0.00 other uint64[] 28 4 4K 64K 104K 2K 480M 0.01 other uint64[] 32 3 16K 64K 271K 512 928M 0.02 other uint64[] Dataset ztest/ds_0 [OTHER], ID 5, cr_txg 7, 1.45M, 14 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 71.5K 512 32K 21.88 DMU dnode 1 1 16K 1K 8K 512 1K 100.00 other ZAP 2 5 8K 512 108K 1K 912M 0.00 other uint64[] 4 4 16K 4K 28.0K 3.00K 384M 0.00 other uint64[] 10 4 4K 64K 108K 1K 864M 0.01 other uint64[] 12 4 4K 64K 0 2K 464M 0.00 other uint64[] 16 4 4K 32K 383K 1K 1008M 0.02 other uint64[] 18 9 16K 8K 622K 512 4.00E 0.00 other uint64[] 19 2 16K 16K 24.0K 2K 32K 100.00 other ZAP 23 8 8K 8K 56.0K 512 595T 0.00 other uint64[] (K=fletcher2) (Z=gzip-7) 24 8 4K 1K 60.0K 512 1.19T 0.00 other uint64[] 25 1 16K 1K 8K 1K 1K 100.00 other ZAP 27 1 4K 16K 0 2K 16K 100.00 other uint64[] 32 1 4K 8K 0 2K 8K 0.00 other uint64[] (K=on) (Z=zle) 36 6 4K 64K 0 512 486G 0.00 other uint64[] Dataset ztest [ZPL], ID 51, cr_txg 1, 7.97K, 0 objects Object lvl iblk dblk dsize dnsize lsize %full type 0 7 16K 16K 0 512 16K 0.00 DMU dnode Verified large_blocks feature refcount of 0 is correct Verified large_dnode feature refcount of 13 is correct Verified sha512 feature refcount of 0 is correct Verified skein feature refcount of 2 is correct Verified edonr feature refcount of 1 is correct Verified userobj_accounting feature refcount of 0 is correct Traversing all blocks to verify checksums and verify nothing leaked ... loading space map for vdev 2 of 3, metaslab 0 of 23 .... No leaks (block sum matches space maps exactly) bp count: 1062 ganged count: 10 bp logical: 14778368 avg: 13915 bp physical: 5126144 avg: 4826 compression: 2.88 bp allocated: 27193344 avg: 25605 compression: 0.54 bp deduped: 0 ref>1: 0 deduplication: 1.00 SPA allocated: 27193344 used: 2.35% additional, non-pointer bps of type 0: 173 Dittoed blocks on same vdev: 610 Blocks LSIZE PSIZE ASIZE avg comp %Total Type - - - - - - - unallocated 2 32K 8K 72.0K 36.0K 4.00 0.27 object directory - - - - - - - object array 1 16K 4K 36.0K 36.0K 4.00 0.14 packed nvlist - - - - - - - packed nvlist size 1 16K 4K 36.0K 36.0K 4.00 0.14 bpobj - - - - - - - bpobj header - - - - - - - SPA space map header 22 124K 84.0K 756K 34.4K 1.48 2.85 SPA space map 5 20.0K 20.0K 60.0K 12.0K 1.00 0.23 ZIL intent log 112 1.75M 480K 3.01M 27.5K 3.73 11.61 DMU dnode 15 30.0K 30.0K 384K 25.6K 1.00 1.45 DMU objset - - - - - - - DSL directory 17 9.5K 1.50K 36.0K 2.12K 6.33 0.14 DSL directory child map - - - - - - - DSL dataset snap map 17 8.50K 512 36.0K 2.12K 17.00 0.14 DSL props - - - - - - - DSL dataset - - - - - - - ZFS znode - - - - - - - ZFS V0 ACL - - - - - - - ZFS plain file - - - - - - - ZFS directory - - - - - - - ZFS master node - - - - - - - ZFS delete queue - - - - - - - zvol object - - - - - - - zvol prop - - - - - - - other uint8[] 743 11.5M 4.12M 20.0M 27.5K 2.78 76.95 other uint64[] 48 360K 103K 972K 20.2K 3.50 3.66 other ZAP - - - - - - - persistent error log 1 128K 4K 36.0K 36.0K 32.00 0.14 SPA history - - - - - - - SPA history offsets - - - - - - - Pool properties - - - - - - - DSL permissions - - - - - - - ZFS ACL - - - - - - - ZFS SYSACL - - - - - - - FUID table - - - - - - - FUID table size 1 1K 1K 36.0K 36.0K 1.00 0.14 DSL dataset next clones - - - - - - - scan work queue - - - - - - - ZFS user/group used - - - - - - - ZFS user/group quota - - - - - - - snapshot refcount tags 2 8K 4K 36.0K 18.0K 2.00 0.14 DDT ZAP algorithm 2 32K 8K 72.0K 36.0K 4.00 0.27 DDT statistics - - - - - - - System attributes - - - - - - - SA master node - - - - - - - SA attr registration - - - - - - - SA attr layouts - - - - - - - scan translations - - - - - - - deduplicated block - - - - - - - DSL deadlist map - - - - - - - DSL deadlist map hdr 1 1K 1K 36.0K 36.0K 1.00 0.14 DSL dir clones - - - - - - - bpobj subobj 5 44.0K 20.0K 180K 36.0K 2.20 0.68 deferred free - - - - - - - dedup ditto 31 48.5K 16K 252K 8.13K 3.03 0.95 other 1.04K 14.1M 4.89M 25.9M 25.0K 2.88 100.00 Total capacity operations bandwidth ---- errors ---- description used avail read write read write read write cksum ztest 25.9M 1.05G 6.58K 0 41.4M 0 0 0 0 raidz2 22.3M 346M 5.49K 0 30.7M 0 0 0 0 /var/tmp/ztest.0a 1.67K 0 9.43M 0 0 0 0 /var/tmp/ztest.1a 1.59K 0 8.94M 0 0 0 34 /var/tmp/ztest.2a 2.23K 0 12.3M 0 0 0 0 raidz2 1.63M 366M 649 0 5.21M 0 0 0 0 /var/tmp/ztest.3a 195 0 1.65M 0 0 0 0 /var/tmp/ztest.4a 195 0 1.65M 0 0 0 0 /var/tmp/ztest.5a 259 0 1.91M 0 0 0 0 raidz2 1.96M 366M 471 0 5.44M 0 0 0 0 /var/tmp/ztest.6a 139 0 1.66M 0 0 0 0 /var/tmp/ztest.7a 139 0 1.66M 0 0 0 0 /var/tmp/ztest.8a 193 0 2.11M 0 0 0 0 ``` We have hit this 4 times recently: zloop-161216-162658/ztest.out:ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] zloop-161217-173015/ztest.out:ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] zloop-161218-091534/ztest.out:ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf] zloop-161217-001943/ztest.out:ASSERT at ztest.c:2499:ztest_io()/usr/sbin/ztest[0x40a7cf]
test
ztest dmu read assert on eio version information distribution name centos distribution version linux kernel architecture zfs version master dec rpm modinfo say spl version master dec rpm modinfo say git build data revision refs remotes origin master built branches refs remotes origin master build of revision refs remotes origin master problem running zloop we hit an assert dmu read zd zd os object offset blocksize data assert at ztest c ztest io usr sbin ztest reproduce sbin ztest vvvvv m r r v a t p s f var tmp sbin ztest vvvvv m r r v a t p s f var tmp log detail ztest out starting main threads sec in ztest verify dnode bt sec in ztest dmu object alloc free sec in ztest split pool sec in ztest zil commit sec in ztest zil commit sec in ztest dmu commit callbacks sec in ztest zap sec in ztest vdev attach detach sec in ztest fault inject sec in ztest spa prop get set sec in ztest dmu read write zcopy sec in ztest vdev attach detach sec in ztest zil commit sec in ztest spa prop get set sec in ztest zap parallel sec in ztest dmu commit callbacks sec in ztest fault inject sec in ztest dmu read write sec in ztest dmu read write sec in ztest zil remount sec in ztest dmu object alloc free sec in ztest dmu commit callbacks sec in ztest spa create destroy sec in ztest zil commit setting dataset ztest temp to sync always sec in ztest dmu object alloc free sec in ztest dmu snapshot create destroy sec in ztest dmu objset create destroy sec in ztest zap sec in ztest split pool setting dataset ztest temp to sync always sec in ztest fzap sec in ztest dmu read write zcopy sec in ztest dmu read write sec in ztest fault inject setting dataset ztest temp to sync always sec in ztest dmu object alloc free sec in ztest dmu read write zcopy sec in ztest spa create destroy sec in ztest zap parallel sec in ztest dmu snapshot create destroy sec in ztest zap parallel sec in ztest spa create destroy sec in ztest dsl prop get set sec in ztest dsl prop get set sec in ztest dsl prop get set sec in ztest dmu commit callbacks sec in ztest dsl prop get set sec in ztest split pool sec in ztest dmu read write sec in ztest dmu object alloc free sec in ztest spa create destroy sec in ztest dmu objset create destroy sec in ztest dmu read write zcopy sec in ztest zap sec in ztest dmu read write sec in ztest dsl prop get set sec in ztest dmu snapshot create destroy sec in ztest dmu snapshot create destroy sec in ztest zil commit sec in ztest dsl prop get set sec in ztest zil remount sec in ztest dmu object alloc free sec in ztest zap sec in ztest dmu read write sec in ztest dmu read write zcopy sec in ztest dmu objset create destroy sec in ztest dmu snapshot hold sec in ztest zil commit unable to expand vdev state error could not expand lun because the vdev configuration changed sec in ztest vdev lun growth sec in ztest dsl prop get set sec in ztest dsl prop get set sec in ztest zap sec in ztest split pool sec in ztest dmu commit callbacks sec in ztest zil commit sec in ztest zil commit sec in ztest zil commit sec in ztest vdev attach detach sec in ztest dmu snapshot hold sec in ztest verify dnode bt sec in ztest zap parallel sec in ztest dsl prop get set sec in ztest spa prop get set sec in ztest zil commit sec in ztest dsl dataset promote busy sec in ztest dmu read write zcopy sec in ztest dmu read write zcopy sec in ztest dsl prop get set sec in ztest dsl prop get set sec in ztest split pool sec in ztest split pool sec in ztest dmu commit callbacks sec in ztest zap sec in ztest fault inject sec in ztest zil commit sec in ztest dmu read write zcopy sec in ztest dmu commit callbacks sec in ztest dsl prop get set sec in ztest dmu snapshot hold sec in ztest dmu snapshot hold sec in ztest dmu commit callbacks sec in ztest zap sec in ztest zil commit sec in ztest fzap sec in ztest dmu snapshot create destroy sec in ztest dmu read write zcopy sec in ztest dmu object alloc free sec in ztest zil commit sec in ztest dmu object alloc free sec in ztest spa create destroy sec in ztest zil remount sec in ztest dmu commit callbacks sec in ztest dmu read write zcopy sec in ztest zap sec in ztest vdev attach detach sec in ztest dmu object alloc free sec in ztest split pool sec in ztest dmu read write dmu read zd zd os object offset blocksize data assert at ztest c ztest io usr sbin ztest libpthread so libc so gsignal libc so abort usr sbin ztest usr sbin ztest usr sbin ztest usr sbin ztest libzpool so zk thread helper libpthread so libc so clone sec in ztest dmu object alloc free child died with signal status using host libthread db library libthread db so core was generated by usr sbin ztest program terminated with signal aborted in raise from libc so backtrace in raise from libc so in abort from libc so in libspl assert in ztest io in ztest dmu write parallel in ztest thread in zk thread helper from libzpool so in start thread from libpthread so in clone from libc so thread thread lwp in pthread rwlock wrlock from libpthread so in ztest spa rename in ztest thread in zk thread helper from libzpool so in start thread from libpthread so in clone from libc so thread thread lwp in from libpthread so in vn rdwr from libzpool so in vdev file io strategy from libzpool so in taskq thread from libzpool so in zk thread helper from libzpool so in start thread from libpthread so in clone from libc so thread thread lwp in pthread join from libpthread so in zk thread join from libzpool so in ztest run in main thread thread lwp in from libpthread so in vn rdwr from libzpool so in vdev file io strategy from libzpool so in taskq thread from libzpool so in zk thread helper from libzpool so in start thread from libpthread so in clone from libc so thread thread lwp in atomic inc ulong nv from libuutil so in spa open common from libzpool so in dmu objset find from libzpool so in ztest dmu objset create destroy in ztest thread in zk thread helper from libzpool so in start thread from libpthread so in clone from libc so thread thread lwp in raise from libc so in abort from libc so in libspl assert in ztest io in ztest dmu write parallel in ztest thread in zk thread helper from libzpool so in start thread from libpthread so in clone from libc so zdb dataset mos id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode object directory dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map bpobj spa space map spa space map spa space map spa space map spa space map spa space map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map spa space map spa space map spa space map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map dsl directory dsl props dsl directory dsl props dsl directory child map dsl directory dsl props dsl directory child map dsl directory dsl props dsl directory child map bpobj dsl directory dsl props dsl directory child map dsl dataset dsl dataset snap map dsl deadlist map dsl dataset dsl deadlist map bpobj dsl dataset dsl dataset snap map dsl deadlist map bpobj dsl dataset next clones dsl dir clones packed nvlist zap zap zap bpobj z uncompressed spa history pool properties zap zap zap zap zap zap object array spa space map spa space map spa space map dsl directory child map zap dsl dataset snap map dsl deadlist map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map zap dsl deadlist map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map ddt statistics ddt zap algorithm dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map zap zap zap zap object array spa space map spa space map spa space map zap zap zap zap dsl directory dsl props dsl directory child map dsl dataset dsl dataset snap map dsl deadlist map object array spa space map spa space map dsl directory dsl props dsl directory child map zap dsl dataset snap map dsl deadlist map dataset ztest temp id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other other dataset ztest temp id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other other dataset ztest ds id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other zap other zap other other other other other zap other other other dataset ztest ds id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other zap other zap other zap other dataset ztest temp id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other other dataset ztest temp id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other other dataset ztest ds id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other zap other zap other k z gzip other other other other zap other zap dataset ztest ds id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other k z on other other zap other zap dataset ztest temp id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other other dataset ztest ds id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other zap other zap other other other zap other other dataset ztest ds id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other other other zap other other zap other k z zle other other other other other other other zap other zap dataset ztest temp id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap dataset ztest temp id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other other dataset ztest ds id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode other zap other other other other other other other zap other k z gzip other other zap other other k on z zle other dataset ztest id cr txg objects object lvl iblk dblk dsize dnsize lsize full type dmu dnode verified large blocks feature refcount of is correct verified large dnode feature refcount of is correct verified feature refcount of is correct verified skein feature refcount of is correct verified edonr feature refcount of is correct verified userobj accounting feature refcount of is correct traversing all blocks to verify checksums and verify nothing leaked loading space map for vdev of metaslab of no leaks block sum matches space maps exactly bp count ganged count bp logical avg bp physical avg compression bp allocated avg compression bp deduped ref deduplication spa allocated used additional non pointer bps of type dittoed blocks on same vdev blocks lsize psize asize avg comp total type unallocated object directory object array packed nvlist packed nvlist size bpobj bpobj header spa space map header spa space map zil intent log dmu dnode dmu objset dsl directory dsl directory child map dsl dataset snap map dsl props dsl dataset zfs znode zfs acl zfs plain file zfs directory zfs master node zfs delete queue zvol object zvol prop other other other zap persistent error log spa history spa history offsets pool properties dsl permissions zfs acl zfs sysacl fuid table fuid table size dsl dataset next clones scan work queue zfs user group used zfs user group quota snapshot refcount tags ddt zap algorithm ddt statistics system attributes sa master node sa attr registration sa attr layouts scan translations deduplicated block dsl deadlist map dsl deadlist map hdr dsl dir clones bpobj subobj deferred free dedup ditto other total capacity operations bandwidth errors description used avail read write read write read write cksum ztest var tmp ztest var tmp ztest var tmp ztest var tmp ztest var tmp ztest var tmp ztest var tmp ztest var tmp ztest var tmp ztest we have hit this times recently zloop ztest out assert at ztest c ztest io usr sbin ztest zloop ztest out assert at ztest c ztest io usr sbin ztest zloop ztest out assert at ztest c ztest io usr sbin ztest zloop ztest out assert at ztest c ztest io usr sbin ztest
1
329,457
28,244,451,750
IssuesEvent
2023-04-06 09:39:26
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_out_warning_linalg_lu_cuda (__main__.TestCommonCUDA)
triaged module: flaky-tests skipped module: unknown
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_out_warning_linalg_lu_cuda&suite=TestCommonCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/9654871304). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 3 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_out_warning_linalg_lu_cuda` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
1.0
DISABLED test_out_warning_linalg_lu_cuda (__main__.TestCommonCUDA) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_out_warning_linalg_lu_cuda&suite=TestCommonCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/9654871304). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 3 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_out_warning_linalg_lu_cuda` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
test
disabled test out warning linalg lu cuda main testcommoncuda platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test out warning linalg lu cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs
1
167,474
13,031,550,712
IssuesEvent
2020-07-28 01:34:48
apache/pulsar
https://api.github.com/repos/apache/pulsar
closed
Intermittent test failure PulsarWorkerAssignmentTest. testFunctionAssignments
component/test flaky-tests type/bug
``` org.apache.pulsar.functions.worker.PulsarWorkerAssignmentTest.testFunctionAssignments Error Message expected [2] but found [1] Stacktrace java.lang.AssertionError: expected [2] but found [1] at org.testng.Assert.fail(Assert.java:96) at org.testng.Assert.failNotEquals(Assert.java:776) at org.testng.Assert.assertEqualsImpl(Assert.java:137) at org.testng.Assert.assertEquals(Assert.java:118) at org.testng.Assert.assertEquals(Assert.java:652) at org.testng.Assert.assertEquals(Assert.java:662) at org.apache.pulsar.functions.worker.PulsarWorkerAssignmentTest.testFunctionAssignments(PulsarWorkerAssignmentTest.java:201) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) at org.testng.internal.InvokeMethodRunnable.runOne(InvokeMethodRunnable.java:54) at org.testng.internal.InvokeMethodRunnable.run(InvokeMethodRunnable.java:44) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ```
2.0
Intermittent test failure PulsarWorkerAssignmentTest. testFunctionAssignments - ``` org.apache.pulsar.functions.worker.PulsarWorkerAssignmentTest.testFunctionAssignments Error Message expected [2] but found [1] Stacktrace java.lang.AssertionError: expected [2] but found [1] at org.testng.Assert.fail(Assert.java:96) at org.testng.Assert.failNotEquals(Assert.java:776) at org.testng.Assert.assertEqualsImpl(Assert.java:137) at org.testng.Assert.assertEquals(Assert.java:118) at org.testng.Assert.assertEquals(Assert.java:652) at org.testng.Assert.assertEquals(Assert.java:662) at org.apache.pulsar.functions.worker.PulsarWorkerAssignmentTest.testFunctionAssignments(PulsarWorkerAssignmentTest.java:201) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) at org.testng.internal.InvokeMethodRunnable.runOne(InvokeMethodRunnable.java:54) at org.testng.internal.InvokeMethodRunnable.run(InvokeMethodRunnable.java:44) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ```
test
intermittent test failure pulsarworkerassignmenttest testfunctionassignments org apache pulsar functions worker pulsarworkerassignmenttest testfunctionassignments error message expected but found stacktrace java lang assertionerror expected but found at org testng assert fail assert java at org testng assert failnotequals assert java at org testng assert assertequalsimpl assert java at org testng assert assertequals assert java at org testng assert assertequals assert java at org testng assert assertequals assert java at org apache pulsar functions worker pulsarworkerassignmenttest testfunctionassignments pulsarworkerassignmenttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org testng internal methodinvocationhelper invokemethod methodinvocationhelper java at org testng internal invokemethodrunnable runone invokemethodrunnable java at org testng internal invokemethodrunnable run invokemethodrunnable java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
1
145,066
11,647,361,636
IssuesEvent
2020-03-01 14:47:11
yuanshenjian/yuanshenjian.github.io
https://api.github.com/repos/yuanshenjian/yuanshenjian.github.io
opened
从另一个角度告诉你单元测试的意义
/unit-test-from-other-sight/ Gitalk
https://yuanshenjian.cn/unit-test-from-other-sight/ 袁慎建, Yuan Shenjian, eXtreme Programming, 极限编程, , , 从另一个角度告诉你单元测试的意义, , eXtreme ProgrammingTDD, 袁慎建的博客, ThoughtWorks, 首页, 博客, 袁慎建的博客, 基本最无敌 三个维度看单元测试 能力建设 生产效率 ...
1.0
从另一个角度告诉你单元测试的意义 - https://yuanshenjian.cn/unit-test-from-other-sight/ 袁慎建, Yuan Shenjian, eXtreme Programming, 极限编程, , , 从另一个角度告诉你单元测试的意义, , eXtreme ProgrammingTDD, 袁慎建的博客, ThoughtWorks, 首页, 博客, 袁慎建的博客, 基本最无敌 三个维度看单元测试 能力建设 生产效率 ...
test
从另一个角度告诉你单元测试的意义 袁慎建 yuan shenjian extreme programming 极限编程 从另一个角度告诉你单元测试的意义 extreme programmingtdd 袁慎建的博客 thoughtworks 首页 博客 袁慎建的博客 基本最无敌 三个维度看单元测试 能力建设 生产效率
1
10,284
4,734,441,619
IssuesEvent
2016-10-19 14:14:56
Krzmbrzl/SQDev
https://api.github.com/repos/Krzmbrzl/SQDev
closed
Create SQDev-perspective
completed in Dev-Build feature
A custom SQDev prespective in order to get rid of all the confusing junk that is not necessary for the plugin and to improve user-friendliness
1.0
Create SQDev-perspective - A custom SQDev prespective in order to get rid of all the confusing junk that is not necessary for the plugin and to improve user-friendliness
non_test
create sqdev perspective a custom sqdev prespective in order to get rid of all the confusing junk that is not necessary for the plugin and to improve user friendliness
0
213,749
16,534,786,674
IssuesEvent
2021-05-27 10:30:11
ubtue/DatenProbleme
https://api.github.com/repos/ubtue/DatenProbleme
closed
ISSN 2196-8020 | Kerygma und Dogma | Summary (V&R Test)
Einspielung_Zotero_AUTO Fehlerquelle: Translator Zotero_AUTO_RSS ready for testing
https://www.vr-elibrary.de/toc/kedo/67/2 Bei den Artikeln mit englischen Abstract bleibt immer der Begriff "Summary" stehen. Da das nächste Wort ohne Blank folgt, ist dies etwas störend. https://www.vr-elibrary.de/doi/10.13109/kedo.2021.67.2.148 ![grafik](https://user-images.githubusercontent.com/29707105/119675524-67483500-be3d-11eb-9afb-541b58716cc7.png) Bei den Artikel mit deutschem Abstract hingegen wurde der Begriff "Zusammenfassung" entfernt. https://www.vr-elibrary.de/doi/10.13109/kedo.2021.67.2.101 ![grafik](https://user-images.githubusercontent.com/29707105/119675808-a6768600-be3d-11eb-8551-4d5787c1fd16.png) Es wäre nicht schlecht, wenn man dies auch für die englischen Abstracts so einrichten könnte.
1.0
ISSN 2196-8020 | Kerygma und Dogma | Summary (V&R Test) - https://www.vr-elibrary.de/toc/kedo/67/2 Bei den Artikeln mit englischen Abstract bleibt immer der Begriff "Summary" stehen. Da das nächste Wort ohne Blank folgt, ist dies etwas störend. https://www.vr-elibrary.de/doi/10.13109/kedo.2021.67.2.148 ![grafik](https://user-images.githubusercontent.com/29707105/119675524-67483500-be3d-11eb-9afb-541b58716cc7.png) Bei den Artikel mit deutschem Abstract hingegen wurde der Begriff "Zusammenfassung" entfernt. https://www.vr-elibrary.de/doi/10.13109/kedo.2021.67.2.101 ![grafik](https://user-images.githubusercontent.com/29707105/119675808-a6768600-be3d-11eb-8551-4d5787c1fd16.png) Es wäre nicht schlecht, wenn man dies auch für die englischen Abstracts so einrichten könnte.
test
issn kerygma und dogma summary v r test bei den artikeln mit englischen abstract bleibt immer der begriff summary stehen da das nächste wort ohne blank folgt ist dies etwas störend bei den artikel mit deutschem abstract hingegen wurde der begriff zusammenfassung entfernt es wäre nicht schlecht wenn man dies auch für die englischen abstracts so einrichten könnte
1
92,306
8,359,167,948
IssuesEvent
2018-10-03 07:13:52
istio/istio
https://api.github.com/repos/istio/istio
closed
Running E2E simple tests with Helm takes 15 minutes extra and doesn't delete namespace
area/environments area/test and release stale
**Describe the bug** When I run `make test e2e_simple_run TAG='0.8.0' E2E_ARGS='--installer=helm'`, the tests keep waiting with this message: ``` Deleting namespace simple-auth-test-2ff8de94cbcf40109b24 ``` It seems like the test code doesn't delete namespace and keeps waiting for 10 minutes and then moves on anyways. **Expected behavior** The tests should delete the namespace and not take so long to finish. **Steps to reproduce the bug** Checkout to `0.8.0` tag and run ``` make test e2e_simple_run TAG='0.8.0' E2E_ARGS='--installer=helm' ``` **Version** Istio: 0.8.0 Kubectl and Kubernetes: 1.10.4 Helm client & server: 2.9.1
1.0
Running E2E simple tests with Helm takes 15 minutes extra and doesn't delete namespace - **Describe the bug** When I run `make test e2e_simple_run TAG='0.8.0' E2E_ARGS='--installer=helm'`, the tests keep waiting with this message: ``` Deleting namespace simple-auth-test-2ff8de94cbcf40109b24 ``` It seems like the test code doesn't delete namespace and keeps waiting for 10 minutes and then moves on anyways. **Expected behavior** The tests should delete the namespace and not take so long to finish. **Steps to reproduce the bug** Checkout to `0.8.0` tag and run ``` make test e2e_simple_run TAG='0.8.0' E2E_ARGS='--installer=helm' ``` **Version** Istio: 0.8.0 Kubectl and Kubernetes: 1.10.4 Helm client & server: 2.9.1
test
running simple tests with helm takes minutes extra and doesn t delete namespace describe the bug when i run make test simple run tag args installer helm the tests keep waiting with this message deleting namespace simple auth test it seems like the test code doesn t delete namespace and keeps waiting for minutes and then moves on anyways expected behavior the tests should delete the namespace and not take so long to finish steps to reproduce the bug checkout to tag and run make test simple run tag args installer helm version istio kubectl and kubernetes helm client server
1
134,242
10,887,534,229
IssuesEvent
2019-11-18 14:43:17
forseti-security/forseti-security
https://api.github.com/repos/forseti-security/forseti-security
closed
Update Inventory CAI export GCS test
module: testing priority: p3 triaged: yes
## Story The [Inventory CAI export GCS test](https://github.com/forseti-security/forseti-security/blob/master/integration_tests/tests/forseti/controls/inventory-cai_gcs_export_file.rb) checks that the CAI export is saved as part of the inventory process. The GCS bucket this checks in is hard-coded. ## Proposed Solution Use the [newly added output](https://github.com/forseti-security/terraform-google-forseti/pull/374) instead of hard-coding `gs://forseti-cai-export-#{suffix}`. ## Acceptance Criteria Existing test passes.
1.0
Update Inventory CAI export GCS test - ## Story The [Inventory CAI export GCS test](https://github.com/forseti-security/forseti-security/blob/master/integration_tests/tests/forseti/controls/inventory-cai_gcs_export_file.rb) checks that the CAI export is saved as part of the inventory process. The GCS bucket this checks in is hard-coded. ## Proposed Solution Use the [newly added output](https://github.com/forseti-security/terraform-google-forseti/pull/374) instead of hard-coding `gs://forseti-cai-export-#{suffix}`. ## Acceptance Criteria Existing test passes.
test
update inventory cai export gcs test story the checks that the cai export is saved as part of the inventory process the gcs bucket this checks in is hard coded proposed solution use the instead of hard coding gs forseti cai export suffix acceptance criteria existing test passes
1
588,047
17,645,898,039
IssuesEvent
2021-08-20 05:59:18
vrtex-team/pytorch_geometric
https://api.github.com/repos/vrtex-team/pytorch_geometric
opened
Fix `negative_sampling` to always return `num_neg_samples`
bug / fix Priority P1
Currently, `torch_geometric.utils.negative_sampling` may return less than `num_neg_samples` due to duplicates in sampling.
1.0
Fix `negative_sampling` to always return `num_neg_samples` - Currently, `torch_geometric.utils.negative_sampling` may return less than `num_neg_samples` due to duplicates in sampling.
non_test
fix negative sampling to always return num neg samples currently torch geometric utils negative sampling may return less than num neg samples due to duplicates in sampling
0
49,693
6,036,815,292
IssuesEvent
2017-06-09 17:03:23
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
kubectl/cmd/util rolling_updater_test.go TestUpdateWithRetries: Use timeout defined in RollingUpdaterConfig
area/kubelet area/test area/test-infra kind/cleanup needs-sig priority/awaiting-more-evidence team/ux (deprecated - do not use)
From a shippable build that failed spuriously for #12168 https://app.shippable.com/builds/55c1b5430e93780b00588c5a ``` ok github.com/GoogleCloudPlatform/kubernetes/pkg/kubectl/cmd/util 1.373s coverage: 32.3% of statements --- FAIL: TestUpdateWithRetries (0.61s) rolling_updater_test.go:958: unexpected error: timed out waiting for the condition rolling_updater_test.go:963: Remaining updates [0xc20814f320] gets [] FAIL coverage: 56.3% of statements FAIL github.com/GoogleCloudPlatform/kubernetes/pkg/kubectl 2.260s ``` In looking at the rolling updater code, it appears to have a whopping 500ms timeout. I'm not sure why we'd expect to be able to contact the API server to update an RC, then contact it again to get the new RC, and do it all within 500 ms on an ongoing reliable basis given network links, etcd, etc. We have evidence that even in unit tests that's not always possible. @ixdy @jlowdermilk
2.0
kubectl/cmd/util rolling_updater_test.go TestUpdateWithRetries: Use timeout defined in RollingUpdaterConfig - From a shippable build that failed spuriously for #12168 https://app.shippable.com/builds/55c1b5430e93780b00588c5a ``` ok github.com/GoogleCloudPlatform/kubernetes/pkg/kubectl/cmd/util 1.373s coverage: 32.3% of statements --- FAIL: TestUpdateWithRetries (0.61s) rolling_updater_test.go:958: unexpected error: timed out waiting for the condition rolling_updater_test.go:963: Remaining updates [0xc20814f320] gets [] FAIL coverage: 56.3% of statements FAIL github.com/GoogleCloudPlatform/kubernetes/pkg/kubectl 2.260s ``` In looking at the rolling updater code, it appears to have a whopping 500ms timeout. I'm not sure why we'd expect to be able to contact the API server to update an RC, then contact it again to get the new RC, and do it all within 500 ms on an ongoing reliable basis given network links, etcd, etc. We have evidence that even in unit tests that's not always possible. @ixdy @jlowdermilk
test
kubectl cmd util rolling updater test go testupdatewithretries use timeout defined in rollingupdaterconfig from a shippable build that failed spuriously for ok github com googlecloudplatform kubernetes pkg kubectl cmd util coverage of statements fail testupdatewithretries rolling updater test go unexpected error timed out waiting for the condition rolling updater test go remaining updates gets fail coverage of statements fail github com googlecloudplatform kubernetes pkg kubectl in looking at the rolling updater code it appears to have a whopping timeout i m not sure why we d expect to be able to contact the api server to update an rc then contact it again to get the new rc and do it all within ms on an ongoing reliable basis given network links etcd etc we have evidence that even in unit tests that s not always possible ixdy jlowdermilk
1
259,665
22,504,583,202
IssuesEvent
2022-06-23 14:31:19
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
opened
Teste de generalizacao para a tag Terceiro Setor - Dados de Parcerias - Belmiro Braga
generalization test development
DoD: Realizar o teste de Generalização do validador da tag Terceiro Setor - Dados de Parcerias para o Município de Belmiro Braga.
1.0
Teste de generalizacao para a tag Terceiro Setor - Dados de Parcerias - Belmiro Braga - DoD: Realizar o teste de Generalização do validador da tag Terceiro Setor - Dados de Parcerias para o Município de Belmiro Braga.
test
teste de generalizacao para a tag terceiro setor dados de parcerias belmiro braga dod realizar o teste de generalização do validador da tag terceiro setor dados de parcerias para o município de belmiro braga
1
186,783
14,408,452,142
IssuesEvent
2020-12-03 23:49:53
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Use test_user with minimum privileges for functional ui tests under rollup_job folder
automation test_ui_functional test_xpack_functional
Meta issue https://github.com/elastic/kibana/issues/60815 - [ ] x-pack/test/functional/apps/rollup_job/hybrid_index_helper.js - [ ] x-pack/test/functional/apps/rollup_job/hybrid_index_pattern.js - [ ] x-pack/test/functional/apps/rollup_job/tsvb.js similar to https://github.com/elastic/kibana/pull/79567 cc @LeeDr
2.0
Use test_user with minimum privileges for functional ui tests under rollup_job folder - Meta issue https://github.com/elastic/kibana/issues/60815 - [ ] x-pack/test/functional/apps/rollup_job/hybrid_index_helper.js - [ ] x-pack/test/functional/apps/rollup_job/hybrid_index_pattern.js - [ ] x-pack/test/functional/apps/rollup_job/tsvb.js similar to https://github.com/elastic/kibana/pull/79567 cc @LeeDr
test
use test user with minimum privileges for functional ui tests under rollup job folder meta issue x pack test functional apps rollup job hybrid index helper js x pack test functional apps rollup job hybrid index pattern js x pack test functional apps rollup job tsvb js similar to cc leedr
1
562,603
16,664,543,781
IssuesEvent
2021-06-06 23:23:06
drashland/deno-drash-middleware
https://api.github.com/repos/drashland/deno-drash-middleware
opened
Add a GraphQL Middleware
Priority: Low Type: Enhancement
## Summary What: Add a middleware that can add as a graphql server, to intercept requests to make queries from, and serve a graphql playground Why: Graphql is very popular nowadays, having it as a middleware for drash would be a great tool for users, should they wish to use graphql, they can just pipe this middleware into their drash app, and away they go :) ## Acceptance Criteria Below is a list of tasks that must be completed before this issue can be closed. - [ ] Write documentation - [ ] Write unit tests - [ ] Write integration tests ## Example Pseudo Code (for implementation) ```typescript // Add example pseudo code for implementation ```
1.0
Add a GraphQL Middleware - ## Summary What: Add a middleware that can add as a graphql server, to intercept requests to make queries from, and serve a graphql playground Why: Graphql is very popular nowadays, having it as a middleware for drash would be a great tool for users, should they wish to use graphql, they can just pipe this middleware into their drash app, and away they go :) ## Acceptance Criteria Below is a list of tasks that must be completed before this issue can be closed. - [ ] Write documentation - [ ] Write unit tests - [ ] Write integration tests ## Example Pseudo Code (for implementation) ```typescript // Add example pseudo code for implementation ```
non_test
add a graphql middleware summary what add a middleware that can add as a graphql server to intercept requests to make queries from and serve a graphql playground why graphql is very popular nowadays having it as a middleware for drash would be a great tool for users should they wish to use graphql they can just pipe this middleware into their drash app and away they go acceptance criteria below is a list of tasks that must be completed before this issue can be closed write documentation write unit tests write integration tests example pseudo code for implementation typescript add example pseudo code for implementation
0
272,571
23,682,013,446
IssuesEvent
2022-08-28 23:14:13
Iridium-Development/IridiumSkyblock
https://api.github.com/repos/Iridium-Development/IridiumSkyblock
closed
Bungeecord breaks /is visit
bug unconfirmed needs testing
**Describe the bug** You can't teleport to islands of offline players when using Bungeecord. **To Reproduce** 1. Run `/is visit <name of an offline player>` 2. "No island found for that player" **Expected behavior** You should be teleported to the island. **Server Version and Plugin Version** IridiumSkyblock version: 3.1.4 Server version: git-Paper-349 (MC: 1.17.1) **Additional context** Might be related to Bungeecord, but it is set up correctly (online mode Bungeecord, offline mode server, `settings.bungeecord=true` in the `spigot.yml`).
1.0
Bungeecord breaks /is visit - **Describe the bug** You can't teleport to islands of offline players when using Bungeecord. **To Reproduce** 1. Run `/is visit <name of an offline player>` 2. "No island found for that player" **Expected behavior** You should be teleported to the island. **Server Version and Plugin Version** IridiumSkyblock version: 3.1.4 Server version: git-Paper-349 (MC: 1.17.1) **Additional context** Might be related to Bungeecord, but it is set up correctly (online mode Bungeecord, offline mode server, `settings.bungeecord=true` in the `spigot.yml`).
test
bungeecord breaks is visit describe the bug you can t teleport to islands of offline players when using bungeecord to reproduce run is visit no island found for that player expected behavior you should be teleported to the island server version and plugin version iridiumskyblock version server version git paper mc additional context might be related to bungeecord but it is set up correctly online mode bungeecord offline mode server settings bungeecord true in the spigot yml
1
63,822
18,011,429,467
IssuesEvent
2021-09-16 09:03:21
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
ORDER BY in SQL not ordered
Type: Defect Source: Internal Team: SQL
Using `Hazelcast Platform 5.0-SNAPSHOT (20210915 - 645a193)` If I run ``` SELECT * FROM "m" ORDER BY 'count' ``` results are not in order. Reproducer ``` package neil; import com.hazelcast.config.Config; import com.hazelcast.core.Hazelcast; import com.hazelcast.core.HazelcastInstance; import com.hazelcast.core.HazelcastJsonValue; import com.hazelcast.map.IMap; import com.hazelcast.sql.SqlResult; public class Application { public static void main(String[] args) throws Exception { Config config = new Config(); config.getJetConfig().setEnabled(true); HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config); IMap<Long, HazelcastJsonValue> map = hazelcastInstance.getMap("m"); String mapping = "CREATE OR REPLACE MAPPING \"" + map.getName() + "\"" + " ( " + " __key BIGINT," + " \"count\" BIGINT" + " ) " + "TYPE IMap " + " OPTIONS ( " + " 'keyFormat' = 'java'," + " 'keyJavaClass' = '" + Long.class.getCanonicalName() + "'," + " 'valueFormat' = 'json-flat'," + " 'valueJavaClass' = '" + HazelcastJsonValue.class.getCanonicalName() + "'" + " )"; System.out.println(mapping); hazelcastInstance.getSql().execute(mapping); for (int i = 0; i < 10; i++) { Long key = Long.valueOf("" + i); HazelcastJsonValue value = new HazelcastJsonValue("{" + " \"count\": " + i + "}"); map.set(key, value); } String query = "SELECT * FROM \"" + map.getName() + "\" ORDER BY 'count'"; System.out.println(query); SqlResult result = hazelcastInstance.getSql().execute(query); result.forEach(System.out::println); hazelcastInstance.shutdown(); } } ``` Sample output ``` [__key BIGINT=2, count BIGINT=2] [__key BIGINT=5, count BIGINT=5] [__key BIGINT=8, count BIGINT=8] [__key BIGINT=4, count BIGINT=4] [__key BIGINT=6, count BIGINT=6] [__key BIGINT=7, count BIGINT=7] [__key BIGINT=3, count BIGINT=3] [__key BIGINT=1, count BIGINT=1] [__key BIGINT=0, count BIGINT=0] [__key BIGINT=9, count BIGINT=9] ```
1.0
ORDER BY in SQL not ordered - Using `Hazelcast Platform 5.0-SNAPSHOT (20210915 - 645a193)` If I run ``` SELECT * FROM "m" ORDER BY 'count' ``` results are not in order. Reproducer ``` package neil; import com.hazelcast.config.Config; import com.hazelcast.core.Hazelcast; import com.hazelcast.core.HazelcastInstance; import com.hazelcast.core.HazelcastJsonValue; import com.hazelcast.map.IMap; import com.hazelcast.sql.SqlResult; public class Application { public static void main(String[] args) throws Exception { Config config = new Config(); config.getJetConfig().setEnabled(true); HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config); IMap<Long, HazelcastJsonValue> map = hazelcastInstance.getMap("m"); String mapping = "CREATE OR REPLACE MAPPING \"" + map.getName() + "\"" + " ( " + " __key BIGINT," + " \"count\" BIGINT" + " ) " + "TYPE IMap " + " OPTIONS ( " + " 'keyFormat' = 'java'," + " 'keyJavaClass' = '" + Long.class.getCanonicalName() + "'," + " 'valueFormat' = 'json-flat'," + " 'valueJavaClass' = '" + HazelcastJsonValue.class.getCanonicalName() + "'" + " )"; System.out.println(mapping); hazelcastInstance.getSql().execute(mapping); for (int i = 0; i < 10; i++) { Long key = Long.valueOf("" + i); HazelcastJsonValue value = new HazelcastJsonValue("{" + " \"count\": " + i + "}"); map.set(key, value); } String query = "SELECT * FROM \"" + map.getName() + "\" ORDER BY 'count'"; System.out.println(query); SqlResult result = hazelcastInstance.getSql().execute(query); result.forEach(System.out::println); hazelcastInstance.shutdown(); } } ``` Sample output ``` [__key BIGINT=2, count BIGINT=2] [__key BIGINT=5, count BIGINT=5] [__key BIGINT=8, count BIGINT=8] [__key BIGINT=4, count BIGINT=4] [__key BIGINT=6, count BIGINT=6] [__key BIGINT=7, count BIGINT=7] [__key BIGINT=3, count BIGINT=3] [__key BIGINT=1, count BIGINT=1] [__key BIGINT=0, count BIGINT=0] [__key BIGINT=9, count BIGINT=9] ```
non_test
order by in sql not ordered using hazelcast platform snapshot if i run select from m order by count results are not in order reproducer package neil import com hazelcast config config import com hazelcast core hazelcast import com hazelcast core hazelcastinstance import com hazelcast core hazelcastjsonvalue import com hazelcast map imap import com hazelcast sql sqlresult public class application public static void main string args throws exception config config new config config getjetconfig setenabled true hazelcastinstance hazelcastinstance hazelcast newhazelcastinstance config imap map hazelcastinstance getmap m string mapping create or replace mapping map getname key bigint count bigint type imap options keyformat java keyjavaclass long class getcanonicalname valueformat json flat valuejavaclass hazelcastjsonvalue class getcanonicalname system out println mapping hazelcastinstance getsql execute mapping for int i i i long key long valueof i hazelcastjsonvalue value new hazelcastjsonvalue count i map set key value string query select from map getname order by count system out println query sqlresult result hazelcastinstance getsql execute query result foreach system out println hazelcastinstance shutdown sample output
0
169,127
13,125,435,787
IssuesEvent
2020-08-06 06:40:02
kyma-project/kyma
https://api.github.com/repos/kyma-project/kyma
opened
Fix Token Request creation failure in the integration tests
area/application-connector test-failing
**Description** `application-connector` and `connection-token-handler` are failing sometimes due to Token Request creation failure. It seems like not only these tests are failing because of this issue and it seems like it may be a problem with the Connection Token Handler component If you see them failing, add the log in the comment below, please. It'd be super helpful for further investigation. You can even add a suggestion on what could have happened or how it may be fixed **Expected result** Token Requests are created correctly and the tests are not failing **Actual result** `application-connector` tests logs: ``` suite.go:186: Waiting for Info URL in Token Request... require.go:794: Error Trace: suite.go:190 suite.go:196 application_access_test.go:27 Error: Received unexpected error: timeout waiting for condition Test: TestApplicationAccess ``` `connection-token-handler` tests logs: ``` --- FAIL: TestTokenRequests/should_create_token_request_CR_with_token_when_tenant_and_group_provided (25.05s) require.go:1159: Error Trace: k8sresource_test.go:79 k8sresource_test.go:59 Error: Should be true Test: TestTokenRequests/should_create_token_request_CR_with_token_when_tenant_and_group_provided ``` **Logs*** - https://status.build.kyma-project.io/view/gcs/kyma-prow-logs/logs/post-master-kyma-gke-central-connector/1290977690907578368 - https://status.build.kyma-project.io/view/gcs/kyma-prow-logs/logs/kyma-integration-gardener-aws/1290286539225436160
1.0
Fix Token Request creation failure in the integration tests - **Description** `application-connector` and `connection-token-handler` are failing sometimes due to Token Request creation failure. It seems like not only these tests are failing because of this issue and it seems like it may be a problem with the Connection Token Handler component If you see them failing, add the log in the comment below, please. It'd be super helpful for further investigation. You can even add a suggestion on what could have happened or how it may be fixed **Expected result** Token Requests are created correctly and the tests are not failing **Actual result** `application-connector` tests logs: ``` suite.go:186: Waiting for Info URL in Token Request... require.go:794: Error Trace: suite.go:190 suite.go:196 application_access_test.go:27 Error: Received unexpected error: timeout waiting for condition Test: TestApplicationAccess ``` `connection-token-handler` tests logs: ``` --- FAIL: TestTokenRequests/should_create_token_request_CR_with_token_when_tenant_and_group_provided (25.05s) require.go:1159: Error Trace: k8sresource_test.go:79 k8sresource_test.go:59 Error: Should be true Test: TestTokenRequests/should_create_token_request_CR_with_token_when_tenant_and_group_provided ``` **Logs*** - https://status.build.kyma-project.io/view/gcs/kyma-prow-logs/logs/post-master-kyma-gke-central-connector/1290977690907578368 - https://status.build.kyma-project.io/view/gcs/kyma-prow-logs/logs/kyma-integration-gardener-aws/1290286539225436160
test
fix token request creation failure in the integration tests description application connector and connection token handler are failing sometimes due to token request creation failure it seems like not only these tests are failing because of this issue and it seems like it may be a problem with the connection token handler component if you see them failing add the log in the comment below please it d be super helpful for further investigation you can even add a suggestion on what could have happened or how it may be fixed expected result token requests are created correctly and the tests are not failing actual result application connector tests logs suite go waiting for info url in token request require go error trace suite go suite go application access test go error received unexpected error timeout waiting for condition test testapplicationaccess connection token handler tests logs fail testtokenrequests should create token request cr with token when tenant and group provided require go error trace test go test go error should be true test testtokenrequests should create token request cr with token when tenant and group provided logs
1
67,742
7,059,253,672
IssuesEvent
2018-01-05 00:21:29
nodejs/node
https://api.github.com/repos/nodejs/node
closed
OSX: sequential/test-http-writable-true-after-close
CI / flaky test http macos test
* **Version**: master * **Platform**: OSX * **Subsystem**: http https://ci.nodejs.org/job/node-test-commit-osx/13264/nodes=osx1010/console ``` not ok 1960 sequential/test-http-writable-true-after-close --- duration_ms: 0.138 severity: fail stack: |- Mismatched <anonymous> function calls. Expected exactly 1, actual 0. at Object.exports.mustCall (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/common/index.js:488:10) at ClientRequest.common.mustCall (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/sequential/test-http-writable-true-after-close.js:16:29) at ClientRequest.<anonymous> (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/common/index.js:520:15) at Object.onceWrapper (events.js:333:30) at emitOne (events.js:126:13) at ClientRequest.emit (events.js:224:7) at HTTPParser.parserOnIncomingClient [as onIncoming] (_http_client.js:550:21) at HTTPParser.parserOnHeadersComplete (_http_common.js:117:23) at Socket.socketOnData (_http_client.js:439:20) ``` Failed in CI run for test change that was not related.
2.0
OSX: sequential/test-http-writable-true-after-close - * **Version**: master * **Platform**: OSX * **Subsystem**: http https://ci.nodejs.org/job/node-test-commit-osx/13264/nodes=osx1010/console ``` not ok 1960 sequential/test-http-writable-true-after-close --- duration_ms: 0.138 severity: fail stack: |- Mismatched <anonymous> function calls. Expected exactly 1, actual 0. at Object.exports.mustCall (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/common/index.js:488:10) at ClientRequest.common.mustCall (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/sequential/test-http-writable-true-after-close.js:16:29) at ClientRequest.<anonymous> (/Users/iojs/build/workspace/node-test-commit-osx/nodes/osx1010/test/common/index.js:520:15) at Object.onceWrapper (events.js:333:30) at emitOne (events.js:126:13) at ClientRequest.emit (events.js:224:7) at HTTPParser.parserOnIncomingClient [as onIncoming] (_http_client.js:550:21) at HTTPParser.parserOnHeadersComplete (_http_common.js:117:23) at Socket.socketOnData (_http_client.js:439:20) ``` Failed in CI run for test change that was not related.
test
osx sequential test http writable true after close version master platform osx subsystem http not ok sequential test http writable true after close duration ms severity fail stack mismatched function calls expected exactly actual at object exports mustcall users iojs build workspace node test commit osx nodes test common index js at clientrequest common mustcall users iojs build workspace node test commit osx nodes test sequential test http writable true after close js at clientrequest users iojs build workspace node test commit osx nodes test common index js at object oncewrapper events js at emitone events js at clientrequest emit events js at httpparser parseronincomingclient http client js at httpparser parseronheaderscomplete http common js at socket socketondata http client js failed in ci run for test change that was not related
1
154,829
12,230,788,292
IssuesEvent
2020-05-04 06:05:03
linkedpipes/dcat-ap-forms
https://api.github.com/repos/linkedpipes/dcat-ap-forms
closed
Misleading title: "Editace datové sady"
bug test
![image](https://user-images.githubusercontent.com/718827/80570413-02809800-89fb-11ea-88b3-8250370ba91a.png) This is displayed even when creating a new dataset, which may be misleading. This should be "Nová datová sada v NKOD", unless it actually is edit (related to #64 )
1.0
Misleading title: "Editace datové sady" - ![image](https://user-images.githubusercontent.com/718827/80570413-02809800-89fb-11ea-88b3-8250370ba91a.png) This is displayed even when creating a new dataset, which may be misleading. This should be "Nová datová sada v NKOD", unless it actually is edit (related to #64 )
test
misleading title editace datové sady this is displayed even when creating a new dataset which may be misleading this should be nová datová sada v nkod unless it actually is edit related to
1
291,401
25,144,022,469
IssuesEvent
2022-11-10 02:42:34
catalyst-cooperative/rmi-ferc1-eia
https://api.github.com/repos/catalyst-cooperative/rmi-ferc1-eia
closed
enable 1-year of data processing
rmi tests
In order to enable github action testing (bc everything is too beeeg). Run through each processing step, diagnose current issues blocking 1-year processing and implement solution: - [x] `pudl_rmi.coordinate.plant_parts_eia` (this one almost certainly already fine) - [ ] `pudl_rmi.coordinate.deprish`: this one literally has no `year` arg's rn - [ ] `pudl_rmi.coordinate.deprish_to_eia`: probably the main issue is in the overrides - [ ] `pudl_rmi.coordinate.ferc1_to_eia`: I suspect we'll need to add masks for the input data (or maybe that'll be handled w/ `pudl_out`) and the training data - [ ] `pudl_rmi.coordinate.deprish_to_ferc1`: I thiiink if everything upstream is restricted to a year this should be okay.
1.0
enable 1-year of data processing - In order to enable github action testing (bc everything is too beeeg). Run through each processing step, diagnose current issues blocking 1-year processing and implement solution: - [x] `pudl_rmi.coordinate.plant_parts_eia` (this one almost certainly already fine) - [ ] `pudl_rmi.coordinate.deprish`: this one literally has no `year` arg's rn - [ ] `pudl_rmi.coordinate.deprish_to_eia`: probably the main issue is in the overrides - [ ] `pudl_rmi.coordinate.ferc1_to_eia`: I suspect we'll need to add masks for the input data (or maybe that'll be handled w/ `pudl_out`) and the training data - [ ] `pudl_rmi.coordinate.deprish_to_ferc1`: I thiiink if everything upstream is restricted to a year this should be okay.
test
enable year of data processing in order to enable github action testing bc everything is too beeeg run through each processing step diagnose current issues blocking year processing and implement solution pudl rmi coordinate plant parts eia this one almost certainly already fine pudl rmi coordinate deprish this one literally has no year arg s rn pudl rmi coordinate deprish to eia probably the main issue is in the overrides pudl rmi coordinate to eia i suspect we ll need to add masks for the input data or maybe that ll be handled w pudl out and the training data pudl rmi coordinate deprish to i thiiink if everything upstream is restricted to a year this should be okay
1