Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
26,136
4,593,627,316
IssuesEvent
2016-09-21 02:06:33
afisher1/GridLAB-D
https://api.github.com/repos/afisher1/GridLAB-D
opened
#63 add_module and add_class do not work properly,
defect
The new module needs to be listed in SRCDIRS ,
1.0
#63 add_module and add_class do not work properly, - The new module needs to be listed in SRCDIRS ,
defect
add module and add class do not work properly the new module needs to be listed in srcdirs
1
5,826
2,610,216,204
IssuesEvent
2015-02-26 19:08:55
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
сенина ответы 2011
auto-migrated Priority-Medium Type-Defect
``` '''Валентин Романов''' Привет всем не подскажите где можно найти .сенина ответы 2011. где то видел уже '''Валерий Беляев''' Вот держи линк http://bit.ly/1csOg26 '''Владислав Рогов''' Спасибо вроде то но просит телефон вводить '''Алан Козлов''' Не это не влияет на баланс '''Георгий Меркушев''' Не это не влияет на баланс Информация о файле: сенина ответы 2011 Загружен: В этом месяце Скачан раз: 838 Рейтинг: 1111 Средняя скорость скачивания: 839 Похожих файлов: 19 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 6:31
1.0
сенина ответы 2011 - ``` '''Валентин Романов''' Привет всем не подскажите где можно найти .сенина ответы 2011. где то видел уже '''Валерий Беляев''' Вот держи линк http://bit.ly/1csOg26 '''Владислав Рогов''' Спасибо вроде то но просит телефон вводить '''Алан Козлов''' Не это не влияет на баланс '''Георгий Меркушев''' Не это не влияет на баланс Информация о файле: сенина ответы 2011 Загружен: В этом месяце Скачан раз: 838 Рейтинг: 1111 Средняя скорость скачивания: 839 Похожих файлов: 19 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 6:31
defect
сенина ответы валентин романов привет всем не подскажите где можно найти сенина ответы где то видел уже валерий беляев вот держи линк владислав рогов спасибо вроде то но просит телефон вводить алан козлов не это не влияет на баланс георгий меркушев не это не влияет на баланс информация о файле сенина ответы загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
97,452
20,263,954,860
IssuesEvent
2022-02-15 10:15:09
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] com_contact email
No Code Attached Yet
The new form-control css -classes are really great. They give us the possibility to improve the UX significantly. Unfortunately there is a small problem with the email-field. Even if the input is not a valid email address the field gets the css-classes `class="form-control validate-email required valid form-control-success"`
1.0
[4.0] com_contact email - The new form-control css -classes are really great. They give us the possibility to improve the UX significantly. Unfortunately there is a small problem with the email-field. Even if the input is not a valid email address the field gets the css-classes `class="form-control validate-email required valid form-control-success"`
non_defect
com contact email the new form control css classes are really great they give us the possibility to improve the ux significantly unfortunately there is a small problem with the email field even if the input is not a valid email address the field gets the css classes class form control validate email required valid form control success
0
282,207
30,889,222,804
IssuesEvent
2023-08-04 02:25:02
madhans23/linux-4.1.15
https://api.github.com/repos/madhans23/linux-4.1.15
reopened
CVE-2016-6480 (Medium) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2016-6480 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/aacraid/commctrl.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/aacraid/commctrl.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Race condition in the ioctl_send_fib function in drivers/scsi/aacraid/commctrl.c in the Linux kernel through 4.7 allows local users to cause a denial of service (out-of-bounds access or system crash) by changing a certain size value, aka a "double fetch" vulnerability. <p>Publish Date: 2016-08-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6480>CVE-2016-6480</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6480">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6480</a></p> <p>Release Date: 2016-08-06</p> <p>Fix Resolution: v4.8-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-6480 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2016-6480 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/aacraid/commctrl.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/aacraid/commctrl.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Race condition in the ioctl_send_fib function in drivers/scsi/aacraid/commctrl.c in the Linux kernel through 4.7 allows local users to cause a denial of service (out-of-bounds access or system crash) by changing a certain size value, aka a "double fetch" vulnerability. <p>Publish Date: 2016-08-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6480>CVE-2016-6480</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6480">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6480</a></p> <p>Release Date: 2016-08-06</p> <p>Fix Resolution: v4.8-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files drivers scsi aacraid commctrl c drivers scsi aacraid commctrl c vulnerability details race condition in the ioctl send fib function in drivers scsi aacraid commctrl c in the linux kernel through allows local users to cause a denial of service out of bounds access or system crash by changing a certain size value aka a double fetch vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
111,600
14,112,381,316
IssuesEvent
2020-11-07 05:00:10
shibafuku/fukulog
https://api.github.com/repos/shibafuku/fukulog
opened
コンテンツページのワイヤーフレーム作成
design
各コンテンツ(mornibg、lunch、denner、yado)ページのワイヤーフレームは同じになるので ・コンテンツクリック後のページ ・店名クリック後のページ の2種類を手書きしたものをスキャンして作成する
1.0
コンテンツページのワイヤーフレーム作成 - 各コンテンツ(mornibg、lunch、denner、yado)ページのワイヤーフレームは同じになるので ・コンテンツクリック後のページ ・店名クリック後のページ の2種類を手書きしたものをスキャンして作成する
non_defect
コンテンツページのワイヤーフレーム作成 各コンテンツ(mornibg、lunch、denner、yado)ページのワイヤーフレームは同じになるので ・コンテンツクリック後のページ ・店名クリック後のページ
0
47,530
13,056,224,244
IssuesEvent
2020-07-30 04:02:38
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
port photonics_1.73 does not build on SL5 64 bit (Trac #686)
Migrated from Trac booking defect
Scientific Linux release 5.8 (Boron) ```text $ gcc --version gcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52) Copyright (C) 2006 Free Software Foundation, Inc. This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. ``` ```text $ ./bin/port -vd install photonics_1.73 DEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73 DEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73 DEBUG: Requested variant x86_64 is not provided by port photonics_1.73. DEBUG: Executing variant linux provides linux DEBUG: Executing com.apple.main (photonics_1.73) DEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz DEBUG: Skipping unarchive (photonics_1.73) since no archive found DEBUG: Skipping completed com.apple.unarchive (photonics_1.73) ---> Fetching photonics_1.73 DEBUG: Executing com.apple.fetch (photonics_1.73) ---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73 ---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2' % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M ---> Verifying checksum(s) for photonics_1.73 DEBUG: Executing com.apple.checksum (photonics_1.73) ---> Checksumming photonics-1.73.tar.bz2 DEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2 DEBUG: setting option extract.cmd to /usr/bin/bzip2 ---> Extracting photonics_1.73 DEBUG: Executing com.apple.extract (photonics_1.73) ---> Extracting photonics-1.73.tar.bz2 DEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -' DEBUG: Executing com.apple.patch (photonics_1.73) ---> Configuring photonics_1.73 DEBUG: Executing com.apple.configure (photonics_1.73) DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize' checking for a BSD-compatible install... /usr/bin/install -c checking whether build environment is sane... yes checking for a thread-safe mkdir -p... /bin/mkdir -p checking for gawk... gawk checking whether make sets $(MAKE)... yes checking for gawk... (cached) gawk checking for gcc... gcc checking for C compiler default output file name... a.out checking whether the C compiler works... yes checking whether we are cross compiling... no checking for suffix of executables... checking for suffix of object files... o checking whether we are using the GNU C compiler... yes checking whether gcc accepts -g... yes checking for gcc option to accept ISO C89... none needed checking for style of include used by make... GNU checking dependency style of gcc... gcc3 checking how to run the C preprocessor... gcc -E checking for g++... g++ checking whether we are using the GNU C++ compiler... yes checking whether g++ accepts -g... yes checking dependency style of g++... gcc3 checking how to run the C++ preprocessor... g++ -E checking for a BSD-compatible install... /usr/bin/install -c checking whether ln -s works... yes checking whether make sets $(MAKE)... (cached) yes checking build system type... x86_64-unknown-linux-gnu checking host system type... x86_64-unknown-linux-gnu checking for a sed that does not truncate output... /bin/sed checking for grep that handles long lines and -e... /bin/grep checking for egrep... /bin/grep -E checking for fgrep... /bin/grep -F checking for ld used by gcc... /usr/bin/ld checking if the linker (/usr/bin/ld) is GNU ld... yes checking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B checking the name lister (/usr/bin/nm -B) interface... BSD nm checking the maximum length of command line arguments... 98304 checking whether the shell understands some XSI constructs... yes checking whether the shell understands "+="... yes checking for /usr/bin/ld option to reload object files... -r checking how to recognize dependent libraries... pass_all checking for ar... ar checking for strip... strip checking for ranlib... ranlib checking command to parse /usr/bin/nm -B output from gcc object... ok checking for ANSI C header files... yes checking for sys/types.h... yes checking for sys/stat.h... yes checking for stdlib.h... yes checking for string.h... yes checking for memory.h... yes checking for strings.h... yes checking for inttypes.h... yes checking for stdint.h... yes checking for unistd.h... yes checking for dlfcn.h... yes checking whether we are using the GNU C++ compiler... (cached) yes checking whether g++ accepts -g... (cached) yes checking dependency style of g++... (cached) gcc3 checking how to run the C++ preprocessor... g++ -E checking for objdir... .libs checking if gcc supports -fno-rtti -fno-exceptions... no checking for gcc option to produce PIC... -fPIC -DPIC checking if gcc PIC flag -fPIC -DPIC works... yes checking if gcc static flag -static works... yes checking if gcc supports -c -o file.o... yes checking if gcc supports -c -o file.o... (cached) yes checking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes checking whether -lc should be explicitly linked in... no checking dynamic linker characteristics... GNU/Linux ld.so checking how to hardcode library paths into programs... immediate checking whether stripping libraries is possible... yes checking if libtool supports shared libraries... yes checking whether to build shared libraries... yes checking whether to build static libraries... yes checking for ld used by g++... /usr/bin/ld -m elf_x86_64 checking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes checking for g++ option to produce PIC... -fPIC -DPIC checking if g++ PIC flag -fPIC -DPIC works... yes checking if g++ static flag -static works... yes checking if g++ supports -c -o file.o... yes checking if g++ supports -c -o file.o... (cached) yes checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes checking dynamic linker characteristics... GNU/Linux ld.so checking how to hardcode library paths into programs... immediate checking for ANSI C header files... (cached) yes checking for stdbool.h that conforms to C99... yes checking for _Bool... yes checking limits.h usability... yes checking limits.h presence... yes checking for limits.h... yes checking malloc.h usability... yes checking malloc.h presence... yes checking for malloc.h... yes checking for an ANSI C-conforming const... yes checking for size_t... yes checking for int8_t... yes checking for int16_t... yes checking for int32_t... yes checking for int64_t... yes checking for off_t... yes checking for stdlib.h... (cached) yes checking for GNU libc compatible malloc... yes checking for stdlib.h... (cached) yes checking for unistd.h... (cached) yes checking for getpagesize... yes checking for working mmap... yes checking for stdlib.h... (cached) yes checking for GNU libc compatible realloc... yes checking for working strtod... yes checking for strstr... yes checking for strtod... (cached) yes checking for strtol... yes checking for strerror... yes checking for memset... yes checking for floor... no checking for library containing floor... -lm checking for pow... yes checking for sqrt... yes checking whether to enable debug mode... yes disabled cernlib dependent code... yes configure: creating ./config.status config.status: creating Makefile config.status: creating lib/Makefile config.status: creating src/Makefile config.status: creating ice/Makefile config.status: creating scripts/Makefile config.status: creating amasim/Makefile config.status: creating level2/Makefile config.status: creating config.h config.status: executing depfiles commands config.status: executing libtool commands ------------------------------------------------------------------- Photonics: "1.73: pyrosoma r4" please refer to the 'INSTALL' file for further instructions. Hints: Building photonics......................'make' Compiling tool directory................'make tool' Performing post compile test............'make tests' All of the above........................'make everything' Clean objects and binaries..............'make clean' Clean tool directory....................'make toolclean' Remove traces of previous configure.....'make distclean' Also consider trying 'scripts/install_in_icetray.sh --help' ---> Building photonics_1.73 with target all DEBUG: Executing com.apple.build (photonics_1.73) DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all' make all-recursive make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73' Making all in lib make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib' /bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o boundary.c:1: error: bad value (native) for -march= switch boundary.c:1: error: bad value (native) for -mtune= switch make[2]: *** [boundary.lo] Error 1 make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib' make[1]: *** [all-recursive] Error 1 make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73' make: *** [all] Error 2 Error: Target com.apple.build returned: shell command "cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all" returned error 2 Command output: make all-recursive make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73' Making all in lib make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib' /bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o boundary.c:1: error: bad value (native) for -march= switch boundary.c:1: error: bad value (native) for -mtune= switch make[2]: *** [boundary.lo] Error 1 make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib' make[1]: *** [all-recursive] Error 1 make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73' make: *** [all] Error 2 Warning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install icecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS ``` Migrated from https://code.icecube.wisc.edu/ticket/686 ```json { "status": "closed", "changetime": "2012-06-22T16:05:07", "description": "Scientific Linux release 5.8 (Boron)\n\n{{{\n$ gcc --version\ngcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)\nCopyright (C) 2006 Free Software Foundation, Inc.\nThis is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n}}}\n\n{{{\n$ ./bin/port -vd install photonics_1.73\nDEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Requested variant x86_64 is not provided by port photonics_1.73.\nDEBUG: Executing variant linux provides linux\nDEBUG: Executing com.apple.main (photonics_1.73)\nDEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz\nDEBUG: Skipping unarchive (photonics_1.73) since no archive found\nDEBUG: Skipping completed com.apple.unarchive (photonics_1.73)\n---> Fetching photonics_1.73\nDEBUG: Executing com.apple.fetch (photonics_1.73)\n---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\n---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M\n---> Verifying checksum(s) for photonics_1.73\nDEBUG: Executing com.apple.checksum (photonics_1.73)\n---> Checksumming photonics-1.73.tar.bz2\nDEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2\nDEBUG: setting option extract.cmd to /usr/bin/bzip2\n---> Extracting photonics_1.73\nDEBUG: Executing com.apple.extract (photonics_1.73)\n---> Extracting photonics-1.73.tar.bz2\nDEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work\" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'\nDEBUG: Executing com.apple.patch (photonics_1.73)\n---> Configuring photonics_1.73\nDEBUG: Executing com.apple.configure (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether build environment is sane... yes\nchecking for a thread-safe mkdir -p... /bin/mkdir -p\nchecking for gawk... gawk\nchecking whether make sets $(MAKE)... yes\nchecking for gawk... (cached) gawk\nchecking for gcc... gcc\nchecking for C compiler default output file name... a.out\nchecking whether the C compiler works... yes\nchecking whether we are cross compiling... no\nchecking for suffix of executables... \nchecking for suffix of object files... o\nchecking whether we are using the GNU C compiler... yes\nchecking whether gcc accepts -g... yes\nchecking for gcc option to accept ISO C89... none needed\nchecking for style of include used by make... GNU\nchecking dependency style of gcc... gcc3\nchecking how to run the C preprocessor... gcc -E\nchecking for g++... g++\nchecking whether we are using the GNU C++ compiler... yes\nchecking whether g++ accepts -g... yes\nchecking dependency style of g++... gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether ln -s works... yes\nchecking whether make sets $(MAKE)... (cached) yes\nchecking build system type... x86_64-unknown-linux-gnu\nchecking host system type... x86_64-unknown-linux-gnu\nchecking for a sed that does not truncate output... /bin/sed\nchecking for grep that handles long lines and -e... /bin/grep\nchecking for egrep... /bin/grep -E\nchecking for fgrep... /bin/grep -F\nchecking for ld used by gcc... /usr/bin/ld\nchecking if the linker (/usr/bin/ld) is GNU ld... yes\nchecking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B\nchecking the name lister (/usr/bin/nm -B) interface... BSD nm\nchecking the maximum length of command line arguments... 98304\nchecking whether the shell understands some XSI constructs... yes\nchecking whether the shell understands \"+=\"... yes\nchecking for /usr/bin/ld option to reload object files... -r\nchecking how to recognize dependent libraries... pass_all\nchecking for ar... ar\nchecking for strip... strip\nchecking for ranlib... ranlib\nchecking command to parse /usr/bin/nm -B output from gcc object... ok\nchecking for ANSI C header files... yes\nchecking for sys/types.h... yes\nchecking for sys/stat.h... yes\nchecking for stdlib.h... yes\nchecking for string.h... yes\nchecking for memory.h... yes\nchecking for strings.h... yes\nchecking for inttypes.h... yes\nchecking for stdint.h... yes\nchecking for unistd.h... yes\nchecking for dlfcn.h... yes\nchecking whether we are using the GNU C++ compiler... (cached) yes\nchecking whether g++ accepts -g... (cached) yes\nchecking dependency style of g++... (cached) gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for objdir... .libs\nchecking if gcc supports -fno-rtti -fno-exceptions... no\nchecking for gcc option to produce PIC... -fPIC -DPIC\nchecking if gcc PIC flag -fPIC -DPIC works... yes\nchecking if gcc static flag -static works... yes\nchecking if gcc supports -c -o file.o... yes\nchecking if gcc supports -c -o file.o... (cached) yes\nchecking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking whether -lc should be explicitly linked in... no\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking whether stripping libraries is possible... yes\nchecking if libtool supports shared libraries... yes\nchecking whether to build shared libraries... yes\nchecking whether to build static libraries... yes\nchecking for ld used by g++... /usr/bin/ld -m elf_x86_64\nchecking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking for g++ option to produce PIC... -fPIC -DPIC\nchecking if g++ PIC flag -fPIC -DPIC works... yes\nchecking if g++ static flag -static works... yes\nchecking if g++ supports -c -o file.o... yes\nchecking if g++ supports -c -o file.o... (cached) yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking for ANSI C header files... (cached) yes\nchecking for stdbool.h that conforms to C99... yes\nchecking for _Bool... yes\nchecking limits.h usability... yes\nchecking limits.h presence... yes\nchecking for limits.h... yes\nchecking malloc.h usability... yes\nchecking malloc.h presence... yes\nchecking for malloc.h... yes\nchecking for an ANSI C-conforming const... yes\nchecking for size_t... yes\nchecking for int8_t... yes\nchecking for int16_t... yes\nchecking for int32_t... yes\nchecking for int64_t... yes\nchecking for off_t... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible malloc... yes\nchecking for stdlib.h... (cached) yes\nchecking for unistd.h... (cached) yes\nchecking for getpagesize... yes\nchecking for working mmap... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible realloc... yes\nchecking for working strtod... yes\nchecking for strstr... yes\nchecking for strtod... (cached) yes\nchecking for strtol... yes\nchecking for strerror... yes\nchecking for memset... yes\nchecking for floor... no\nchecking for library containing floor... -lm\nchecking for pow... yes\nchecking for sqrt... yes\nchecking whether to enable debug mode... yes\ndisabled cernlib dependent code... yes\nconfigure: creating ./config.status\nconfig.status: creating Makefile\nconfig.status: creating lib/Makefile\nconfig.status: creating src/Makefile\nconfig.status: creating ice/Makefile\nconfig.status: creating scripts/Makefile\nconfig.status: creating amasim/Makefile\nconfig.status: creating level2/Makefile\nconfig.status: creating config.h\nconfig.status: executing depfiles commands\nconfig.status: executing libtool commands\n-------------------------------------------------------------------\n Photonics: \"1.73: pyrosoma r4\"\n please refer to the 'INSTALL' file for further instructions.\n Hints:\n\n Building photonics......................'make'\n Compiling tool directory................'make tool'\n Performing post compile test............'make tests'\n All of the above........................'make everything'\n\n Clean objects and binaries..............'make clean'\n Clean tool directory....................'make toolclean'\n Remove traces of previous configure.....'make distclean'\n\n Also consider trying 'scripts/install_in_icetray.sh --help'\n\n\n---> Building photonics_1.73 with target all\nDEBUG: Executing com.apple.build (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all'\nmake all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\nError: Target com.apple.build returned: shell command \"cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all\" returned error 2\nCommand output: make all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\n\nWarning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install\nicecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS\n}}}", "reporter": "boersma", "cc": "", "resolution": "worksforme", "_ts": "1340381107000000", "component": "booking", "summary": "port photonics_1.73 does not build on SL5 64 bit", "priority": "normal", "keywords": "photonics I3_PORTS gcc", "time": "2012-06-22T15:13:34", "milestone": "", "owner": "", "type": "defect" } ```
1.0
port photonics_1.73 does not build on SL5 64 bit (Trac #686) - Scientific Linux release 5.8 (Boron) ```text $ gcc --version gcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52) Copyright (C) 2006 Free Software Foundation, Inc. This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. ``` ```text $ ./bin/port -vd install photonics_1.73 DEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73 DEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73 DEBUG: Requested variant x86_64 is not provided by port photonics_1.73. DEBUG: Executing variant linux provides linux DEBUG: Executing com.apple.main (photonics_1.73) DEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz DEBUG: Skipping unarchive (photonics_1.73) since no archive found DEBUG: Skipping completed com.apple.unarchive (photonics_1.73) ---> Fetching photonics_1.73 DEBUG: Executing com.apple.fetch (photonics_1.73) ---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73 ---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2' % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M ---> Verifying checksum(s) for photonics_1.73 DEBUG: Executing com.apple.checksum (photonics_1.73) ---> Checksumming photonics-1.73.tar.bz2 DEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2 DEBUG: setting option extract.cmd to /usr/bin/bzip2 ---> Extracting photonics_1.73 DEBUG: Executing com.apple.extract (photonics_1.73) ---> Extracting photonics-1.73.tar.bz2 DEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -' DEBUG: Executing com.apple.patch (photonics_1.73) ---> Configuring photonics_1.73 DEBUG: Executing com.apple.configure (photonics_1.73) DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize' checking for a BSD-compatible install... /usr/bin/install -c checking whether build environment is sane... yes checking for a thread-safe mkdir -p... /bin/mkdir -p checking for gawk... gawk checking whether make sets $(MAKE)... yes checking for gawk... (cached) gawk checking for gcc... gcc checking for C compiler default output file name... a.out checking whether the C compiler works... yes checking whether we are cross compiling... no checking for suffix of executables... checking for suffix of object files... o checking whether we are using the GNU C compiler... yes checking whether gcc accepts -g... yes checking for gcc option to accept ISO C89... none needed checking for style of include used by make... GNU checking dependency style of gcc... gcc3 checking how to run the C preprocessor... gcc -E checking for g++... g++ checking whether we are using the GNU C++ compiler... yes checking whether g++ accepts -g... yes checking dependency style of g++... gcc3 checking how to run the C++ preprocessor... g++ -E checking for a BSD-compatible install... /usr/bin/install -c checking whether ln -s works... yes checking whether make sets $(MAKE)... (cached) yes checking build system type... x86_64-unknown-linux-gnu checking host system type... x86_64-unknown-linux-gnu checking for a sed that does not truncate output... /bin/sed checking for grep that handles long lines and -e... /bin/grep checking for egrep... /bin/grep -E checking for fgrep... /bin/grep -F checking for ld used by gcc... /usr/bin/ld checking if the linker (/usr/bin/ld) is GNU ld... yes checking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B checking the name lister (/usr/bin/nm -B) interface... BSD nm checking the maximum length of command line arguments... 98304 checking whether the shell understands some XSI constructs... yes checking whether the shell understands "+="... yes checking for /usr/bin/ld option to reload object files... -r checking how to recognize dependent libraries... pass_all checking for ar... ar checking for strip... strip checking for ranlib... ranlib checking command to parse /usr/bin/nm -B output from gcc object... ok checking for ANSI C header files... yes checking for sys/types.h... yes checking for sys/stat.h... yes checking for stdlib.h... yes checking for string.h... yes checking for memory.h... yes checking for strings.h... yes checking for inttypes.h... yes checking for stdint.h... yes checking for unistd.h... yes checking for dlfcn.h... yes checking whether we are using the GNU C++ compiler... (cached) yes checking whether g++ accepts -g... (cached) yes checking dependency style of g++... (cached) gcc3 checking how to run the C++ preprocessor... g++ -E checking for objdir... .libs checking if gcc supports -fno-rtti -fno-exceptions... no checking for gcc option to produce PIC... -fPIC -DPIC checking if gcc PIC flag -fPIC -DPIC works... yes checking if gcc static flag -static works... yes checking if gcc supports -c -o file.o... yes checking if gcc supports -c -o file.o... (cached) yes checking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes checking whether -lc should be explicitly linked in... no checking dynamic linker characteristics... GNU/Linux ld.so checking how to hardcode library paths into programs... immediate checking whether stripping libraries is possible... yes checking if libtool supports shared libraries... yes checking whether to build shared libraries... yes checking whether to build static libraries... yes checking for ld used by g++... /usr/bin/ld -m elf_x86_64 checking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes checking for g++ option to produce PIC... -fPIC -DPIC checking if g++ PIC flag -fPIC -DPIC works... yes checking if g++ static flag -static works... yes checking if g++ supports -c -o file.o... yes checking if g++ supports -c -o file.o... (cached) yes checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes checking dynamic linker characteristics... GNU/Linux ld.so checking how to hardcode library paths into programs... immediate checking for ANSI C header files... (cached) yes checking for stdbool.h that conforms to C99... yes checking for _Bool... yes checking limits.h usability... yes checking limits.h presence... yes checking for limits.h... yes checking malloc.h usability... yes checking malloc.h presence... yes checking for malloc.h... yes checking for an ANSI C-conforming const... yes checking for size_t... yes checking for int8_t... yes checking for int16_t... yes checking for int32_t... yes checking for int64_t... yes checking for off_t... yes checking for stdlib.h... (cached) yes checking for GNU libc compatible malloc... yes checking for stdlib.h... (cached) yes checking for unistd.h... (cached) yes checking for getpagesize... yes checking for working mmap... yes checking for stdlib.h... (cached) yes checking for GNU libc compatible realloc... yes checking for working strtod... yes checking for strstr... yes checking for strtod... (cached) yes checking for strtol... yes checking for strerror... yes checking for memset... yes checking for floor... no checking for library containing floor... -lm checking for pow... yes checking for sqrt... yes checking whether to enable debug mode... yes disabled cernlib dependent code... yes configure: creating ./config.status config.status: creating Makefile config.status: creating lib/Makefile config.status: creating src/Makefile config.status: creating ice/Makefile config.status: creating scripts/Makefile config.status: creating amasim/Makefile config.status: creating level2/Makefile config.status: creating config.h config.status: executing depfiles commands config.status: executing libtool commands ------------------------------------------------------------------- Photonics: "1.73: pyrosoma r4" please refer to the 'INSTALL' file for further instructions. Hints: Building photonics......................'make' Compiling tool directory................'make tool' Performing post compile test............'make tests' All of the above........................'make everything' Clean objects and binaries..............'make clean' Clean tool directory....................'make toolclean' Remove traces of previous configure.....'make distclean' Also consider trying 'scripts/install_in_icetray.sh --help' ---> Building photonics_1.73 with target all DEBUG: Executing com.apple.build (photonics_1.73) DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all' make all-recursive make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73' Making all in lib make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib' /bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o boundary.c:1: error: bad value (native) for -march= switch boundary.c:1: error: bad value (native) for -mtune= switch make[2]: *** [boundary.lo] Error 1 make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib' make[1]: *** [all-recursive] Error 1 make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73' make: *** [all] Error 2 Error: Target com.apple.build returned: shell command "cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all" returned error 2 Command output: make all-recursive make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73' Making all in lib make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib' /bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o boundary.c:1: error: bad value (native) for -march= switch boundary.c:1: error: bad value (native) for -mtune= switch make[2]: *** [boundary.lo] Error 1 make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib' make[1]: *** [all-recursive] Error 1 make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73' make: *** [all] Error 2 Warning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install icecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS ``` Migrated from https://code.icecube.wisc.edu/ticket/686 ```json { "status": "closed", "changetime": "2012-06-22T16:05:07", "description": "Scientific Linux release 5.8 (Boron)\n\n{{{\n$ gcc --version\ngcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)\nCopyright (C) 2006 Free Software Foundation, Inc.\nThis is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n}}}\n\n{{{\n$ ./bin/port -vd install photonics_1.73\nDEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Requested variant x86_64 is not provided by port photonics_1.73.\nDEBUG: Executing variant linux provides linux\nDEBUG: Executing com.apple.main (photonics_1.73)\nDEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz\nDEBUG: Skipping unarchive (photonics_1.73) since no archive found\nDEBUG: Skipping completed com.apple.unarchive (photonics_1.73)\n---> Fetching photonics_1.73\nDEBUG: Executing com.apple.fetch (photonics_1.73)\n---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\n---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M\n---> Verifying checksum(s) for photonics_1.73\nDEBUG: Executing com.apple.checksum (photonics_1.73)\n---> Checksumming photonics-1.73.tar.bz2\nDEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2\nDEBUG: setting option extract.cmd to /usr/bin/bzip2\n---> Extracting photonics_1.73\nDEBUG: Executing com.apple.extract (photonics_1.73)\n---> Extracting photonics-1.73.tar.bz2\nDEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work\" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'\nDEBUG: Executing com.apple.patch (photonics_1.73)\n---> Configuring photonics_1.73\nDEBUG: Executing com.apple.configure (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether build environment is sane... yes\nchecking for a thread-safe mkdir -p... /bin/mkdir -p\nchecking for gawk... gawk\nchecking whether make sets $(MAKE)... yes\nchecking for gawk... (cached) gawk\nchecking for gcc... gcc\nchecking for C compiler default output file name... a.out\nchecking whether the C compiler works... yes\nchecking whether we are cross compiling... no\nchecking for suffix of executables... \nchecking for suffix of object files... o\nchecking whether we are using the GNU C compiler... yes\nchecking whether gcc accepts -g... yes\nchecking for gcc option to accept ISO C89... none needed\nchecking for style of include used by make... GNU\nchecking dependency style of gcc... gcc3\nchecking how to run the C preprocessor... gcc -E\nchecking for g++... g++\nchecking whether we are using the GNU C++ compiler... yes\nchecking whether g++ accepts -g... yes\nchecking dependency style of g++... gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether ln -s works... yes\nchecking whether make sets $(MAKE)... (cached) yes\nchecking build system type... x86_64-unknown-linux-gnu\nchecking host system type... x86_64-unknown-linux-gnu\nchecking for a sed that does not truncate output... /bin/sed\nchecking for grep that handles long lines and -e... /bin/grep\nchecking for egrep... /bin/grep -E\nchecking for fgrep... /bin/grep -F\nchecking for ld used by gcc... /usr/bin/ld\nchecking if the linker (/usr/bin/ld) is GNU ld... yes\nchecking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B\nchecking the name lister (/usr/bin/nm -B) interface... BSD nm\nchecking the maximum length of command line arguments... 98304\nchecking whether the shell understands some XSI constructs... yes\nchecking whether the shell understands \"+=\"... yes\nchecking for /usr/bin/ld option to reload object files... -r\nchecking how to recognize dependent libraries... pass_all\nchecking for ar... ar\nchecking for strip... strip\nchecking for ranlib... ranlib\nchecking command to parse /usr/bin/nm -B output from gcc object... ok\nchecking for ANSI C header files... yes\nchecking for sys/types.h... yes\nchecking for sys/stat.h... yes\nchecking for stdlib.h... yes\nchecking for string.h... yes\nchecking for memory.h... yes\nchecking for strings.h... yes\nchecking for inttypes.h... yes\nchecking for stdint.h... yes\nchecking for unistd.h... yes\nchecking for dlfcn.h... yes\nchecking whether we are using the GNU C++ compiler... (cached) yes\nchecking whether g++ accepts -g... (cached) yes\nchecking dependency style of g++... (cached) gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for objdir... .libs\nchecking if gcc supports -fno-rtti -fno-exceptions... no\nchecking for gcc option to produce PIC... -fPIC -DPIC\nchecking if gcc PIC flag -fPIC -DPIC works... yes\nchecking if gcc static flag -static works... yes\nchecking if gcc supports -c -o file.o... yes\nchecking if gcc supports -c -o file.o... (cached) yes\nchecking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking whether -lc should be explicitly linked in... no\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking whether stripping libraries is possible... yes\nchecking if libtool supports shared libraries... yes\nchecking whether to build shared libraries... yes\nchecking whether to build static libraries... yes\nchecking for ld used by g++... /usr/bin/ld -m elf_x86_64\nchecking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking for g++ option to produce PIC... -fPIC -DPIC\nchecking if g++ PIC flag -fPIC -DPIC works... yes\nchecking if g++ static flag -static works... yes\nchecking if g++ supports -c -o file.o... yes\nchecking if g++ supports -c -o file.o... (cached) yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking for ANSI C header files... (cached) yes\nchecking for stdbool.h that conforms to C99... yes\nchecking for _Bool... yes\nchecking limits.h usability... yes\nchecking limits.h presence... yes\nchecking for limits.h... yes\nchecking malloc.h usability... yes\nchecking malloc.h presence... yes\nchecking for malloc.h... yes\nchecking for an ANSI C-conforming const... yes\nchecking for size_t... yes\nchecking for int8_t... yes\nchecking for int16_t... yes\nchecking for int32_t... yes\nchecking for int64_t... yes\nchecking for off_t... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible malloc... yes\nchecking for stdlib.h... (cached) yes\nchecking for unistd.h... (cached) yes\nchecking for getpagesize... yes\nchecking for working mmap... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible realloc... yes\nchecking for working strtod... yes\nchecking for strstr... yes\nchecking for strtod... (cached) yes\nchecking for strtol... yes\nchecking for strerror... yes\nchecking for memset... yes\nchecking for floor... no\nchecking for library containing floor... -lm\nchecking for pow... yes\nchecking for sqrt... yes\nchecking whether to enable debug mode... yes\ndisabled cernlib dependent code... yes\nconfigure: creating ./config.status\nconfig.status: creating Makefile\nconfig.status: creating lib/Makefile\nconfig.status: creating src/Makefile\nconfig.status: creating ice/Makefile\nconfig.status: creating scripts/Makefile\nconfig.status: creating amasim/Makefile\nconfig.status: creating level2/Makefile\nconfig.status: creating config.h\nconfig.status: executing depfiles commands\nconfig.status: executing libtool commands\n-------------------------------------------------------------------\n Photonics: \"1.73: pyrosoma r4\"\n please refer to the 'INSTALL' file for further instructions.\n Hints:\n\n Building photonics......................'make'\n Compiling tool directory................'make tool'\n Performing post compile test............'make tests'\n All of the above........................'make everything'\n\n Clean objects and binaries..............'make clean'\n Clean tool directory....................'make toolclean'\n Remove traces of previous configure.....'make distclean'\n\n Also consider trying 'scripts/install_in_icetray.sh --help'\n\n\n---> Building photonics_1.73 with target all\nDEBUG: Executing com.apple.build (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all'\nmake all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\nError: Target com.apple.build returned: shell command \"cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all\" returned error 2\nCommand output: make all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\n\nWarning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install\nicecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS\n}}}", "reporter": "boersma", "cc": "", "resolution": "worksforme", "_ts": "1340381107000000", "component": "booking", "summary": "port photonics_1.73 does not build on SL5 64 bit", "priority": "normal", "keywords": "photonics I3_PORTS gcc", "time": "2012-06-22T15:13:34", "milestone": "", "owner": "", "type": "defect" } ```
defect
port photonics does not build on bit trac scientific linux release boron text gcc version gcc gcc red hat copyright c free software foundation inc this is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose text bin port vd install photonics debug found port in file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics debug changing to port directory net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics debug requested variant is not provided by port photonics debug executing variant linux provides linux debug executing com apple main photonics debug no tgz archive net software icecube ports var db dports packages linux photonics tgz debug skipping unarchive photonics since no archive found debug skipping completed com apple unarchive photonics fetching photonics debug executing com apple fetch photonics photonics tar doesn t seem to exist in net software icecube ports var db dports distfiles photonics attempting to fetch photonics tar from debug assembled command cd net software icecube ports var db dports distfiles photonics curl f l o photonics tar tmp total received xferd average speed time time time current dload upload total spent left speed verifying checksum s for photonics debug executing com apple checksum photonics checksumming photonics tar debug correct checksum for photonics tar debug setting option extract cmd to usr bin extracting photonics debug executing com apple extract photonics extracting photonics tar debug setting option extract args to net software icecube ports var db dports distfiles photonics photonics tar debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work usr bin dc net software icecube ports var db dports distfiles photonics photonics tar tar no same owner xf debug executing com apple patch photonics configuring photonics debug executing com apple configure photonics debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic configure prefix net software icecube ports enable optimize checking for a bsd compatible install usr bin install c checking whether build environment is sane yes checking for a thread safe mkdir p bin mkdir p checking for gawk gawk checking whether make sets make yes checking for gawk cached gawk checking for gcc gcc checking for c compiler default output file name a out checking whether the c compiler works yes checking whether we are cross compiling no checking for suffix of executables checking for suffix of object files o checking whether we are using the gnu c compiler yes checking whether gcc accepts g yes checking for gcc option to accept iso none needed checking for style of include used by make gnu checking dependency style of gcc checking how to run the c preprocessor gcc e checking for g g checking whether we are using the gnu c compiler yes checking whether g accepts g yes checking dependency style of g checking how to run the c preprocessor g e checking for a bsd compatible install usr bin install c checking whether ln s works yes checking whether make sets make cached yes checking build system type unknown linux gnu checking host system type unknown linux gnu checking for a sed that does not truncate output bin sed checking for grep that handles long lines and e bin grep checking for egrep bin grep e checking for fgrep bin grep f checking for ld used by gcc usr bin ld checking if the linker usr bin ld is gnu ld yes checking for bsd or ms compatible name lister nm usr bin nm b checking the name lister usr bin nm b interface bsd nm checking the maximum length of command line arguments checking whether the shell understands some xsi constructs yes checking whether the shell understands yes checking for usr bin ld option to reload object files r checking how to recognize dependent libraries pass all checking for ar ar checking for strip strip checking for ranlib ranlib checking command to parse usr bin nm b output from gcc object ok checking for ansi c header files yes checking for sys types h yes checking for sys stat h yes checking for stdlib h yes checking for string h yes checking for memory h yes checking for strings h yes checking for inttypes h yes checking for stdint h yes checking for unistd h yes checking for dlfcn h yes checking whether we are using the gnu c compiler cached yes checking whether g accepts g cached yes checking dependency style of g cached checking how to run the c preprocessor g e checking for objdir libs checking if gcc supports fno rtti fno exceptions no checking for gcc option to produce pic fpic dpic checking if gcc pic flag fpic dpic works yes checking if gcc static flag static works yes checking if gcc supports c o file o yes checking if gcc supports c o file o cached yes checking whether the gcc linker usr bin ld m elf supports shared libraries yes checking whether lc should be explicitly linked in no checking dynamic linker characteristics gnu linux ld so checking how to hardcode library paths into programs immediate checking whether stripping libraries is possible yes checking if libtool supports shared libraries yes checking whether to build shared libraries yes checking whether to build static libraries yes checking for ld used by g usr bin ld m elf checking if the linker usr bin ld m elf is gnu ld yes checking whether the g linker usr bin ld m elf supports shared libraries yes checking for g option to produce pic fpic dpic checking if g pic flag fpic dpic works yes checking if g static flag static works yes checking if g supports c o file o yes checking if g supports c o file o cached yes checking whether the g linker usr bin ld m elf supports shared libraries yes checking dynamic linker characteristics gnu linux ld so checking how to hardcode library paths into programs immediate checking for ansi c header files cached yes checking for stdbool h that conforms to yes checking for bool yes checking limits h usability yes checking limits h presence yes checking for limits h yes checking malloc h usability yes checking malloc h presence yes checking for malloc h yes checking for an ansi c conforming const yes checking for size t yes checking for t yes checking for t yes checking for t yes checking for t yes checking for off t yes checking for stdlib h cached yes checking for gnu libc compatible malloc yes checking for stdlib h cached yes checking for unistd h cached yes checking for getpagesize yes checking for working mmap yes checking for stdlib h cached yes checking for gnu libc compatible realloc yes checking for working strtod yes checking for strstr yes checking for strtod cached yes checking for strtol yes checking for strerror yes checking for memset yes checking for floor no checking for library containing floor lm checking for pow yes checking for sqrt yes checking whether to enable debug mode yes disabled cernlib dependent code yes configure creating config status config status creating makefile config status creating lib makefile config status creating src makefile config status creating ice makefile config status creating scripts makefile config status creating amasim makefile config status creating makefile config status creating config h config status executing depfiles commands config status executing libtool commands photonics pyrosoma please refer to the install file for further instructions hints building photonics make compiling tool directory make tool performing post compile test make tests all of the above make everything clean objects and binaries make clean clean tool directory make toolclean remove traces of previous configure make distclean also consider trying scripts install in icetray sh help building photonics with target all debug executing com apple build photonics debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all make all recursive make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics making all in lib make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c libtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o boundary c error bad value native for march switch boundary c error bad value native for mtune switch make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics make error error target com apple build returned shell command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all returned error command output make all recursive make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics making all in lib make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c libtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o boundary c error bad value native for march switch boundary c error bad value native for mtune switch make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics make error warning the following items did not execute for photonics com apple activate com apple build com apple destroot com apple archive com apple install icecubemgr net software icecube ports migrated from json status closed changetime description scientific linux release boron n n n gcc version ngcc gcc red hat ncopyright c free software foundation inc nthis is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose n n n n bin port vd install photonics ndebug found port in file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics ndebug changing to port directory net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics ndebug requested variant is not provided by port photonics ndebug executing variant linux provides linux ndebug executing com apple main photonics ndebug no tgz archive net software icecube ports var db dports packages linux photonics tgz ndebug skipping unarchive photonics since no archive found ndebug skipping completed com apple unarchive photonics n fetching photonics ndebug executing com apple fetch photonics n photonics tar doesn t seem to exist in net software icecube ports var db dports distfiles photonics n attempting to fetch photonics tar from assembled command cd net software icecube ports var db dports distfiles photonics curl f l o photonics tar tmp total received xferd average speed time time time current n dload upload total spent left speed n verifying checksum s for photonics ndebug executing com apple checksum photonics n checksumming photonics tar ndebug correct checksum for photonics tar ndebug setting option extract cmd to usr bin n extracting photonics ndebug executing com apple extract photonics n extracting photonics tar ndebug setting option extract args to net software icecube ports var db dports distfiles photonics photonics tar ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work usr bin dc net software icecube ports var db dports distfiles photonics photonics tar tar no same owner xf ndebug executing com apple patch photonics n configuring photonics ndebug executing com apple configure photonics ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic configure prefix net software icecube ports enable optimize nchecking for a bsd compatible install usr bin install c nchecking whether build environment is sane yes nchecking for a thread safe mkdir p bin mkdir p nchecking for gawk gawk nchecking whether make sets make yes nchecking for gawk cached gawk nchecking for gcc gcc nchecking for c compiler default output file name a out nchecking whether the c compiler works yes nchecking whether we are cross compiling no nchecking for suffix of executables nchecking for suffix of object files o nchecking whether we are using the gnu c compiler yes nchecking whether gcc accepts g yes nchecking for gcc option to accept iso none needed nchecking for style of include used by make gnu nchecking dependency style of gcc nchecking how to run the c preprocessor gcc e nchecking for g g nchecking whether we are using the gnu c compiler yes nchecking whether g accepts g yes nchecking dependency style of g nchecking how to run the c preprocessor g e nchecking for a bsd compatible install usr bin install c nchecking whether ln s works yes nchecking whether make sets make cached yes nchecking build system type unknown linux gnu nchecking host system type unknown linux gnu nchecking for a sed that does not truncate output bin sed nchecking for grep that handles long lines and e bin grep nchecking for egrep bin grep e nchecking for fgrep bin grep f nchecking for ld used by gcc usr bin ld nchecking if the linker usr bin ld is gnu ld yes nchecking for bsd or ms compatible name lister nm usr bin nm b nchecking the name lister usr bin nm b interface bsd nm nchecking the maximum length of command line arguments nchecking whether the shell understands some xsi constructs yes nchecking whether the shell understands yes nchecking for usr bin ld option to reload object files r nchecking how to recognize dependent libraries pass all nchecking for ar ar nchecking for strip strip nchecking for ranlib ranlib nchecking command to parse usr bin nm b output from gcc object ok nchecking for ansi c header files yes nchecking for sys types h yes nchecking for sys stat h yes nchecking for stdlib h yes nchecking for string h yes nchecking for memory h yes nchecking for strings h yes nchecking for inttypes h yes nchecking for stdint h yes nchecking for unistd h yes nchecking for dlfcn h yes nchecking whether we are using the gnu c compiler cached yes nchecking whether g accepts g cached yes nchecking dependency style of g cached nchecking how to run the c preprocessor g e nchecking for objdir libs nchecking if gcc supports fno rtti fno exceptions no nchecking for gcc option to produce pic fpic dpic nchecking if gcc pic flag fpic dpic works yes nchecking if gcc static flag static works yes nchecking if gcc supports c o file o yes nchecking if gcc supports c o file o cached yes nchecking whether the gcc linker usr bin ld m elf supports shared libraries yes nchecking whether lc should be explicitly linked in no nchecking dynamic linker characteristics gnu linux ld so nchecking how to hardcode library paths into programs immediate nchecking whether stripping libraries is possible yes nchecking if libtool supports shared libraries yes nchecking whether to build shared libraries yes nchecking whether to build static libraries yes nchecking for ld used by g usr bin ld m elf nchecking if the linker usr bin ld m elf is gnu ld yes nchecking whether the g linker usr bin ld m elf supports shared libraries yes nchecking for g option to produce pic fpic dpic nchecking if g pic flag fpic dpic works yes nchecking if g static flag static works yes nchecking if g supports c o file o yes nchecking if g supports c o file o cached yes nchecking whether the g linker usr bin ld m elf supports shared libraries yes nchecking dynamic linker characteristics gnu linux ld so nchecking how to hardcode library paths into programs immediate nchecking for ansi c header files cached yes nchecking for stdbool h that conforms to yes nchecking for bool yes nchecking limits h usability yes nchecking limits h presence yes nchecking for limits h yes nchecking malloc h usability yes nchecking malloc h presence yes nchecking for malloc h yes nchecking for an ansi c conforming const yes nchecking for size t yes nchecking for t yes nchecking for t yes nchecking for t yes nchecking for t yes nchecking for off t yes nchecking for stdlib h cached yes nchecking for gnu libc compatible malloc yes nchecking for stdlib h cached yes nchecking for unistd h cached yes nchecking for getpagesize yes nchecking for working mmap yes nchecking for stdlib h cached yes nchecking for gnu libc compatible realloc yes nchecking for working strtod yes nchecking for strstr yes nchecking for strtod cached yes nchecking for strtol yes nchecking for strerror yes nchecking for memset yes nchecking for floor no nchecking for library containing floor lm nchecking for pow yes nchecking for sqrt yes nchecking whether to enable debug mode yes ndisabled cernlib dependent code yes nconfigure creating config status nconfig status creating makefile nconfig status creating lib makefile nconfig status creating src makefile nconfig status creating ice makefile nconfig status creating scripts makefile nconfig status creating amasim makefile nconfig status creating makefile nconfig status creating config h nconfig status executing depfiles commands nconfig status executing libtool commands n n photonics pyrosoma n please refer to the install file for further instructions n hints n n building photonics make n compiling tool directory make tool n performing post compile test make tests n all of the above make everything n n clean objects and binaries make clean n clean tool directory make toolclean n remove traces of previous configure make distclean n n also consider trying scripts install in icetray sh help n n n building photonics with target all ndebug executing com apple build photonics ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all nmake all recursive nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmaking all in lib nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib n bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c nlibtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o nboundary c error bad value native for march switch nboundary c error bad value native for mtune switch nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmake error nerror target com apple build returned shell command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all returned error ncommand output make all recursive nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmaking all in lib nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib n bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c nlibtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o nboundary c error bad value native for march switch nboundary c error bad value native for mtune switch nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmake error n nwarning the following items did not execute for photonics com apple activate com apple build com apple destroot com apple archive com apple install nicecubemgr net software icecube ports n reporter boersma cc resolution worksforme ts component booking summary port photonics does not build on bit priority normal keywords photonics ports gcc time milestone owner type defect
1
48,405
20,144,015,331
IssuesEvent
2022-02-09 04:23:10
Azure/azure-powershell
https://api.github.com/repos/Azure/azure-powershell
closed
Enable-AzFrontDoorCustomDomainHttps doesnt return error / removes https settings from endpoint
Service Attention Network - Front Door customer-reported
We have a case where one of our certs isnt compatible with azure frontdoor and the error we get through the azure portal is: 'Failed to update the custom https configuration for the frontend host 'ENDPOINTNAME'. Error: The private key is not RSA or it is unreadable. Only RSA private key is supported for BYOC to secure a custom domain.' (we have a support case open about this) This error is not returned when running the Enable-AzFrontDoorCustomDomainHttps cmdlet but FAR WORSE after the cmdlet is run it removes the currently configured cert from the endpoint so rather than us getting an updated cert as expected we are left with a broken config and a cert error is presented to our users.
1.0
Enable-AzFrontDoorCustomDomainHttps doesnt return error / removes https settings from endpoint - We have a case where one of our certs isnt compatible with azure frontdoor and the error we get through the azure portal is: 'Failed to update the custom https configuration for the frontend host 'ENDPOINTNAME'. Error: The private key is not RSA or it is unreadable. Only RSA private key is supported for BYOC to secure a custom domain.' (we have a support case open about this) This error is not returned when running the Enable-AzFrontDoorCustomDomainHttps cmdlet but FAR WORSE after the cmdlet is run it removes the currently configured cert from the endpoint so rather than us getting an updated cert as expected we are left with a broken config and a cert error is presented to our users.
non_defect
enable azfrontdoorcustomdomainhttps doesnt return error removes https settings from endpoint we have a case where one of our certs isnt compatible with azure frontdoor and the error we get through the azure portal is failed to update the custom https configuration for the frontend host endpointname error the private key is not rsa or it is unreadable only rsa private key is supported for byoc to secure a custom domain we have a support case open about this this error is not returned when running the enable azfrontdoorcustomdomainhttps cmdlet but far worse after the cmdlet is run it removes the currently configured cert from the endpoint so rather than us getting an updated cert as expected we are left with a broken config and a cert error is presented to our users
0
433,686
12,508,571,074
IssuesEvent
2020-06-02 15:45:12
LLK/scratch-paint
https://api.github.com/repos/LLK/scratch-paint
closed
Paint editor grid contrast is too low
High Impact Medium Severity priority 2
### Expected Behavior We lowered the contrast of the checkerboard in the background of the paint editor to reduce eye-strain, but now there are complaints that the contrast is too low. ### Actual Behavior There are reports on the forums that it is hard to see when painting with white, or that the grid lines appear to be "missing" on some devices ### Steps to Reproduce View paint editor <img width="665" alt="Screen Shot 2020-05-24 at 23 04 16" src="https://user-images.githubusercontent.com/2855464/82774454-ebbf3c80-9e12-11ea-98dc-3747a48309b6.png"> ### Operating System and Browser all
1.0
Paint editor grid contrast is too low - ### Expected Behavior We lowered the contrast of the checkerboard in the background of the paint editor to reduce eye-strain, but now there are complaints that the contrast is too low. ### Actual Behavior There are reports on the forums that it is hard to see when painting with white, or that the grid lines appear to be "missing" on some devices ### Steps to Reproduce View paint editor <img width="665" alt="Screen Shot 2020-05-24 at 23 04 16" src="https://user-images.githubusercontent.com/2855464/82774454-ebbf3c80-9e12-11ea-98dc-3747a48309b6.png"> ### Operating System and Browser all
non_defect
paint editor grid contrast is too low expected behavior we lowered the contrast of the checkerboard in the background of the paint editor to reduce eye strain but now there are complaints that the contrast is too low actual behavior there are reports on the forums that it is hard to see when painting with white or that the grid lines appear to be missing on some devices steps to reproduce view paint editor img width alt screen shot at src operating system and browser all
0
20,987
3,441,868,298
IssuesEvent
2015-12-14 20:13:28
wdg/blacktree-secrets
https://api.github.com/repos/wdg/blacktree-secrets
closed
Secret #36768 will eat your mac
auto-migrated Priority-Medium Type-Defect
``` There is something wrong with secret #36768. Its datatype is boolean where it would be a string "YES" or "NO". Using a boolean here will result in windows not showing, as described in http://www.macosxhints.com/article.php?story=20090228011201290 My suggestion: Set datatype to "string" Set title to "Window appearance" Set default value to "YES" Set Values to "Leopard" = YES; "Tiger" = NO; Keep dangerous checked, as it won't eat your mac anymore, but it's still not entirely safe. I tried to change the secret, but it's way too old I tried to submit a new secret, but it got marked as a duplicate of #36768 ``` Original issue reported on code.google.com by `yorndej...@gmail.com` on 13 Jul 2010 at 12:21
1.0
Secret #36768 will eat your mac - ``` There is something wrong with secret #36768. Its datatype is boolean where it would be a string "YES" or "NO". Using a boolean here will result in windows not showing, as described in http://www.macosxhints.com/article.php?story=20090228011201290 My suggestion: Set datatype to "string" Set title to "Window appearance" Set default value to "YES" Set Values to "Leopard" = YES; "Tiger" = NO; Keep dangerous checked, as it won't eat your mac anymore, but it's still not entirely safe. I tried to change the secret, but it's way too old I tried to submit a new secret, but it got marked as a duplicate of #36768 ``` Original issue reported on code.google.com by `yorndej...@gmail.com` on 13 Jul 2010 at 12:21
defect
secret will eat your mac there is something wrong with secret its datatype is boolean where it would be a string yes or no using a boolean here will result in windows not showing as described in my suggestion set datatype to string set title to window appearance set default value to yes set values to leopard yes tiger no keep dangerous checked as it won t eat your mac anymore but it s still not entirely safe i tried to change the secret but it s way too old i tried to submit a new secret but it got marked as a duplicate of original issue reported on code google com by yorndej gmail com on jul at
1
148,850
19,552,576,467
IssuesEvent
2022-01-03 01:14:45
madhans23/linux-4.15
https://api.github.com/repos/madhans23/linux-4.15
opened
WS-2021-0566 (Medium) detected in linux-stagingv5.15
security vulnerability
## WS-2021-0566 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stagingv5.15</b></p></summary> <p> <p>hwmon staging tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git>https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> net: hns3: fix use-after-free bug in hclgevf_send_mbx_msg This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven. This ID is fixed in Linux Kernel version v5.15.11 by commit 4f4a353f6fe033807cd026a5de81c67469ff19b0, it was introduced in version v5.8 by commit d8355240cf8fb8b9e002b5c8458578435cea85c2. For more details please see the references link. <p>Publish Date: 2021-12-27 <p>URL: <a href=https://github.com/gregkh/linux/commit/4f4a353f6fe033807cd026a5de81c67469ff19b0>WS-2021-0566</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002741">https://osv.dev/vulnerability/GSD-2021-1002741</a></p> <p>Release Date: 2021-12-27</p> <p>Fix Resolution: v5.15.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0566 (Medium) detected in linux-stagingv5.15 - ## WS-2021-0566 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stagingv5.15</b></p></summary> <p> <p>hwmon staging tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git>https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> net: hns3: fix use-after-free bug in hclgevf_send_mbx_msg This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven. This ID is fixed in Linux Kernel version v5.15.11 by commit 4f4a353f6fe033807cd026a5de81c67469ff19b0, it was introduced in version v5.8 by commit d8355240cf8fb8b9e002b5c8458578435cea85c2. For more details please see the references link. <p>Publish Date: 2021-12-27 <p>URL: <a href=https://github.com/gregkh/linux/commit/4f4a353f6fe033807cd026a5de81c67469ff19b0>WS-2021-0566</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002741">https://osv.dev/vulnerability/GSD-2021-1002741</a></p> <p>Release Date: 2021-12-27</p> <p>Fix Resolution: v5.15.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
ws medium detected in linux ws medium severity vulnerability vulnerable library linux hwmon staging tree library home page a href found in base branch master vulnerable source files drivers net ethernet hisilicon hclgevf mbx c vulnerability details net fix use after free bug in hclgevf send mbx msg this is an automated id intended to aid in discovery of potential security vulnerabilities the actual impact and attack plausibility have not yet been proven this id is fixed in linux kernel version by commit it was introduced in version by commit for more details please see the references link publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
70,552
23,232,456,289
IssuesEvent
2022-08-03 08:50:40
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Can't verify user when option to send keys to verified devices only is selected
T-Defect A-E2EE-SAS-Verification crypto-team
### Steps to reproduce 1. Go to Settings > Security & Privacy 3. Enable 'Encrypt to verified devices only' 4. Go back to a DM with one of your contact, click on his profile 5. Tap on verify. ### Outcome #### What did you expect? That you initiate a verification process. #### What happened instead? It's failing because the recipient won't be able to decrypt the verification request <img width="388" alt="image" src="https://user-images.githubusercontent.com/9841565/182565977-7387b034-7f10-4e07-ad98-97d1236cbb72.png"> So it's quite annoying, I enabled 'Encrypt to verified devices only' and I have unverified users in a room. So you would think that verify them will fix the issue, but you can't ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store 1.4.32 ### Homeserver local synapse ### Will you send logs? No ### Are you willing to provide a PR? Yes
1.0
Can't verify user when option to send keys to verified devices only is selected - ### Steps to reproduce 1. Go to Settings > Security & Privacy 3. Enable 'Encrypt to verified devices only' 4. Go back to a DM with one of your contact, click on his profile 5. Tap on verify. ### Outcome #### What did you expect? That you initiate a verification process. #### What happened instead? It's failing because the recipient won't be able to decrypt the verification request <img width="388" alt="image" src="https://user-images.githubusercontent.com/9841565/182565977-7387b034-7f10-4e07-ad98-97d1236cbb72.png"> So it's quite annoying, I enabled 'Encrypt to verified devices only' and I have unverified users in a room. So you would think that verify them will fix the issue, but you can't ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store 1.4.32 ### Homeserver local synapse ### Will you send logs? No ### Are you willing to provide a PR? Yes
defect
can t verify user when option to send keys to verified devices only is selected steps to reproduce go to settings security privacy enable encrypt to verified devices only go back to a dm with one of your contact click on his profile tap on verify outcome what did you expect that you initiate a verification process what happened instead it s failing because the recipient won t be able to decrypt the verification request img width alt image src so it s quite annoying i enabled encrypt to verified devices only and i have unverified users in a room so you would think that verify them will fix the issue but you can t your phone model no response operating system version no response application version and app store homeserver local synapse will you send logs no are you willing to provide a pr yes
1
315
2,525,200,567
IssuesEvent
2015-01-20 22:51:43
AtlasOfLivingAustralia/biocache-hubs
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-hubs
closed
Occurrence records with invalid layer sampling?
bug priority-medium status-started type-defect
*migrated from:* https://code.google.com/p/ala/issues/detail?id=659 *date:* Thu May 1 22:17:26 2014 *author:* leebel...@gmail.com --- Summaries of occurrence queries from Kristen Williams are suggesting (serious?) anomalies in the sampling of spatial layers of some records during processing. For example, if a query is constructed on Country = n/a AND IBRA = n/a AND IMCRA = n/a AND el884 (bathymetry and elevation) = n/a it should only return records that occur outside Australia. The following query however [http://biocache.ala.org.au/occurrences/search?fq=-ibra%3A*&fq=-imcra%3A*&fq=-country%3A*&fq=-el848%3A*&wkt=POLYGON((96.6%20-58.0,159.8%20-58.0,159.8%20-10.5,96.6%20-10.5,96.6%20-58.0))#tab_mapView](http://biocache.ala.org.au/occurrences/search?fq=-ibra%3A*&fq=-imcra%3A*&fq=-country%3A*&fq=-el848%3A*&wkt=POLYGON((96.6%20-58.0,159.8%20-58.0,159.8%20-10.5,96.6%20-10.5,96.6%20-58.0))#tab_mapView) produces 40,954 records, most of which ARE in 'Australia' and are terrestrial. A few records are marine (correctly or not as I haven't checked yet) but should still be on the bathy/topo layer (and IMCRA). There are also a few on offshore islands, which should also be on the bathy/topo layer, if not IMBRA or IMCRA - and they should also be 'Australia' but that may be another issue. The WKT (bounding box) here just includes Australia and associated islands. PS: Diagnosis of classification of the records would be made easier if `#658` were solved.
1.0
Occurrence records with invalid layer sampling? - *migrated from:* https://code.google.com/p/ala/issues/detail?id=659 *date:* Thu May 1 22:17:26 2014 *author:* leebel...@gmail.com --- Summaries of occurrence queries from Kristen Williams are suggesting (serious?) anomalies in the sampling of spatial layers of some records during processing. For example, if a query is constructed on Country = n/a AND IBRA = n/a AND IMCRA = n/a AND el884 (bathymetry and elevation) = n/a it should only return records that occur outside Australia. The following query however [http://biocache.ala.org.au/occurrences/search?fq=-ibra%3A*&fq=-imcra%3A*&fq=-country%3A*&fq=-el848%3A*&wkt=POLYGON((96.6%20-58.0,159.8%20-58.0,159.8%20-10.5,96.6%20-10.5,96.6%20-58.0))#tab_mapView](http://biocache.ala.org.au/occurrences/search?fq=-ibra%3A*&fq=-imcra%3A*&fq=-country%3A*&fq=-el848%3A*&wkt=POLYGON((96.6%20-58.0,159.8%20-58.0,159.8%20-10.5,96.6%20-10.5,96.6%20-58.0))#tab_mapView) produces 40,954 records, most of which ARE in 'Australia' and are terrestrial. A few records are marine (correctly or not as I haven't checked yet) but should still be on the bathy/topo layer (and IMCRA). There are also a few on offshore islands, which should also be on the bathy/topo layer, if not IMBRA or IMCRA - and they should also be 'Australia' but that may be another issue. The WKT (bounding box) here just includes Australia and associated islands. PS: Diagnosis of classification of the records would be made easier if `#658` were solved.
defect
occurrence records with invalid layer sampling migrated from date thu may author leebel gmail com summaries of occurrence queries from kristen williams are suggesting serious anomalies in the sampling of spatial layers of some records during processing for example if a query is constructed on country n a and ibra n a and imcra n a and bathymetry and elevation n a it should only return records that occur outside australia the following query however produces records most of which are in australia and are terrestrial a few records are marine correctly or not as i haven t checked yet but should still be on the bathy topo layer and imcra there are also a few on offshore islands which should also be on the bathy topo layer if not imbra or imcra and they should also be australia but that may be another issue the wkt bounding box here just includes australia and associated islands ps diagnosis of classification of the records would be made easier if were solved
1
29,192
14,003,839,126
IssuesEvent
2020-10-28 16:21:25
apache/trafficcontrol
https://api.github.com/repos/apache/trafficcontrol
opened
Reduce the number of DB requests made in ValidateServerCapabilities()
Traffic Ops good first issue improvement performance
## I'm submitting a ... - improvement request (usability, performance, tech debt, etc.) ## Traffic Control components affected ... - Traffic Ops ## Current behavior: TO currently makes a DB request per server in the `ValidateServerCapabilities()` function: https://github.com/apache/trafficcontrol/blob/b99eda51b03505f48d67aadf481d16470455a7f4/traffic_ops/traffic_ops_golang/deliveryservice/servers/servers.go#L541. When a delivery service is being assigned to a large number of servers, this slows down the total execution time unnecessarily. ## New behavior: Instead of making a DB request per server, the function should make a single DB request to get data for all the servers at once, then perform the validation. This will make it much faster to assign a delivery service to large numbers of servers at a time. ## Minimal reproduction of the problem with instructions: Assign a delivery service with required capabilities to a large number of servers with those capabilities at a time.
True
Reduce the number of DB requests made in ValidateServerCapabilities() - ## I'm submitting a ... - improvement request (usability, performance, tech debt, etc.) ## Traffic Control components affected ... - Traffic Ops ## Current behavior: TO currently makes a DB request per server in the `ValidateServerCapabilities()` function: https://github.com/apache/trafficcontrol/blob/b99eda51b03505f48d67aadf481d16470455a7f4/traffic_ops/traffic_ops_golang/deliveryservice/servers/servers.go#L541. When a delivery service is being assigned to a large number of servers, this slows down the total execution time unnecessarily. ## New behavior: Instead of making a DB request per server, the function should make a single DB request to get data for all the servers at once, then perform the validation. This will make it much faster to assign a delivery service to large numbers of servers at a time. ## Minimal reproduction of the problem with instructions: Assign a delivery service with required capabilities to a large number of servers with those capabilities at a time.
non_defect
reduce the number of db requests made in validateservercapabilities i m submitting a improvement request usability performance tech debt etc traffic control components affected traffic ops current behavior to currently makes a db request per server in the validateservercapabilities function when a delivery service is being assigned to a large number of servers this slows down the total execution time unnecessarily new behavior instead of making a db request per server the function should make a single db request to get data for all the servers at once then perform the validation this will make it much faster to assign a delivery service to large numbers of servers at a time minimal reproduction of the problem with instructions assign a delivery service with required capabilities to a large number of servers with those capabilities at a time
0
77,850
27,193,790,096
IssuesEvent
2023-02-20 02:12:44
hyperledger/iroha
https://api.github.com/repos/hyperledger/iroha
closed
[BUG] `This peer is faulty` after 1.5 hours of iroha downtime.
Bug iroha2 Dev defect QA-confirmed
### OS and Environment MacOS, Docker Hub ### GIT commit hash 4ee707a3 ### Minimum working example / Steps to reproduce 1. `docker compose up` 2. Just wait 1.5 hours ![image](https://user-images.githubusercontent.com/111361420/219669674-f40ce383-e46f-4ed1-ba8b-17e10a87fd99.png) ### Actual result ```bash iroha-iroha1-1 | 2023-02-17T12:09:46.555187Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader iroha-iroha0-1 | 2023-02-17T12:09:46.674023Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer iroha-iroha2-1 | 2023-02-17T12:09:48.588423Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail iroha-iroha3-1 | 2023-02-17T12:09:49.220245Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer iroha-iroha1-1 | 2023-02-17T12:09:49.573967Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader iroha-iroha0-1 | 2023-02-17T12:09:49.675306Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer iroha-iroha0-1 | 2023-02-17T12:09:50.645770Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:50.764237Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:51.625035Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail iroha-iroha3-1 | 2023-02-17T12:09:52.171543Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer iroha-iroha1-1 | 2023-02-17T12:09:52.515576Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader iroha-iroha0-1 | 2023-02-17T12:09:52.615514Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer iroha-iroha1-1 | 2023-02-17T12:09:52.687375Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha3-1 | 2023-02-17T12:09:52.694003Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha1-1 | 2023-02-17T12:09:53.232625Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha0-1 | 2023-02-17T12:09:53.242050Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:53.591693Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha0-1 | 2023-02-17T12:09:53.599164Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:53.703294Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha3-1 | 2023-02-17T12:09:53.709998Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:54.578374Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail iroha-iroha3-1 | 2023-02-17T12:09:55.124516Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer iroha-iroha1-1 | 2023-02-17T12:09:55.458997Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader iroha-iroha0-1 | 2023-02-17T12:09:55.575338Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer iroha-iroha1-1 | 2023-02-17T12:09:55.613468Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha3-1 | 2023-02-17T12:09:55.659568Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel ``` ### Expected result No errors during downtime. ### Who can help to reproduce? @astrokov7
1.0
[BUG] `This peer is faulty` after 1.5 hours of iroha downtime. - ### OS and Environment MacOS, Docker Hub ### GIT commit hash 4ee707a3 ### Minimum working example / Steps to reproduce 1. `docker compose up` 2. Just wait 1.5 hours ![image](https://user-images.githubusercontent.com/111361420/219669674-f40ce383-e46f-4ed1-ba8b-17e10a87fd99.png) ### Actual result ```bash iroha-iroha1-1 | 2023-02-17T12:09:46.555187Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader iroha-iroha0-1 | 2023-02-17T12:09:46.674023Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer iroha-iroha2-1 | 2023-02-17T12:09:48.588423Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail iroha-iroha3-1 | 2023-02-17T12:09:49.220245Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer iroha-iroha1-1 | 2023-02-17T12:09:49.573967Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader iroha-iroha0-1 | 2023-02-17T12:09:49.675306Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer iroha-iroha0-1 | 2023-02-17T12:09:50.645770Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:50.764237Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:51.625035Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail iroha-iroha3-1 | 2023-02-17T12:09:52.171543Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer iroha-iroha1-1 | 2023-02-17T12:09:52.515576Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader iroha-iroha0-1 | 2023-02-17T12:09:52.615514Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer iroha-iroha1-1 | 2023-02-17T12:09:52.687375Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha3-1 | 2023-02-17T12:09:52.694003Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha1-1 | 2023-02-17T12:09:53.232625Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha0-1 | 2023-02-17T12:09:53.242050Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:53.591693Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha0-1 | 2023-02-17T12:09:53.599164Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:53.703294Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha3-1 | 2023-02-17T12:09:53.709998Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha2-1 | 2023-02-17T12:09:54.578374Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail iroha-iroha3-1 | 2023-02-17T12:09:55.124516Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer iroha-iroha1-1 | 2023-02-17T12:09:55.458997Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader iroha-iroha0-1 | 2023-02-17T12:09:55.575338Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer iroha-iroha1-1 | 2023-02-17T12:09:55.613468Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel iroha-iroha3-1 | 2023-02-17T12:09:55.659568Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel ``` ### Expected result No errors during downtime. ### Who can help to reproduce? @astrokov7
defect
this peer is faulty after hours of iroha downtime os and environment macos docker hub git commit hash minimum working example steps to reproduce docker compose up just wait hours actual result bash iroha info run iroha core sumeragi main loop no new transactions requesting view change role leader iroha info run iroha core sumeragi main loop no new transactions requesting view change role validatingpeer iroha info run iroha core sumeragi main loop no new transactions requesting view change role proxytail iroha info run iroha core sumeragi main loop no new transactions requesting view change role observingpeer iroha info run iroha core sumeragi main loop no new transactions requesting view change role leader iroha info run iroha core sumeragi main loop no new transactions requesting view change role validatingpeer iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha info run iroha core sumeragi main loop no new transactions requesting view change role proxytail iroha info run iroha core sumeragi main loop no new transactions requesting view change role observingpeer iroha info run iroha core sumeragi main loop no new transactions requesting view change role leader iroha info run iroha core sumeragi main loop no new transactions requesting view change role validatingpeer iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha info run iroha core sumeragi main loop no new transactions requesting view change role proxytail iroha info run iroha core sumeragi main loop no new transactions requesting view change role observingpeer iroha info run iroha core sumeragi main loop no new transactions requesting view change role leader iroha info run iroha core sumeragi main loop no new transactions requesting view change role validatingpeer iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel expected result no errors during downtime who can help to reproduce
1
249,596
7,963,887,107
IssuesEvent
2018-07-13 19:14:15
phetsims/equality-explorer
https://api.github.com/repos/phetsims/equality-explorer
closed
disable showAnswers in production versions
priority:5-deferred
phetsims/joist#406 is a prerequisite to this. In EqualityExplorerQueryParameters, set `showAnswers` to false if we're running a production version, so that answers can't be revealed.
1.0
disable showAnswers in production versions - phetsims/joist#406 is a prerequisite to this. In EqualityExplorerQueryParameters, set `showAnswers` to false if we're running a production version, so that answers can't be revealed.
non_defect
disable showanswers in production versions phetsims joist is a prerequisite to this in equalityexplorerqueryparameters set showanswers to false if we re running a production version so that answers can t be revealed
0
79,635
28,494,306,324
IssuesEvent
2023-04-18 13:17:26
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Cannot send message
T-Defect X-Regression S-Major O-Occasional
### Steps to reproduce 1. Be in an encrypted room 2. Try to send a message ### Outcome #### What did you expect? Message sent #### What happened instead? Error ``` caught (in promise) TypeError: Cannot read properties of null (reading 'getEvents') at TimelineWindow.getEvents (timeline-window.ts:375:37) at TimelinePanel_TimelinePanel.getEvents (TimelinePanel.tsx:1546:64) at TimelinePanel_TimelinePanel.reloadEvents (TimelinePanel.tsx:1533:28) at MatrixClient.<anonymous> (TimelinePanel.tsx:882:14) at MatrixClient.emit (events.js:158:1) at MatrixClient.emit (typed-event-emitter.ts:59:22) at Room.forSource (ReEmitter.ts:53:29) at Room.emit (events.js:158:1) at Room.emit (typed-event-emitter.ts:59:22) at Room.addPendingEvent (room.ts:2411:14) ``` Got this a couple of times today. ### Operating system Ubuntu 22.04.2 LTS ### Application version Element Nightly version: 2023041201 Olm version: 3.2.12 ### How did you install the app? https://packages.element.io/debian/ ### Homeserver https://element.ems.host ### Will you send logs? Yes
1.0
Cannot send message - ### Steps to reproduce 1. Be in an encrypted room 2. Try to send a message ### Outcome #### What did you expect? Message sent #### What happened instead? Error ``` caught (in promise) TypeError: Cannot read properties of null (reading 'getEvents') at TimelineWindow.getEvents (timeline-window.ts:375:37) at TimelinePanel_TimelinePanel.getEvents (TimelinePanel.tsx:1546:64) at TimelinePanel_TimelinePanel.reloadEvents (TimelinePanel.tsx:1533:28) at MatrixClient.<anonymous> (TimelinePanel.tsx:882:14) at MatrixClient.emit (events.js:158:1) at MatrixClient.emit (typed-event-emitter.ts:59:22) at Room.forSource (ReEmitter.ts:53:29) at Room.emit (events.js:158:1) at Room.emit (typed-event-emitter.ts:59:22) at Room.addPendingEvent (room.ts:2411:14) ``` Got this a couple of times today. ### Operating system Ubuntu 22.04.2 LTS ### Application version Element Nightly version: 2023041201 Olm version: 3.2.12 ### How did you install the app? https://packages.element.io/debian/ ### Homeserver https://element.ems.host ### Will you send logs? Yes
defect
cannot send message steps to reproduce be in an encrypted room try to send a message outcome what did you expect message sent what happened instead error caught in promise typeerror cannot read properties of null reading getevents at timelinewindow getevents timeline window ts at timelinepanel timelinepanel getevents timelinepanel tsx at timelinepanel timelinepanel reloadevents timelinepanel tsx at matrixclient timelinepanel tsx at matrixclient emit events js at matrixclient emit typed event emitter ts at room forsource reemitter ts at room emit events js at room emit typed event emitter ts at room addpendingevent room ts got this a couple of times today operating system ubuntu lts application version element nightly version olm version how did you install the app homeserver will you send logs yes
1
11,964
18,722,503,454
IssuesEvent
2021-11-03 13:20:34
alper-turgal/SWE-573
https://api.github.com/repos/alper-turgal/SWE-573
closed
Write the first full version of requirements
requirements-task
- Review the preliminary requirements - Delete the ones rejected by the product owner - Add the confirmed requirements - Add some requirements about the admin
1.0
Write the first full version of requirements - - Review the preliminary requirements - Delete the ones rejected by the product owner - Add the confirmed requirements - Add some requirements about the admin
non_defect
write the first full version of requirements review the preliminary requirements delete the ones rejected by the product owner add the confirmed requirements add some requirements about the admin
0
14,211
17,094,345,025
IssuesEvent
2021-07-08 22:33:23
jenkinsci/configuration-as-code-plugin
https://api.github.com/repos/jenkinsci/configuration-as-code-plugin
opened
Unable to login to Jenkins after Jenkins upgrade to 2.289.2
plugin-compatibility
HI, After upgrading the Jenkins version to 2.289.2, the jenkins is not working as it throws the below error java.lang.IllegalArgumentException: Permission can not be null for sid:{USER ID} There was no changes to the Permission Matrix
True
Unable to login to Jenkins after Jenkins upgrade to 2.289.2 - HI, After upgrading the Jenkins version to 2.289.2, the jenkins is not working as it throws the below error java.lang.IllegalArgumentException: Permission can not be null for sid:{USER ID} There was no changes to the Permission Matrix
non_defect
unable to login to jenkins after jenkins upgrade to hi after upgrading the jenkins version to the jenkins is not working as it throws the below error java lang illegalargumentexception permission can not be null for sid user id there was no changes to the permission matrix
0
338,327
10,227,503,581
IssuesEvent
2019-08-16 20:59:46
Alluxio/alluxio
https://api.github.com/repos/Alluxio/alluxio
closed
Display three the same records in BlockInfo tab of worker node WEBUI
area-webui priority-high target-2.0.1 type-bug
**Alluxio Version:** 2.1.0-SNAPSHOT **Describe the bug** when entering page of **BlockInfo** of worker node , each file has three the same records in WEBUI. clicked the link of file which three the same records, display the same information in WEBUI. ![图片](https://user-images.githubusercontent.com/303954/61579980-1ed3d800-ab3f-11e9-8273-8a1cace55f4b.png) ![图片](https://user-images.githubusercontent.com/303954/61579997-462aa500-ab3f-11e9-8653-adc78ce2065f.png) **To Reproduce** enter page of BlockInfo of worker node **Expected behavior** A clear and concise description of what you expected to happen. **Urgency** Describe the impact and urgency of the bug. **Additional context** Add any other context about the problem here.
1.0
Display three the same records in BlockInfo tab of worker node WEBUI - **Alluxio Version:** 2.1.0-SNAPSHOT **Describe the bug** when entering page of **BlockInfo** of worker node , each file has three the same records in WEBUI. clicked the link of file which three the same records, display the same information in WEBUI. ![图片](https://user-images.githubusercontent.com/303954/61579980-1ed3d800-ab3f-11e9-8273-8a1cace55f4b.png) ![图片](https://user-images.githubusercontent.com/303954/61579997-462aa500-ab3f-11e9-8653-adc78ce2065f.png) **To Reproduce** enter page of BlockInfo of worker node **Expected behavior** A clear and concise description of what you expected to happen. **Urgency** Describe the impact and urgency of the bug. **Additional context** Add any other context about the problem here.
non_defect
display three the same records in blockinfo tab of worker node webui alluxio version snapshot describe the bug when entering page of blockinfo of worker node each file has three the same records in webui clicked the link of file which three the same records display the same information in webui to reproduce enter page of blockinfo of worker node expected behavior a clear and concise description of what you expected to happen urgency describe the impact and urgency of the bug additional context add any other context about the problem here
0
86,013
16,813,310,355
IssuesEvent
2021-06-17 02:37:58
QLou895/Terrified_2.0
https://api.github.com/repos/QLou895/Terrified_2.0
opened
Peripheral
Code Dev eyeTrackAPI
### Eye Tracking Interaction - [ ] Implement peripheral stuff through a cross-hair for initial interface & testing - [ ] Acquire face mesh (especially targeting eyes to affect screen) - [ ] User "blinks" -> create input: (suggestion to darken room and exaggerate shadows and _something comes alive_) - [ ] See monster in peripheral -> look at monster & disappears - [ ] Looking Left -> something appears or looks like it occurred on the Right in peripheral - [ ]
1.0
Peripheral - ### Eye Tracking Interaction - [ ] Implement peripheral stuff through a cross-hair for initial interface & testing - [ ] Acquire face mesh (especially targeting eyes to affect screen) - [ ] User "blinks" -> create input: (suggestion to darken room and exaggerate shadows and _something comes alive_) - [ ] See monster in peripheral -> look at monster & disappears - [ ] Looking Left -> something appears or looks like it occurred on the Right in peripheral - [ ]
non_defect
peripheral eye tracking interaction implement peripheral stuff through a cross hair for initial interface testing acquire face mesh especially targeting eyes to affect screen user blinks create input suggestion to darken room and exaggerate shadows and something comes alive see monster in peripheral look at monster disappears looking left something appears or looks like it occurred on the right in peripheral
0
51,214
13,207,395,671
IssuesEvent
2020-08-14 22:56:38
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
use JEBEventServiceIsolator.PopEvent(...) in JEBEventService (Trac #51)
Incomplete Migration Migrated from Trac defect jeb-server
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/51">https://code.icecube.wisc.edu/projects/icecube/ticket/51</a>, reported by tschmidtand owned by tschmidt</em></summary> <p> ```json { "status": "closed", "changetime": "2007-11-09T22:48:41", "_ts": "1194648521000000", "description": "Add method PopEvent(...) to JEBEventServiceIsolator and use it in JEBEventService instead of directly retrieving an event from the event service. it's working like it is done, but with using the isolator, one decouples the event service much better. Note: the event serice may not be threadsafe.\n", "reporter": "tschmidt", "cc": "", "resolution": "duplicate", "time": "2007-06-07T15:17:33", "component": "jeb-server", "summary": "use JEBEventServiceIsolator.PopEvent(...) in JEBEventService", "priority": "normal", "keywords": "", "milestone": "", "owner": "tschmidt", "type": "defect" } ``` </p> </details>
1.0
use JEBEventServiceIsolator.PopEvent(...) in JEBEventService (Trac #51) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/51">https://code.icecube.wisc.edu/projects/icecube/ticket/51</a>, reported by tschmidtand owned by tschmidt</em></summary> <p> ```json { "status": "closed", "changetime": "2007-11-09T22:48:41", "_ts": "1194648521000000", "description": "Add method PopEvent(...) to JEBEventServiceIsolator and use it in JEBEventService instead of directly retrieving an event from the event service. it's working like it is done, but with using the isolator, one decouples the event service much better. Note: the event serice may not be threadsafe.\n", "reporter": "tschmidt", "cc": "", "resolution": "duplicate", "time": "2007-06-07T15:17:33", "component": "jeb-server", "summary": "use JEBEventServiceIsolator.PopEvent(...) in JEBEventService", "priority": "normal", "keywords": "", "milestone": "", "owner": "tschmidt", "type": "defect" } ``` </p> </details>
defect
use jebeventserviceisolator popevent in jebeventservice trac migrated from json status closed changetime ts description add method popevent to jebeventserviceisolator and use it in jebeventservice instead of directly retrieving an event from the event service it s working like it is done but with using the isolator one decouples the event service much better note the event serice may not be threadsafe n reporter tschmidt cc resolution duplicate time component jeb server summary use jebeventserviceisolator popevent in jebeventservice priority normal keywords milestone owner tschmidt type defect
1
2,054
2,603,975,600
IssuesEvent
2015-02-24 19:01:24
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳湿尤能治疗
auto-migrated Priority-Medium Type-Defect
``` 沈阳湿尤能治疗〓沈陽軍區政治部醫院性病〓TEL:024-31023308�� �成立于1946年,68年專注于性傳播疾病的研究和治療。位于沈� ��市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史� ��久、設備精良、技術權威、專家云集,是預防、保健、醫療 、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊�� �院、全國首批醫療規范定點單位,是第四軍醫大學、東南大� ��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤 部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功�� � ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:17
1.0
沈阳湿尤能治疗 - ``` 沈阳湿尤能治疗〓沈陽軍區政治部醫院性病〓TEL:024-31023308�� �成立于1946年,68年專注于性傳播疾病的研究和治療。位于沈� ��市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史� ��久、設備精良、技術權威、專家云集,是預防、保健、醫療 、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊�� �院、全國首批醫療規范定點單位,是第四軍醫大學、東南大� ��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤 部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功�� � ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:17
defect
沈阳湿尤能治疗 沈阳湿尤能治疗〓沈陽軍區政治部醫院性病〓tel: �� � , 。位于沈� �� 。是一所與新中國同建立共輝煌的歷史� ��久、設備精良、技術權威、專家云集,是預防、保健、醫療 、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊�� �院、全國首批醫療規范定點單位,是第四軍醫大學、東南大� ��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤 部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功�� � original issue reported on code google com by gmail com on jun at
1
2,734
3,006,423,510
IssuesEvent
2015-07-27 10:17:10
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
opened
Opencv 2.4.8, MinGW, cv::CascadeClassifier();
affected: 2.4 auto-transferred bug category: build/install priority: normal
Transferred from http://code.opencv.org/issues/3593 ``` || Denis Gottardello on 2014-03-05 17:06 || Priority: Normal || Affected: 2.4.8 (latest release) || Category: build/install || Tracker: Bug || Difficulty: || PR: || Platform: x86 / Windows ``` Opencv 2.4.8, MinGW, cv::CascadeClassifier(); ----------- ``` In order to build OpenCV with MinGW32 is necessary to apply the patch described here: http://stackoverflow.com/questions/21103042/error-while-building-opencv-monitorfromrect-was-not-declared-in-this-scope But the serious problem is that the object *"cv::CascadeClassifier()"* is unusable. The row *"cv::CascadeClassifier *a= new cv::CascadeClassifier();"* or *"cv::CascadeClassifier a;"* causes an application crash! On Linux and Mac all works well. ``` History ------- ##### Denis Gottardello on 2014-03-05 17:07 ``` The problem happens also with BackgroundSubtractorMOG2 and BackgroundSubtractorMOG. ``` ##### Dinar Ahmatnurov on 2014-03-05 17:16 ``` Roman. please have a look - Priority changed from Blocker to Normal - Assignee set to Roman Donchenko - Category set to build/install ``` ##### Dinar Ahmatnurov on 2014-03-05 17:17 ``` - Status changed from New to Open ``` ##### Denis Gottardello on 2014-06-17 19:49 ``` Is this bug still open? ``` ##### Denis Gottardello on 2015-01-01 17:58 ``` Up! ```
1.0
Opencv 2.4.8, MinGW, cv::CascadeClassifier(); - Transferred from http://code.opencv.org/issues/3593 ``` || Denis Gottardello on 2014-03-05 17:06 || Priority: Normal || Affected: 2.4.8 (latest release) || Category: build/install || Tracker: Bug || Difficulty: || PR: || Platform: x86 / Windows ``` Opencv 2.4.8, MinGW, cv::CascadeClassifier(); ----------- ``` In order to build OpenCV with MinGW32 is necessary to apply the patch described here: http://stackoverflow.com/questions/21103042/error-while-building-opencv-monitorfromrect-was-not-declared-in-this-scope But the serious problem is that the object *"cv::CascadeClassifier()"* is unusable. The row *"cv::CascadeClassifier *a= new cv::CascadeClassifier();"* or *"cv::CascadeClassifier a;"* causes an application crash! On Linux and Mac all works well. ``` History ------- ##### Denis Gottardello on 2014-03-05 17:07 ``` The problem happens also with BackgroundSubtractorMOG2 and BackgroundSubtractorMOG. ``` ##### Dinar Ahmatnurov on 2014-03-05 17:16 ``` Roman. please have a look - Priority changed from Blocker to Normal - Assignee set to Roman Donchenko - Category set to build/install ``` ##### Dinar Ahmatnurov on 2014-03-05 17:17 ``` - Status changed from New to Open ``` ##### Denis Gottardello on 2014-06-17 19:49 ``` Is this bug still open? ``` ##### Denis Gottardello on 2015-01-01 17:58 ``` Up! ```
non_defect
opencv mingw cv cascadeclassifier transferred from denis gottardello on priority normal affected latest release category build install tracker bug difficulty pr platform windows opencv mingw cv cascadeclassifier in order to build opencv with is necessary to apply the patch described here but the serious problem is that the object cv cascadeclassifier is unusable the row cv cascadeclassifier a new cv cascadeclassifier or cv cascadeclassifier a causes an application crash on linux and mac all works well history denis gottardello on the problem happens also with and backgroundsubtractormog dinar ahmatnurov on roman please have a look priority changed from blocker to normal assignee set to roman donchenko category set to build install dinar ahmatnurov on status changed from new to open denis gottardello on is this bug still open denis gottardello on up
0
4,565
2,610,117,963
IssuesEvent
2015-02-26 18:36:38
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
text not copied when invoking Scribefire for Safari
auto-migrated Priority-Medium Type-Defect
``` What's the problem? When I select text on a page and invoke the ScribeFire extension, the text is not copied into the blog body. What browser are you using? Safari Version 5.0.1 (6533.17.8) on Mac OSX 10.6.4 What version of ScribeFire are you running? 1.2.0.0 ``` ----- Original issue reported on code.google.com by `fiver.lo...@gmail.com` on 7 Aug 2010 at 10:45
1.0
text not copied when invoking Scribefire for Safari - ``` What's the problem? When I select text on a page and invoke the ScribeFire extension, the text is not copied into the blog body. What browser are you using? Safari Version 5.0.1 (6533.17.8) on Mac OSX 10.6.4 What version of ScribeFire are you running? 1.2.0.0 ``` ----- Original issue reported on code.google.com by `fiver.lo...@gmail.com` on 7 Aug 2010 at 10:45
defect
text not copied when invoking scribefire for safari what s the problem when i select text on a page and invoke the scribefire extension the text is not copied into the blog body what browser are you using safari version on mac osx what version of scribefire are you running original issue reported on code google com by fiver lo gmail com on aug at
1
10,290
26,671,665,199
IssuesEvent
2023-01-26 10:49:20
MicrosoftDocs/architecture-center
https://api.github.com/repos/MicrosoftDocs/architecture-center
closed
Conflicting statement on patching AKS nodes
doc-enhancement assigned-to-author triaged architecture-center/svc reference-architecture/subsvc Pri2
[As per current document security pacthes doesnt get automaticlly applied to nodes by default , in similar document its mentioned that each evening linux node in AKS get sec pacthes vai distro security update and that behaviour is automatically configured as nodes deployed to AKS cluster. Refernce doc links: https://learn.microsoft.com/en-us/azure/aks/concepts-security#linux-nodes] --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 43da9bdf-a3d3-a65b-b555-60dcc5ac2c55 * Version Independent ID: 9913c705-aeb7-e154-4fdf-164ff1e7ff99 * Content: [AKS Day-2 - Patch and upgrade guidance - Azure Architecture Center](https://learn.microsoft.com/en-us/azure/architecture/operator-guides/aks/aks-upgrade-practices) * Content Source: [docs/operator-guides/aks/aks-upgrade-practices.md](https://github.com/microsoftdocs/architecture-center/blob/main/docs/operator-guides/aks/aks-upgrade-practices.md) * Service: **architecture-center** * Sub-service: **reference-architecture** * GitHub Login: @rishabhsaha * Microsoft Alias: **pnp**
2.0
Conflicting statement on patching AKS nodes - [As per current document security pacthes doesnt get automaticlly applied to nodes by default , in similar document its mentioned that each evening linux node in AKS get sec pacthes vai distro security update and that behaviour is automatically configured as nodes deployed to AKS cluster. Refernce doc links: https://learn.microsoft.com/en-us/azure/aks/concepts-security#linux-nodes] --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 43da9bdf-a3d3-a65b-b555-60dcc5ac2c55 * Version Independent ID: 9913c705-aeb7-e154-4fdf-164ff1e7ff99 * Content: [AKS Day-2 - Patch and upgrade guidance - Azure Architecture Center](https://learn.microsoft.com/en-us/azure/architecture/operator-guides/aks/aks-upgrade-practices) * Content Source: [docs/operator-guides/aks/aks-upgrade-practices.md](https://github.com/microsoftdocs/architecture-center/blob/main/docs/operator-guides/aks/aks-upgrade-practices.md) * Service: **architecture-center** * Sub-service: **reference-architecture** * GitHub Login: @rishabhsaha * Microsoft Alias: **pnp**
non_defect
conflicting statement on patching aks nodes as per current document security pacthes doesnt get automaticlly applied to nodes by default in similar document its mentioned that each evening linux node in aks get sec pacthes vai distro security update and that behaviour is automatically configured as nodes deployed to aks cluster refernce doc links document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service reference architecture github login rishabhsaha microsoft alias pnp
0
251,074
7,999,466,359
IssuesEvent
2018-07-22 01:47:33
Zicerite/Gavania-Project
https://api.github.com/repos/Zicerite/Gavania-Project
opened
Make Gunman's Araw rune
Medium Priority Skill Related
Rune 0 = Pistol. 12 ammo, reload speed = 1.5s Rune 1 = Shotgun, 6 ammo, reload speed = 4s Rune 2 = Burst Shot, 15 ammo, 3 shots/burst, reload speed = 2.5s
1.0
Make Gunman's Araw rune - Rune 0 = Pistol. 12 ammo, reload speed = 1.5s Rune 1 = Shotgun, 6 ammo, reload speed = 4s Rune 2 = Burst Shot, 15 ammo, 3 shots/burst, reload speed = 2.5s
non_defect
make gunman s araw rune rune pistol ammo reload speed rune shotgun ammo reload speed rune burst shot ammo shots burst reload speed
0
346,522
24,886,950,537
IssuesEvent
2022-10-28 08:36:43
songivan00/ped
https://api.github.com/repos/songivan00/ped
opened
Suggestion for documentation and feature improvement for find
severity.Medium type.DocumentationBug
The boxed portion of the find feature seems to suggest that only the full name is allowed in this find feature, however, the find feature still works with the complete first name or last name. Maybe using 'full name' might not be the right term. Also, I feel that an improvement to this feature could be to allow substring of the name to work too! ![Screenshot 2022-10-28 at 4.30.18 PM.png](https://raw.githubusercontent.com/songivan00/ped/main/files/1fa61e01-212d-403e-a3e8-54b14e107552.png) <!--session: 1666943941242-6358995f-a616-4a7c-a8f1-6a6e2d00776b--> <!--Version: Web v3.4.4-->
1.0
Suggestion for documentation and feature improvement for find - The boxed portion of the find feature seems to suggest that only the full name is allowed in this find feature, however, the find feature still works with the complete first name or last name. Maybe using 'full name' might not be the right term. Also, I feel that an improvement to this feature could be to allow substring of the name to work too! ![Screenshot 2022-10-28 at 4.30.18 PM.png](https://raw.githubusercontent.com/songivan00/ped/main/files/1fa61e01-212d-403e-a3e8-54b14e107552.png) <!--session: 1666943941242-6358995f-a616-4a7c-a8f1-6a6e2d00776b--> <!--Version: Web v3.4.4-->
non_defect
suggestion for documentation and feature improvement for find the boxed portion of the find feature seems to suggest that only the full name is allowed in this find feature however the find feature still works with the complete first name or last name maybe using full name might not be the right term also i feel that an improvement to this feature could be to allow substring of the name to work too
0
491,646
14,168,372,948
IssuesEvent
2020-11-12 11:38:15
gnosis/conditional-tokens-explorer
https://api.github.com/repos/gnosis/conditional-tokens-explorer
closed
[UI/UX] Search option is not fully colored (some missing pixels at the top and from the right border)
Low priority bug
See comment https://github.com/gnosis/conditional-tokens-explorer/pull/571#issuecomment-723940788 Point 7 ![98675299-212cf780-236b-11eb-940e-05359b82eee1](https://user-images.githubusercontent.com/1144028/98731477-8c32fa00-237c-11eb-83b8-a3673b26ae92.jpg)
1.0
[UI/UX] Search option is not fully colored (some missing pixels at the top and from the right border) - See comment https://github.com/gnosis/conditional-tokens-explorer/pull/571#issuecomment-723940788 Point 7 ![98675299-212cf780-236b-11eb-940e-05359b82eee1](https://user-images.githubusercontent.com/1144028/98731477-8c32fa00-237c-11eb-83b8-a3673b26ae92.jpg)
non_defect
search option is not fully colored some missing pixels at the top and from the right border see comment point
0
269,229
28,960,038,850
IssuesEvent
2023-05-10 01:10:16
dreamboy9/mongo
https://api.github.com/repos/dreamboy9/mongo
reopened
CVE-2022-0686 (High) detected in url-parse-1.5.1.tgz
Mend: dependency security vulnerability
## CVE-2022-0686 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p> <p>Path to dependency file: /buildscripts/libdeps/graph_visualizer_web_stack/package.json</p> <p>Path to vulnerable library: /buildscripts/libdeps/graph_visualizer_web_stack/node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - webpack-dev-server-3.11.1.tgz - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dreamboy9/mongo/commit/60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b">60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.8. <p>Publish Date: 2022-02-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0686>CVE-2022-0686</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686</a></p> <p>Release Date: 2022-02-20</p> <p>Fix Resolution (url-parse): 1.5.8</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0686 (High) detected in url-parse-1.5.1.tgz - ## CVE-2022-0686 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p> <p>Path to dependency file: /buildscripts/libdeps/graph_visualizer_web_stack/package.json</p> <p>Path to vulnerable library: /buildscripts/libdeps/graph_visualizer_web_stack/node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - webpack-dev-server-3.11.1.tgz - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dreamboy9/mongo/commit/60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b">60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.8. <p>Publish Date: 2022-02-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0686>CVE-2022-0686</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686</a></p> <p>Release Date: 2022-02-20</p> <p>Fix Resolution (url-parse): 1.5.8</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in url parse tgz cve high severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file buildscripts libdeps graph visualizer web stack package json path to vulnerable library buildscripts libdeps graph visualizer web stack node modules url parse package json dependency hierarchy react scripts tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution react scripts step up your open source security game with mend
0
46,407
13,055,908,758
IssuesEvent
2020-07-30 03:05:22
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
CoincSuite split_recombine.py example doesn't run (Trac #1162)
Incomplete Migration Migrated from Trac combo reconstruction defect
Migrated from https://code.icecube.wisc.edu/ticket/1162 ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "Here is the error message:\n\n\n{{{\nINFO (Python): Using CoincSuite Recombinations (coincsuite.py:82 in Complete)\nTraceback (most recent call last):\n File \"split_recombine.py\", line 77, in <module>\n Split_Recombine( tray, \"Split_Recombine\", params)\n File \"split_recombine.py\", line 58, in Split_Recombine\n SplitPulses = \"MaskedOfflinePulses\")\n File \"/home/jtatar/StrikeTeam/IceRec/build/lib/I3Tray.py\", line 204, in AddSegment\n return _segment(self, _name, **kwargs)\n File \"/home/jtatar/StrikeTeam/IceRec/build/lib/icecube/CoincSuite/coincsuite.py\", line 169, in Complete\n mininame = lilliput.add_minuit_simplex_minimizer_service(tray)\nNameError: global name 'lilliput' is not defined\n\n}}}\n\n* Default input file does not exist.\n* Please add a better summary of what script does at top of script.", "reporter": "jtatar", "cc": "", "resolution": "fixed", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "CoincSuite split_recombine.py example doesn't run", "priority": "blocker", "keywords": "", "time": "2015-08-18T18:26:13", "milestone": "", "owner": "mzoll", "type": "defect" } ```
1.0
CoincSuite split_recombine.py example doesn't run (Trac #1162) - Migrated from https://code.icecube.wisc.edu/ticket/1162 ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "Here is the error message:\n\n\n{{{\nINFO (Python): Using CoincSuite Recombinations (coincsuite.py:82 in Complete)\nTraceback (most recent call last):\n File \"split_recombine.py\", line 77, in <module>\n Split_Recombine( tray, \"Split_Recombine\", params)\n File \"split_recombine.py\", line 58, in Split_Recombine\n SplitPulses = \"MaskedOfflinePulses\")\n File \"/home/jtatar/StrikeTeam/IceRec/build/lib/I3Tray.py\", line 204, in AddSegment\n return _segment(self, _name, **kwargs)\n File \"/home/jtatar/StrikeTeam/IceRec/build/lib/icecube/CoincSuite/coincsuite.py\", line 169, in Complete\n mininame = lilliput.add_minuit_simplex_minimizer_service(tray)\nNameError: global name 'lilliput' is not defined\n\n}}}\n\n* Default input file does not exist.\n* Please add a better summary of what script does at top of script.", "reporter": "jtatar", "cc": "", "resolution": "fixed", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "CoincSuite split_recombine.py example doesn't run", "priority": "blocker", "keywords": "", "time": "2015-08-18T18:26:13", "milestone": "", "owner": "mzoll", "type": "defect" } ```
defect
coincsuite split recombine py example doesn t run trac migrated from json status closed changetime description here is the error message n n n ninfo python using coincsuite recombinations coincsuite py in complete ntraceback most recent call last n file split recombine py line in n split recombine tray split recombine params n file split recombine py line in split recombine n splitpulses maskedofflinepulses n file home jtatar striketeam icerec build lib py line in addsegment n return segment self name kwargs n file home jtatar striketeam icerec build lib icecube coincsuite coincsuite py line in complete n mininame lilliput add minuit simplex minimizer service tray nnameerror global name lilliput is not defined n n n n default input file does not exist n please add a better summary of what script does at top of script reporter jtatar cc resolution fixed ts component combo reconstruction summary coincsuite split recombine py example doesn t run priority blocker keywords time milestone owner mzoll type defect
1
252,084
27,228,676,758
IssuesEvent
2023-02-21 11:34:24
TIBCOSoftware/apiscout
https://api.github.com/repos/TIBCOSoftware/apiscout
closed
CVE-2019-11250 (Medium) detected in github.com/kubernetes/client-go/transport-v2.0.0 - autoclosed
security vulnerability
## CVE-2019-11250 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/kubernetes/client-go/transport-v2.0.0</b></p></summary> <p>Go client for Kubernetes.</p> <p> Dependency Hierarchy: - github.com/kubernetes/client-go-v2.0.0 (Root Library) - github.com/kubernetes/client-go-v2.0.0 - github.com/kubernetes/client-go-v2.0.0 - :x: **github.com/kubernetes/client-go/transport-v2.0.0** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Kubernetes client-go library logs request headers at verbosity levels of 7 or higher. This can disclose credentials to unauthorized users via logs or command output. Kubernetes components (such as kube-apiserver) prior to v1.16.0, which make use of basic or bearer token authentication, and run at high verbosity levels, are affected. <p>Publish Date: 2019-08-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11250>CVE-2019-11250</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GO-2020-0065">https://osv.dev/vulnerability/GO-2020-0065</a></p> <p>Release Date: 2020-10-16</p> <p>Fix Resolution: v1.16.0-beta.1</p> </p> </details> <p></p>
True
CVE-2019-11250 (Medium) detected in github.com/kubernetes/client-go/transport-v2.0.0 - autoclosed - ## CVE-2019-11250 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/kubernetes/client-go/transport-v2.0.0</b></p></summary> <p>Go client for Kubernetes.</p> <p> Dependency Hierarchy: - github.com/kubernetes/client-go-v2.0.0 (Root Library) - github.com/kubernetes/client-go-v2.0.0 - github.com/kubernetes/client-go-v2.0.0 - :x: **github.com/kubernetes/client-go/transport-v2.0.0** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Kubernetes client-go library logs request headers at verbosity levels of 7 or higher. This can disclose credentials to unauthorized users via logs or command output. Kubernetes components (such as kube-apiserver) prior to v1.16.0, which make use of basic or bearer token authentication, and run at high verbosity levels, are affected. <p>Publish Date: 2019-08-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11250>CVE-2019-11250</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GO-2020-0065">https://osv.dev/vulnerability/GO-2020-0065</a></p> <p>Release Date: 2020-10-16</p> <p>Fix Resolution: v1.16.0-beta.1</p> </p> </details> <p></p>
non_defect
cve medium detected in github com kubernetes client go transport autoclosed cve medium severity vulnerability vulnerable library github com kubernetes client go transport go client for kubernetes dependency hierarchy github com kubernetes client go root library github com kubernetes client go github com kubernetes client go x github com kubernetes client go transport vulnerable library vulnerability details the kubernetes client go library logs request headers at verbosity levels of or higher this can disclose credentials to unauthorized users via logs or command output kubernetes components such as kube apiserver prior to which make use of basic or bearer token authentication and run at high verbosity levels are affected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution beta
0
32,525
15,419,233,412
IssuesEvent
2021-03-05 09:53:10
Piwigo/Piwigo
https://api.github.com/repos/Piwigo/Piwigo
closed
storage usage broken with big galleries
Section: Technical Type: Performance
If your photo gallery contains 1 million photos, the current (11.3) algorithm to calculate the storage space used by each kind of files (photos/videos/formats/other) will reach memory limit. Furthermore, there is no need to calculate the sum in the PHP. The database can do it, more efficiently. On my test database with 960k photos, it takes 800ms to fetch all individuals photos and 4000ms to generate the whole page. With an aggregation query, the SQL takes 1700ms (that's longer) but the page generates in 3000ms (that's faster). Here is the aggregation query I used to test: ``` $query = ' SELECT SUBSTRING_INDEX(path,".",-1) AS ext, SUM(filesize) AS filesize FROM `'.IMAGES_TABLE.'` GROUP BY ext ;'; $file_extensions = query2array($query, 'ext', 'filesize'); ``` Even optimized, it takes too much time in my opinion. Next step would be to cache this.
True
storage usage broken with big galleries - If your photo gallery contains 1 million photos, the current (11.3) algorithm to calculate the storage space used by each kind of files (photos/videos/formats/other) will reach memory limit. Furthermore, there is no need to calculate the sum in the PHP. The database can do it, more efficiently. On my test database with 960k photos, it takes 800ms to fetch all individuals photos and 4000ms to generate the whole page. With an aggregation query, the SQL takes 1700ms (that's longer) but the page generates in 3000ms (that's faster). Here is the aggregation query I used to test: ``` $query = ' SELECT SUBSTRING_INDEX(path,".",-1) AS ext, SUM(filesize) AS filesize FROM `'.IMAGES_TABLE.'` GROUP BY ext ;'; $file_extensions = query2array($query, 'ext', 'filesize'); ``` Even optimized, it takes too much time in my opinion. Next step would be to cache this.
non_defect
storage usage broken with big galleries if your photo gallery contains million photos the current algorithm to calculate the storage space used by each kind of files photos videos formats other will reach memory limit furthermore there is no need to calculate the sum in the php the database can do it more efficiently on my test database with photos it takes to fetch all individuals photos and to generate the whole page with an aggregation query the sql takes that s longer but the page generates in that s faster here is the aggregation query i used to test query select substring index path as ext sum filesize as filesize from images table group by ext file extensions query ext filesize even optimized it takes too much time in my opinion next step would be to cache this
0
267,771
23,318,869,473
IssuesEvent
2022-08-08 14:42:19
gravitational/teleport
https://api.github.com/repos/gravitational/teleport
opened
`TestEC2IsInstanceMetadataAvailable` flakiness
flaky tests
## Failure #### Link(s) to logs - https://console.cloud.google.com/cloud-build/builds/52d61a26-7ba6-4982-a10d-4ca10841ff2a;step=0?project=ci-account #### Relevant snippet ``` === CONT TestEC2IsInstanceMetadataAvailable/response_with_new_id_format ec2_test.go:156: Error Trace: ec2_test.go:156 Error: Should be true Test: TestEC2IsInstanceMetadataAvailable/response_with_new_id_format --- FAIL: TestEC2IsInstanceMetadataAvailable/response_with_new_id_format (0.35s) ```
1.0
`TestEC2IsInstanceMetadataAvailable` flakiness - ## Failure #### Link(s) to logs - https://console.cloud.google.com/cloud-build/builds/52d61a26-7ba6-4982-a10d-4ca10841ff2a;step=0?project=ci-account #### Relevant snippet ``` === CONT TestEC2IsInstanceMetadataAvailable/response_with_new_id_format ec2_test.go:156: Error Trace: ec2_test.go:156 Error: Should be true Test: TestEC2IsInstanceMetadataAvailable/response_with_new_id_format --- FAIL: TestEC2IsInstanceMetadataAvailable/response_with_new_id_format (0.35s) ```
non_defect
flakiness failure link s to logs relevant snippet cont response with new id format test go error trace test go error should be true test response with new id format fail response with new id format
0
186,376
14,394,660,511
IssuesEvent
2020-12-03 01:49:26
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
itsivareddy/terrafrom-Oci: oci/core_virtual_circuit_test.go; 16 LoC
fresh small test
Found a possible issue in [itsivareddy/terrafrom-Oci](https://www.github.com/itsivareddy/terrafrom-Oci) at [oci/core_virtual_circuit_test.go](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_virtual_circuit_test.go#L531-L546) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_virtual_circuit_test.go#L531-L546) <details> <summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary> ```go for _, virtualCircuitId := range virtualCircuitIds { if ok := SweeperDefaultResourceId[virtualCircuitId]; !ok { deleteVirtualCircuitRequest := oci_core.DeleteVirtualCircuitRequest{} deleteVirtualCircuitRequest.VirtualCircuitId = &virtualCircuitId deleteVirtualCircuitRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "core") _, error := virtualNetworkClient.DeleteVirtualCircuit(context.Background(), deleteVirtualCircuitRequest) if error != nil { fmt.Printf("Error deleting VirtualCircuit %s %s, It is possible that the resource is already deleted. Please verify manually \n", virtualCircuitId, error) continue } waitTillCondition(testAccProvider, &virtualCircuitId, virtualCircuitSweepWaitCondition, time.Duration(3*time.Minute), virtualCircuitSweepResponseFetchOperation, "core", true) } } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to virtualCircuitId is reassigned at line 535 </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 075608a9e201ee0e32484da68d5ba5370dfde1be
1.0
itsivareddy/terrafrom-Oci: oci/core_virtual_circuit_test.go; 16 LoC - Found a possible issue in [itsivareddy/terrafrom-Oci](https://www.github.com/itsivareddy/terrafrom-Oci) at [oci/core_virtual_circuit_test.go](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_virtual_circuit_test.go#L531-L546) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_virtual_circuit_test.go#L531-L546) <details> <summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary> ```go for _, virtualCircuitId := range virtualCircuitIds { if ok := SweeperDefaultResourceId[virtualCircuitId]; !ok { deleteVirtualCircuitRequest := oci_core.DeleteVirtualCircuitRequest{} deleteVirtualCircuitRequest.VirtualCircuitId = &virtualCircuitId deleteVirtualCircuitRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "core") _, error := virtualNetworkClient.DeleteVirtualCircuit(context.Background(), deleteVirtualCircuitRequest) if error != nil { fmt.Printf("Error deleting VirtualCircuit %s %s, It is possible that the resource is already deleted. Please verify manually \n", virtualCircuitId, error) continue } waitTillCondition(testAccProvider, &virtualCircuitId, virtualCircuitSweepWaitCondition, time.Duration(3*time.Minute), virtualCircuitSweepResponseFetchOperation, "core", true) } } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to virtualCircuitId is reassigned at line 535 </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 075608a9e201ee0e32484da68d5ba5370dfde1be
non_defect
itsivareddy terrafrom oci oci core virtual circuit test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for virtualcircuitid range virtualcircuitids if ok sweeperdefaultresourceid ok deletevirtualcircuitrequest oci core deletevirtualcircuitrequest deletevirtualcircuitrequest virtualcircuitid virtualcircuitid deletevirtualcircuitrequest requestmetadata retrypolicy getretrypolicy true core error virtualnetworkclient deletevirtualcircuit context background deletevirtualcircuitrequest if error nil fmt printf error deleting virtualcircuit s s it is possible that the resource is already deleted please verify manually n virtualcircuitid error continue waittillcondition testaccprovider virtualcircuitid virtualcircuitsweepwaitcondition time duration time minute virtualcircuitsweepresponsefetchoperation core true below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to virtualcircuitid is reassigned at line leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
75,312
25,763,429,596
IssuesEvent
2022-12-08 22:48:52
vector-im/element-meta
https://api.github.com/repos/vector-im/element-meta
opened
Standardise wording for threads in the labs setting
A-Threads T-Defect O-Occasional S-Tolerable
We're calling them "threads", "threaded messaging" and "thread messages" across the three platforms. We should use "Threaded messages" in all cases
1.0
Standardise wording for threads in the labs setting - We're calling them "threads", "threaded messaging" and "thread messages" across the three platforms. We should use "Threaded messages" in all cases
defect
standardise wording for threads in the labs setting we re calling them threads threaded messaging and thread messages across the three platforms we should use threaded messages in all cases
1
201,065
22,946,722,982
IssuesEvent
2022-07-19 01:14:23
zencoder/castlabs-java-api
https://api.github.com/repos/zencoder/castlabs-java-api
closed
CVE-2020-14062 (High) detected in jackson-databind-2.9.8.jar - autoclosed
security vulnerability
## CVE-2020-14062 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /ory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/zencoder/castlabs-java-api/commit/cedc82ec94663352957fffd68b1d6bb74aba81df">cedc82ec94663352957fffd68b1d6bb74aba81df</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to com.sun.org.apache.xalan.internal.lib.sql.JNDIConnectionPool (aka xalan2). <p>Publish Date: 2020-06-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14062>CVE-2020-14062</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14062">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14062</a></p> <p>Release Date: 2020-06-14</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-14062","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to com.sun.org.apache.xalan.internal.lib.sql.JNDIConnectionPool (aka xalan2).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14062","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-14062 (High) detected in jackson-databind-2.9.8.jar - autoclosed - ## CVE-2020-14062 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /ory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/zencoder/castlabs-java-api/commit/cedc82ec94663352957fffd68b1d6bb74aba81df">cedc82ec94663352957fffd68b1d6bb74aba81df</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to com.sun.org.apache.xalan.internal.lib.sql.JNDIConnectionPool (aka xalan2). <p>Publish Date: 2020-06-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14062>CVE-2020-14062</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14062">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14062</a></p> <p>Release Date: 2020-06-14</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-14062","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to com.sun.org.apache.xalan.internal.lib.sql.JNDIConnectionPool (aka xalan2).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14062","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library ory com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com sun org apache xalan internal lib sql jndiconnectionpool aka publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com sun org apache xalan internal lib sql jndiconnectionpool aka vulnerabilityurl
0
275,408
23,913,596,773
IssuesEvent
2022-09-09 10:30:29
wpeventmanager/wp-event-manager
https://api.github.com/repos/wpeventmanager/wp-event-manager
closed
wp-event-manager / update-banner does not work !
In Testing
Hello, #738 has been close but bug is always here I tested that again https://demo.wp-eventmanager.com/t3a6zsurong9qhc/events/#events admin : I create an event then I update banner using dashboard Banner is not updated in events page But in the dashboard thumbnail is ok Did you saw my video ? If it works on your side could you post a video showing what you are doing ? Thanks Regards
1.0
wp-event-manager / update-banner does not work ! - Hello, #738 has been close but bug is always here I tested that again https://demo.wp-eventmanager.com/t3a6zsurong9qhc/events/#events admin : I create an event then I update banner using dashboard Banner is not updated in events page But in the dashboard thumbnail is ok Did you saw my video ? If it works on your side could you post a video showing what you are doing ? Thanks Regards
non_defect
wp event manager update banner does not work hello has been close but bug is always here i tested that again admin i create an event then i update banner using dashboard banner is not updated in events page but in the dashboard thumbnail is ok did you saw my video if it works on your side could you post a video showing what you are doing thanks regards
0
63,028
6,823,221,385
IssuesEvent
2017-11-07 22:55:51
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Need an easy way to find leaking animations in running app (not in test)
dev: tests performance prod: framework
It's easy to end up leaking animations but not easy to catch them. https://github.com/flutter/flutter/issues/4918#issuecomment-232770246 We have debugAssertNoTransientCallbacks which we use in flutter_test to catch animation leaks, but if you don't happen to have a flutter_test of your widget there isn't a way to catch leaked animations. @jason-simmons and I just tried collecting to an app with a known leak via Observatory and calling debugAssertNoTransientCallbacks while the isolate was paused and it didn't seem to print anything. Regardless, we should make this less easy to get wrong or more easy to fix if you do.
1.0
Need an easy way to find leaking animations in running app (not in test) - It's easy to end up leaking animations but not easy to catch them. https://github.com/flutter/flutter/issues/4918#issuecomment-232770246 We have debugAssertNoTransientCallbacks which we use in flutter_test to catch animation leaks, but if you don't happen to have a flutter_test of your widget there isn't a way to catch leaked animations. @jason-simmons and I just tried collecting to an app with a known leak via Observatory and calling debugAssertNoTransientCallbacks while the isolate was paused and it didn't seem to print anything. Regardless, we should make this less easy to get wrong or more easy to fix if you do.
non_defect
need an easy way to find leaking animations in running app not in test it s easy to end up leaking animations but not easy to catch them we have debugassertnotransientcallbacks which we use in flutter test to catch animation leaks but if you don t happen to have a flutter test of your widget there isn t a way to catch leaked animations jason simmons and i just tried collecting to an app with a known leak via observatory and calling debugassertnotransientcallbacks while the isolate was paused and it didn t seem to print anything regardless we should make this less easy to get wrong or more easy to fix if you do
0
69,469
8,408,950,701
IssuesEvent
2018-10-12 04:48:19
VotingVoices/chorus
https://api.github.com/repos/VotingVoices/chorus
closed
Should we tone down the all-caps headers on the Plan page?
design feedback
Personally I think the all-caps headers come across as shouting, which might contribute to the feedback we've heard (@seattlehedgehog can confirm?) that people find the Plan page a little bit intimidating. Maybe we can tone down the headers by not using all-caps?
1.0
Should we tone down the all-caps headers on the Plan page? - Personally I think the all-caps headers come across as shouting, which might contribute to the feedback we've heard (@seattlehedgehog can confirm?) that people find the Plan page a little bit intimidating. Maybe we can tone down the headers by not using all-caps?
non_defect
should we tone down the all caps headers on the plan page personally i think the all caps headers come across as shouting which might contribute to the feedback we ve heard seattlehedgehog can confirm that people find the plan page a little bit intimidating maybe we can tone down the headers by not using all caps
0
56,722
15,326,099,148
IssuesEvent
2021-02-26 02:49:39
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Forms flagged as deleted are not being archived by the nightly forms migration.
Defect Product Support Team VA Forms
**Describe the defect** When a form is flagged in the nightly forms DB data dump here https://prod.cms.va.gov/sites/default/files/migrate_source/va_forms_data.csv the form detail page is supposed to be archived by the migration. The form `21-0790` is flagged as deleted but the form detail page was never archived. It does not even show as being updated on the date it was deleted. https://prod.cms.va.gov/find-forms/about-form-21-0790 **Facts:** 1. The flag was never updated in the CMS data. 2. The deleted date field value does not have the date from the CSV. 3. It was was not updated by the migration since the deletion date in the CSV. This needs some investigation. https://dsva.slack.com/archives/CBU0KDSB1/p1611590647138900 **Rule Out** - [x] Migration not running nightly: Last run time and Jenkins confirms it is. - [x] Editorial mishap : No record of any mistake in the revision history. - [x] Revision reversion: No record of it. - [x] Bad archive logic: Tested it by hand. Worked as expected, and worked as expected with 12 of the 13 nodes that should be archived. Also no record the node was updated to reflect the deleted flag or the deleted date, which happens in the migration before the archive logic happens. - [x] Toggle of deleted flag in the CSV: This caused no errors and worked as expected. - [x] What if they put in a deleted flag other than 1 (like a string or some other number)? It still archived as expected. - [x] What if there was bad data on the row and then fixed it later? It fixed it when the correct format was restored to the current state of the row. - [x] What if they toggled the deleted on and off a few times? It worked as expected. - [x] What if the archive logic was flawed? I checked it line by line in x-debug AND that would not explain why the deleted date was never updated. - [x] Row id (the migration key) changed in the CSV and then changed back. This would have resulted in duplicate nodes. There are no duplicates for this form. - [x] destination id (Node id) and sourceid (rowid from CSV) mismatch? No that looks fine. - [x] Duplicate destination ids? No. - [x] Duplicate sourceid in map table? No - [x] Duplicate rowid in CSV? No - [ ] The form deleted date is 2020-10-30 So was there anything on that night or the couple nights after that would have been a failure of the site? - [ ] Node save happens first and then the migrate_map is updated. What if the node was saved on one instance and the migration map was updated on the new instance? (catastrophic failure of some sort?) This would cause the migration to think it had made the update and no other changes to the data row would happen. The migration map would have an updated date that does not correspond to a node save revision.
1.0
Forms flagged as deleted are not being archived by the nightly forms migration. - **Describe the defect** When a form is flagged in the nightly forms DB data dump here https://prod.cms.va.gov/sites/default/files/migrate_source/va_forms_data.csv the form detail page is supposed to be archived by the migration. The form `21-0790` is flagged as deleted but the form detail page was never archived. It does not even show as being updated on the date it was deleted. https://prod.cms.va.gov/find-forms/about-form-21-0790 **Facts:** 1. The flag was never updated in the CMS data. 2. The deleted date field value does not have the date from the CSV. 3. It was was not updated by the migration since the deletion date in the CSV. This needs some investigation. https://dsva.slack.com/archives/CBU0KDSB1/p1611590647138900 **Rule Out** - [x] Migration not running nightly: Last run time and Jenkins confirms it is. - [x] Editorial mishap : No record of any mistake in the revision history. - [x] Revision reversion: No record of it. - [x] Bad archive logic: Tested it by hand. Worked as expected, and worked as expected with 12 of the 13 nodes that should be archived. Also no record the node was updated to reflect the deleted flag or the deleted date, which happens in the migration before the archive logic happens. - [x] Toggle of deleted flag in the CSV: This caused no errors and worked as expected. - [x] What if they put in a deleted flag other than 1 (like a string or some other number)? It still archived as expected. - [x] What if there was bad data on the row and then fixed it later? It fixed it when the correct format was restored to the current state of the row. - [x] What if they toggled the deleted on and off a few times? It worked as expected. - [x] What if the archive logic was flawed? I checked it line by line in x-debug AND that would not explain why the deleted date was never updated. - [x] Row id (the migration key) changed in the CSV and then changed back. This would have resulted in duplicate nodes. There are no duplicates for this form. - [x] destination id (Node id) and sourceid (rowid from CSV) mismatch? No that looks fine. - [x] Duplicate destination ids? No. - [x] Duplicate sourceid in map table? No - [x] Duplicate rowid in CSV? No - [ ] The form deleted date is 2020-10-30 So was there anything on that night or the couple nights after that would have been a failure of the site? - [ ] Node save happens first and then the migrate_map is updated. What if the node was saved on one instance and the migration map was updated on the new instance? (catastrophic failure of some sort?) This would cause the migration to think it had made the update and no other changes to the data row would happen. The migration map would have an updated date that does not correspond to a node save revision.
defect
forms flagged as deleted are not being archived by the nightly forms migration describe the defect when a form is flagged in the nightly forms db data dump here the form detail page is supposed to be archived by the migration the form is flagged as deleted but the form detail page was never archived it does not even show as being updated on the date it was deleted facts the flag was never updated in the cms data the deleted date field value does not have the date from the csv it was was not updated by the migration since the deletion date in the csv this needs some investigation rule out migration not running nightly last run time and jenkins confirms it is editorial mishap no record of any mistake in the revision history revision reversion no record of it bad archive logic tested it by hand worked as expected and worked as expected with of the nodes that should be archived also no record the node was updated to reflect the deleted flag or the deleted date which happens in the migration before the archive logic happens toggle of deleted flag in the csv this caused no errors and worked as expected what if they put in a deleted flag other than like a string or some other number it still archived as expected what if there was bad data on the row and then fixed it later it fixed it when the correct format was restored to the current state of the row what if they toggled the deleted on and off a few times it worked as expected what if the archive logic was flawed i checked it line by line in x debug and that would not explain why the deleted date was never updated row id the migration key changed in the csv and then changed back this would have resulted in duplicate nodes there are no duplicates for this form destination id node id and sourceid rowid from csv mismatch no that looks fine duplicate destination ids no duplicate sourceid in map table no duplicate rowid in csv no the form deleted date is so was there anything on that night or the couple nights after that would have been a failure of the site node save happens first and then the migrate map is updated what if the node was saved on one instance and the migration map was updated on the new instance catastrophic failure of some sort this would cause the migration to think it had made the update and no other changes to the data row would happen the migration map would have an updated date that does not correspond to a node save revision
1
34,838
12,303,003,713
IssuesEvent
2020-05-11 17:56:09
nextgenhealthcare/connect
https://api.github.com/repos/nextgenhealthcare/connect
closed
Add HTTPS Listener and HTTPS Sender
Security
Add an HTTPS Listener and HTTPS Sender. This was referenced in MIRTH-74 but the HTTPS part of that request was not added. Imported Issue. Original Details: Jira Issue Key: MIRTH-645 Reporter: jlehew Created: 2007-11-26T09:16:32.000-0800
True
Add HTTPS Listener and HTTPS Sender - Add an HTTPS Listener and HTTPS Sender. This was referenced in MIRTH-74 but the HTTPS part of that request was not added. Imported Issue. Original Details: Jira Issue Key: MIRTH-645 Reporter: jlehew Created: 2007-11-26T09:16:32.000-0800
non_defect
add https listener and https sender add an https listener and https sender this was referenced in mirth but the https part of that request was not added imported issue original details jira issue key mirth reporter jlehew created
0
15,528
2,859,440,490
IssuesEvent
2015-06-03 10:35:38
netty/netty
https://api.github.com/repos/netty/netty
closed
DNS resolver loses ADDITIONAL records
defect
#### Netty Version 4.1.0.Beta5 (using latest code/API per https://github.com/netty/netty/commit/63a02fc04e6ce45aee38b0437ae0279880e5fe40 @trustin ) #### Context Additional records section is important for cutting down the number of necessary DNS questions - mainly to avoid follow-up queries for resolving hostnames (MX, CNAME or NS) to A / AAAA records. Unfortunately we've discovered that these records are **never** included when querying by Netty and we've tried to do MX queries for ~1 million different domains. #### Step to Reproduce First, try this on your shell prompt - ``` [root@localhost]# dig MX hotmail.com ; <<>> DiG 9.8.2rc1-RedHat-9.8.2-0.30.rc1.el6 <<>> MX hotmail.com ;; global options: +cmd ;; Got answer: ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 38795 ;; flags: qr rd ra; QUERY: 1, ANSWER: 4, AUTHORITY: 4, ADDITIONAL: 16 ;; QUESTION SECTION: ;hotmail.com. IN MX ;; ANSWER SECTION: hotmail.com. 3600 IN MX 5 mx4.hotmail.com. hotmail.com. 3600 IN MX 5 mx1.hotmail.com. hotmail.com. 3600 IN MX 5 mx2.hotmail.com. hotmail.com. 3600 IN MX 5 mx3.hotmail.com. ;; AUTHORITY SECTION: hotmail.com. 42729 IN NS ns1.msft.net. hotmail.com. 42729 IN NS ns2.msft.net. hotmail.com. 42729 IN NS ns3.msft.net. hotmail.com. 42729 IN NS ns4.msft.net. ;; ADDITIONAL SECTION: mx1.hotmail.com. 3600 IN A 65.54.188.126 mx1.hotmail.com. 3600 IN A 65.55.33.119 mx1.hotmail.com. 3600 IN A 65.55.33.135 mx1.hotmail.com. 3600 IN A 65.55.37.72 mx1.hotmail.com. 3600 IN A 65.55.37.88 mx1.hotmail.com. 3600 IN A 65.55.37.104 mx1.hotmail.com. 3600 IN A 65.55.37.120 mx1.hotmail.com. 3600 IN A 65.55.92.136 mx1.hotmail.com. 3600 IN A 65.55.92.152 mx1.hotmail.com. 3600 IN A 65.55.92.168 mx1.hotmail.com. 3600 IN A 65.55.92.184 mx1.hotmail.com. 3600 IN A 207.46.8.167 mx1.hotmail.com. 3600 IN A 207.46.8.199 mx1.hotmail.com. 3600 IN A 65.54.188.72 mx1.hotmail.com. 3600 IN A 65.54.188.94 mx1.hotmail.com. 3600 IN A 65.54.188.110 ;; Query time: 14 msec ;; SERVER: 72.13.91.2#53(72.13.91.2) ;; WHEN: Sun May 17 22:08:02 2015 ;; MSG SIZE rcvd: 445 ``` Then do the same query via Netty and observe no A records in the additional section of the response ``` resolver.query(new DefaultDnsQuestion("hotmail.com", DnsRecordType.MX)).addListener(this); ``` #### Possible Causes? First of all, my knowledge of DNS protocol intricacies is limited, so I am just reporting a few things that seem different from the above ```dig``` test. I see that resolver is including some sort of ```OPT``` record in every DNS query ( https://github.com/netty/netty/blob/netty-4.1.0.Beta5/resolver-dns/src/main/java/io/netty/resolver/dns/DnsQueryContext.java#L74 ) and the DNS responses sometimes include an ```OPT``` record in the additional section. Seems to be this might be interfering with normal retrieval of additional records? In any case perhaps it would be a good idea to allow a finer control over the additional section of the question vs. have the resolver always do exactly the same unchangeable thing... #### Environment [root@localhost]# java -version java version "1.7.0_51" Java(TM) SE Runtime Environment (build 1.7.0_51-b13) Java HotSpot(TM) 64-Bit Server VM (build 24.51-b03, mixed mode) [root@localhost]# uname -a Linux localhost 2.6.32-431.el6.x86_64 #1 SMP Fri Nov 22 03:15:09 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
1.0
DNS resolver loses ADDITIONAL records - #### Netty Version 4.1.0.Beta5 (using latest code/API per https://github.com/netty/netty/commit/63a02fc04e6ce45aee38b0437ae0279880e5fe40 @trustin ) #### Context Additional records section is important for cutting down the number of necessary DNS questions - mainly to avoid follow-up queries for resolving hostnames (MX, CNAME or NS) to A / AAAA records. Unfortunately we've discovered that these records are **never** included when querying by Netty and we've tried to do MX queries for ~1 million different domains. #### Step to Reproduce First, try this on your shell prompt - ``` [root@localhost]# dig MX hotmail.com ; <<>> DiG 9.8.2rc1-RedHat-9.8.2-0.30.rc1.el6 <<>> MX hotmail.com ;; global options: +cmd ;; Got answer: ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 38795 ;; flags: qr rd ra; QUERY: 1, ANSWER: 4, AUTHORITY: 4, ADDITIONAL: 16 ;; QUESTION SECTION: ;hotmail.com. IN MX ;; ANSWER SECTION: hotmail.com. 3600 IN MX 5 mx4.hotmail.com. hotmail.com. 3600 IN MX 5 mx1.hotmail.com. hotmail.com. 3600 IN MX 5 mx2.hotmail.com. hotmail.com. 3600 IN MX 5 mx3.hotmail.com. ;; AUTHORITY SECTION: hotmail.com. 42729 IN NS ns1.msft.net. hotmail.com. 42729 IN NS ns2.msft.net. hotmail.com. 42729 IN NS ns3.msft.net. hotmail.com. 42729 IN NS ns4.msft.net. ;; ADDITIONAL SECTION: mx1.hotmail.com. 3600 IN A 65.54.188.126 mx1.hotmail.com. 3600 IN A 65.55.33.119 mx1.hotmail.com. 3600 IN A 65.55.33.135 mx1.hotmail.com. 3600 IN A 65.55.37.72 mx1.hotmail.com. 3600 IN A 65.55.37.88 mx1.hotmail.com. 3600 IN A 65.55.37.104 mx1.hotmail.com. 3600 IN A 65.55.37.120 mx1.hotmail.com. 3600 IN A 65.55.92.136 mx1.hotmail.com. 3600 IN A 65.55.92.152 mx1.hotmail.com. 3600 IN A 65.55.92.168 mx1.hotmail.com. 3600 IN A 65.55.92.184 mx1.hotmail.com. 3600 IN A 207.46.8.167 mx1.hotmail.com. 3600 IN A 207.46.8.199 mx1.hotmail.com. 3600 IN A 65.54.188.72 mx1.hotmail.com. 3600 IN A 65.54.188.94 mx1.hotmail.com. 3600 IN A 65.54.188.110 ;; Query time: 14 msec ;; SERVER: 72.13.91.2#53(72.13.91.2) ;; WHEN: Sun May 17 22:08:02 2015 ;; MSG SIZE rcvd: 445 ``` Then do the same query via Netty and observe no A records in the additional section of the response ``` resolver.query(new DefaultDnsQuestion("hotmail.com", DnsRecordType.MX)).addListener(this); ``` #### Possible Causes? First of all, my knowledge of DNS protocol intricacies is limited, so I am just reporting a few things that seem different from the above ```dig``` test. I see that resolver is including some sort of ```OPT``` record in every DNS query ( https://github.com/netty/netty/blob/netty-4.1.0.Beta5/resolver-dns/src/main/java/io/netty/resolver/dns/DnsQueryContext.java#L74 ) and the DNS responses sometimes include an ```OPT``` record in the additional section. Seems to be this might be interfering with normal retrieval of additional records? In any case perhaps it would be a good idea to allow a finer control over the additional section of the question vs. have the resolver always do exactly the same unchangeable thing... #### Environment [root@localhost]# java -version java version "1.7.0_51" Java(TM) SE Runtime Environment (build 1.7.0_51-b13) Java HotSpot(TM) 64-Bit Server VM (build 24.51-b03, mixed mode) [root@localhost]# uname -a Linux localhost 2.6.32-431.el6.x86_64 #1 SMP Fri Nov 22 03:15:09 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
defect
dns resolver loses additional records netty version using latest code api per trustin context additional records section is important for cutting down the number of necessary dns questions mainly to avoid follow up queries for resolving hostnames mx cname or ns to a aaaa records unfortunately we ve discovered that these records are never included when querying by netty and we ve tried to do mx queries for million different domains step to reproduce first try this on your shell prompt dig mx hotmail com dig redhat mx hotmail com global options cmd got answer header opcode query status noerror id flags qr rd ra query answer authority additional question section hotmail com in mx answer section hotmail com in mx hotmail com hotmail com in mx hotmail com hotmail com in mx hotmail com hotmail com in mx hotmail com authority section hotmail com in ns msft net hotmail com in ns msft net hotmail com in ns msft net hotmail com in ns msft net additional section hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a query time msec server when sun may msg size rcvd then do the same query via netty and observe no a records in the additional section of the response resolver query new defaultdnsquestion hotmail com dnsrecordtype mx addlistener this possible causes first of all my knowledge of dns protocol intricacies is limited so i am just reporting a few things that seem different from the above dig test i see that resolver is including some sort of opt record in every dns query and the dns responses sometimes include an opt record in the additional section seems to be this might be interfering with normal retrieval of additional records in any case perhaps it would be a good idea to allow a finer control over the additional section of the question vs have the resolver always do exactly the same unchangeable thing environment java version java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode uname a linux localhost smp fri nov utc gnu linux
1
35,613
7,794,610,089
IssuesEvent
2018-06-08 03:45:08
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
RICO changed on May 29 Plots (Trac #48)
Migrated from Trac clubb_src defect dschanen@uwm.edu
The RICO case (but no other cases), changed substantially on the nightly plots for 29 May 2009. Did the initial sounding change? Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/48 ```json { "status": "closed", "changetime": "2009-09-29T20:02:55", "description": "The RICO case (but no other cases), changed substantially on the nightly plots for 29 May 2009. Did the initial sounding change?", "reporter": "dschanen@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1254254575000000", "component": "clubb_src", "summary": "RICO changed on May 29 Plots", "priority": "major", "keywords": "RICO, plotgen, night plots", "time": "2009-05-29T17:53:24", "milestone": "", "owner": "dschanen@uwm.edu", "type": "defect" } ```
1.0
RICO changed on May 29 Plots (Trac #48) - The RICO case (but no other cases), changed substantially on the nightly plots for 29 May 2009. Did the initial sounding change? Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/48 ```json { "status": "closed", "changetime": "2009-09-29T20:02:55", "description": "The RICO case (but no other cases), changed substantially on the nightly plots for 29 May 2009. Did the initial sounding change?", "reporter": "dschanen@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1254254575000000", "component": "clubb_src", "summary": "RICO changed on May 29 Plots", "priority": "major", "keywords": "RICO, plotgen, night plots", "time": "2009-05-29T17:53:24", "milestone": "", "owner": "dschanen@uwm.edu", "type": "defect" } ```
defect
rico changed on may plots trac the rico case but no other cases changed substantially on the nightly plots for may did the initial sounding change attachments migrated from json status closed changetime description the rico case but no other cases changed substantially on the nightly plots for may did the initial sounding change reporter dschanen uwm edu cc resolution verified by v larson ts component clubb src summary rico changed on may plots priority major keywords rico plotgen night plots time milestone owner dschanen uwm edu type defect
1
233,891
17,911,286,386
IssuesEvent
2021-09-09 05:50:20
OneTwentyFour/unicorndocs
https://api.github.com/repos/OneTwentyFour/unicorndocs
closed
docs reorganization
documentation
We're working on reorganizing the docs as we began doing in PR #2 ... individual branches will be created and merged eventually to main as we continue working on the documentation. This issue will track our progress as we go.
1.0
docs reorganization - We're working on reorganizing the docs as we began doing in PR #2 ... individual branches will be created and merged eventually to main as we continue working on the documentation. This issue will track our progress as we go.
non_defect
docs reorganization we re working on reorganizing the docs as we began doing in pr individual branches will be created and merged eventually to main as we continue working on the documentation this issue will track our progress as we go
0
447,316
31,682,573,137
IssuesEvent
2023-09-08 01:59:55
DeMarcoLab/fibsem
https://api.github.com/repos/DeMarcoLab/fibsem
closed
[DOC] User Guide
documentation
Create a user guide including: - [x] System Configuration - [x] Connecting / Setup - [x] Imaging + Settings - [x] Movement (Modes, Hotkeys) - [x] Milling (Patterns, Settings, Hotkeys) - [x] Manipulator (Positions, Modes) - [x] Detection (Supervision, Features, Segmentation, Data)
1.0
[DOC] User Guide - Create a user guide including: - [x] System Configuration - [x] Connecting / Setup - [x] Imaging + Settings - [x] Movement (Modes, Hotkeys) - [x] Milling (Patterns, Settings, Hotkeys) - [x] Manipulator (Positions, Modes) - [x] Detection (Supervision, Features, Segmentation, Data)
non_defect
user guide create a user guide including system configuration connecting setup imaging settings movement modes hotkeys milling patterns settings hotkeys manipulator positions modes detection supervision features segmentation data
0
139,373
12,855,311,096
IssuesEvent
2020-07-09 04:49:36
deathlyrage/pot-demo-bugs
https://api.github.com/repos/deathlyrage/pot-demo-bugs
closed
Skin will not load
documentation more info
**Location:** (X=-54359.347656,Y=380192.40625,Z=-2523.67749) **Message:** My skin will not load unless I scroll in up to the dinosaurs scales ![](https://mapbug.alderongames.com/uploads/(X=-54359.347656,Y=380192.40625,Z=-2523.67749).png) **Version:** 8222 (demo-public-test) **Reporter:** Vigil (658-017-677)
1.0
Skin will not load - **Location:** (X=-54359.347656,Y=380192.40625,Z=-2523.67749) **Message:** My skin will not load unless I scroll in up to the dinosaurs scales ![](https://mapbug.alderongames.com/uploads/(X=-54359.347656,Y=380192.40625,Z=-2523.67749).png) **Version:** 8222 (demo-public-test) **Reporter:** Vigil (658-017-677)
non_defect
skin will not load location x y z message my skin will not load unless i scroll in up to the dinosaurs scales version demo public test reporter vigil
0
50,725
13,187,698,487
IssuesEvent
2020-08-13 04:16:33
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
Example scripts in weighting do not run (Trac #1238)
Migrated from Trac combo core defect
neither example scripts work out of the box nugen_weights.py complains about NewNuFlux corsika_weights.py complains about no input files and overwrites itslef with an hdf5 file <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1238">https://code.icecube.wisc.edu/ticket/1238</a>, reported by kjmeagher and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:13:59", "description": "neither example scripts work out of the box\nnugen_weights.py complains about NewNuFlux\ncorsika_weights.py complains about no input files and overwrites itslef with an hdf5 file", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1458335639558230", "component": "combo core", "summary": "Example scripts in weighting do not run", "priority": "blocker", "keywords": "", "time": "2015-08-20T08:28:57", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
Example scripts in weighting do not run (Trac #1238) - neither example scripts work out of the box nugen_weights.py complains about NewNuFlux corsika_weights.py complains about no input files and overwrites itslef with an hdf5 file <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1238">https://code.icecube.wisc.edu/ticket/1238</a>, reported by kjmeagher and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:13:59", "description": "neither example scripts work out of the box\nnugen_weights.py complains about NewNuFlux\ncorsika_weights.py complains about no input files and overwrites itslef with an hdf5 file", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1458335639558230", "component": "combo core", "summary": "Example scripts in weighting do not run", "priority": "blocker", "keywords": "", "time": "2015-08-20T08:28:57", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
example scripts in weighting do not run trac neither example scripts work out of the box nugen weights py complains about newnuflux corsika weights py complains about no input files and overwrites itslef with an file migrated from json status closed changetime description neither example scripts work out of the box nnugen weights py complains about newnuflux ncorsika weights py complains about no input files and overwrites itslef with an file reporter kjmeagher cc resolution fixed ts component combo core summary example scripts in weighting do not run priority blocker keywords time milestone owner jvansanten type defect
1
74,232
25,016,634,527
IssuesEvent
2022-11-03 19:26:16
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Investigation into Content release failures due to runner permission issues
Defect Needs refining Unplanned work Platform CMS Team
## Background There were a [number of Content release failures on 10-27-2022](https://github.com/department-of-veterans-affairs/content-build/actions/workflows/content-release.yml?query=created%3A2022-10-27..2022-10-27). The failures were around lack of disk-write permissions. Example: ``` https://github.com/department-of-veterans-affairs/content-build Removing previously created refs, to avoid conflicts /usr/bin/git rev-parse --symbolic-full-name --verify --quiet HEAD refs/heads/main /usr/bin/git checkout --detach Error: fatal: Unable to create '/home/runner/runner2/_work/content-build/content-build/.git/index.lock': Permission denied Warning: Unable to prepare the existing repository. The repository will be recreated instead. Deleting the contents of '/home/runner/runner2/_work/content-build/content-build' Error: stderr maxBuffer length exceeded ``` These types of failures have happened previously and generally resolve on their own, but on 10-27 the failures lasted longer than previously (nearly 60 minutes) and were more visible to the Sitewide CMS team. We decided to spend some time investigating the issue. A discussion of the issue took place in #platform-cop-devops (closed channel): https://dsva.slack.com/archives/C040CF55H0A/p1666928783525639 ## Findings Any [workflow that uses a self-hosted runner on the Content Build repo](https://github.com/department-of-veterans-affairs/content-build/actions) uses runners with the name [dsva-vagov-content-build-gha-runner](https://console.amazonaws-us-gov.com/ec2/v2/home?region=us-gov-west-1#Instances:instanceState=running;tag:Name=dsva-vagov-content-build-gha-runner;v=3;$case=tags:true%5C,client:false;$regex=tags:false%5C,client:false;sort=desc:launchTime). When these workflows run, they check out and cache code to defined locations on the runner that they're running on. These locations, their permissions, and any changes to those permissions are persistent for the life of the runner. Normally, a user `runner` is designated as the user which runs the workflow job. A select number of these workflows run parts of their workflow within containers. These containers are stood up using the root user. Examples: * [a11y workflow](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/a11y.yml#L160) * [a11y heading order](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/a11y-heading-order.yml#L126) * [continuous integration](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/continuous-integration.yml#L356) When these jobs using containers run as root, they change ownership of the directories for code checkouts & yarn/npm installation to root. This means that, when a subsequent workflow which is not containerized attempts to run, it attempts to install code to a root-owned directory. This fails because the runner user does not have permission to write to those directories. ## Recommendations If we could isolate installation location for everything - node & node_modules and also repository checkouts - then we could allow containerized 'root' jobs to coexist on the same runners as non-containerized jobs. However, [action/setup-node](https://github.com/actions/setup-node) does not provide any mechanism for setting install location. Therefore, the recommendation is: * Ensure either that all jobs treat permissions in the same way, or else ensure that there are instances that can be used only for containerized and only for non-containerized. ## Follow-up tickets * [doing the above work to create the new ASG](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11402) * [doing the work to assign the appropriate ASG to the appropriate jobs](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11403) * [culling any outstanding instances (there are currently something like 96 running instances which is very expensive)](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11405) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [x] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
Investigation into Content release failures due to runner permission issues - ## Background There were a [number of Content release failures on 10-27-2022](https://github.com/department-of-veterans-affairs/content-build/actions/workflows/content-release.yml?query=created%3A2022-10-27..2022-10-27). The failures were around lack of disk-write permissions. Example: ``` https://github.com/department-of-veterans-affairs/content-build Removing previously created refs, to avoid conflicts /usr/bin/git rev-parse --symbolic-full-name --verify --quiet HEAD refs/heads/main /usr/bin/git checkout --detach Error: fatal: Unable to create '/home/runner/runner2/_work/content-build/content-build/.git/index.lock': Permission denied Warning: Unable to prepare the existing repository. The repository will be recreated instead. Deleting the contents of '/home/runner/runner2/_work/content-build/content-build' Error: stderr maxBuffer length exceeded ``` These types of failures have happened previously and generally resolve on their own, but on 10-27 the failures lasted longer than previously (nearly 60 minutes) and were more visible to the Sitewide CMS team. We decided to spend some time investigating the issue. A discussion of the issue took place in #platform-cop-devops (closed channel): https://dsva.slack.com/archives/C040CF55H0A/p1666928783525639 ## Findings Any [workflow that uses a self-hosted runner on the Content Build repo](https://github.com/department-of-veterans-affairs/content-build/actions) uses runners with the name [dsva-vagov-content-build-gha-runner](https://console.amazonaws-us-gov.com/ec2/v2/home?region=us-gov-west-1#Instances:instanceState=running;tag:Name=dsva-vagov-content-build-gha-runner;v=3;$case=tags:true%5C,client:false;$regex=tags:false%5C,client:false;sort=desc:launchTime). When these workflows run, they check out and cache code to defined locations on the runner that they're running on. These locations, their permissions, and any changes to those permissions are persistent for the life of the runner. Normally, a user `runner` is designated as the user which runs the workflow job. A select number of these workflows run parts of their workflow within containers. These containers are stood up using the root user. Examples: * [a11y workflow](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/a11y.yml#L160) * [a11y heading order](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/a11y-heading-order.yml#L126) * [continuous integration](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/continuous-integration.yml#L356) When these jobs using containers run as root, they change ownership of the directories for code checkouts & yarn/npm installation to root. This means that, when a subsequent workflow which is not containerized attempts to run, it attempts to install code to a root-owned directory. This fails because the runner user does not have permission to write to those directories. ## Recommendations If we could isolate installation location for everything - node & node_modules and also repository checkouts - then we could allow containerized 'root' jobs to coexist on the same runners as non-containerized jobs. However, [action/setup-node](https://github.com/actions/setup-node) does not provide any mechanism for setting install location. Therefore, the recommendation is: * Ensure either that all jobs treat permissions in the same way, or else ensure that there are instances that can be used only for containerized and only for non-containerized. ## Follow-up tickets * [doing the above work to create the new ASG](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11402) * [doing the work to assign the appropriate ASG to the appropriate jobs](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11403) * [culling any outstanding instances (there are currently something like 96 running instances which is very expensive)](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11405) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [x] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
defect
investigation into content release failures due to runner permission issues background there were a the failures were around lack of disk write permissions example removing previously created refs to avoid conflicts usr bin git rev parse symbolic full name verify quiet head refs heads main usr bin git checkout detach error fatal unable to create home runner work content build content build git index lock permission denied warning unable to prepare the existing repository the repository will be recreated instead deleting the contents of home runner work content build content build error stderr maxbuffer length exceeded these types of failures have happened previously and generally resolve on their own but on the failures lasted longer than previously nearly minutes and were more visible to the sitewide cms team we decided to spend some time investigating the issue a discussion of the issue took place in platform cop devops closed channel findings any uses runners with the name when these workflows run they check out and cache code to defined locations on the runner that they re running on these locations their permissions and any changes to those permissions are persistent for the life of the runner normally a user runner is designated as the user which runs the workflow job a select number of these workflows run parts of their workflow within containers these containers are stood up using the root user examples when these jobs using containers run as root they change ownership of the directories for code checkouts yarn npm installation to root this means that when a subsequent workflow which is not containerized attempts to run it attempts to install code to a root owned directory this fails because the runner user does not have permission to write to those directories recommendations if we could isolate installation location for everything node node modules and also repository checkouts then we could allow containerized root jobs to coexist on the same runners as non containerized jobs however does not provide any mechanism for setting install location therefore the recommendation is ensure either that all jobs treat permissions in the same way or else ensure that there are instances that can be used only for containerized and only for non containerized follow up tickets cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
1
76,265
26,337,173,809
IssuesEvent
2023-01-10 15:07:52
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Creating polls does not scale well for small windows
T-Defect S-Minor A11y O-Occasional A-Polls
### Steps to reproduce ![image](https://user-images.githubusercontent.com/2403652/164203803-5f91ebba-5cc8-499e-879d-8d482e0054c0.png) ### Outcome #### What did you expect? A responsive modal #### What happened instead? Overlapping buttons and horizontal overscroll ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version 1.10.10 ### Homeserver _No response_ ### Will you send logs? No
1.0
Creating polls does not scale well for small windows - ### Steps to reproduce ![image](https://user-images.githubusercontent.com/2403652/164203803-5f91ebba-5cc8-499e-879d-8d482e0054c0.png) ### Outcome #### What did you expect? A responsive modal #### What happened instead? Overlapping buttons and horizontal overscroll ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version 1.10.10 ### Homeserver _No response_ ### Will you send logs? No
defect
creating polls does not scale well for small windows steps to reproduce outcome what did you expect a responsive modal what happened instead overlapping buttons and horizontal overscroll operating system no response browser information no response url for webapp no response application version homeserver no response will you send logs no
1
69,876
9,344,647,945
IssuesEvent
2019-03-30 00:00:59
rancher/k3s
https://api.github.com/repos/rancher/k3s
closed
Document image preloading
help wanted kind/documentation
You can now put docker image export tars at `/var/lib/rancher/k3s/agent/images` and they will be read and loaded into containerd on start. We need to document this :)
1.0
Document image preloading - You can now put docker image export tars at `/var/lib/rancher/k3s/agent/images` and they will be read and loaded into containerd on start. We need to document this :)
non_defect
document image preloading you can now put docker image export tars at var lib rancher agent images and they will be read and loaded into containerd on start we need to document this
0
14,016
2,789,853,312
IssuesEvent
2015-05-08 21:55:54
google/google-visualization-api-issues
https://api.github.com/repos/google/google-visualization-api-issues
opened
In line chart bug
Priority-Medium Type-Defect
Original [issue 302](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=302) created by orwant on 2010-06-03T11:59:41.000Z: <b>What steps will reproduce the problem? Please provide a link to a</b> <b>demonstration page if at all possible, or attach code.</b> 1.when i called the url of line chart in iphone. its width is so long. it might not decreases. i try all the ways. at same, legends are displayed non-order. plz rectify those two things <b>2.</b> <b>3.</b> <b>What component is this issue related to (PieChart, LineChart, DataTable,</b> <b>Query, etc)?</b> <b>Are you using the test environment (version 1.1)?</b> <b>(If you are not sure, answer NO)</b> <b>What operating system and browser are you using?</b> <b>*********************************************************</b> <b>For developers viewing this issue: please click the 'star' icon to be</b> <b>notified of future changes, and to let us know how many of you are</b> <b>interested in seeing it resolved.</b> <b>*********************************************************</b>
1.0
In line chart bug - Original [issue 302](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=302) created by orwant on 2010-06-03T11:59:41.000Z: <b>What steps will reproduce the problem? Please provide a link to a</b> <b>demonstration page if at all possible, or attach code.</b> 1.when i called the url of line chart in iphone. its width is so long. it might not decreases. i try all the ways. at same, legends are displayed non-order. plz rectify those two things <b>2.</b> <b>3.</b> <b>What component is this issue related to (PieChart, LineChart, DataTable,</b> <b>Query, etc)?</b> <b>Are you using the test environment (version 1.1)?</b> <b>(If you are not sure, answer NO)</b> <b>What operating system and browser are you using?</b> <b>*********************************************************</b> <b>For developers viewing this issue: please click the 'star' icon to be</b> <b>notified of future changes, and to let us know how many of you are</b> <b>interested in seeing it resolved.</b> <b>*********************************************************</b>
defect
in line chart bug original created by orwant on what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code when i called the url of line chart in iphone its width is so long it might not decreases i try all the ways at same legends are displayed non order plz rectify those two things what component is this issue related to piechart linechart datatable query etc are you using the test environment version if you are not sure answer no what operating system and browser are you using for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved
1
41,531
10,510,933,309
IssuesEvent
2019-09-27 14:26:09
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
DSL.field(Name) fields should support schema mapping
C: Functionality E: All Editions P: Medium R: Fixed T: Defect
When using `DSL.field(Name)` with qualified field references, it should be possible to apply schema and table mappings. This is currently not being done. ~We should enable this feature, including a new Setting to turn it off to allow or users to have backwards compatible behaviour.~ ---- See also: https://stackoverflow.com/q/57222533/521799
1.0
DSL.field(Name) fields should support schema mapping - When using `DSL.field(Name)` with qualified field references, it should be possible to apply schema and table mappings. This is currently not being done. ~We should enable this feature, including a new Setting to turn it off to allow or users to have backwards compatible behaviour.~ ---- See also: https://stackoverflow.com/q/57222533/521799
defect
dsl field name fields should support schema mapping when using dsl field name with qualified field references it should be possible to apply schema and table mappings this is currently not being done we should enable this feature including a new setting to turn it off to allow or users to have backwards compatible behaviour see also
1
405,994
27,544,647,403
IssuesEvent
2023-03-07 10:53:35
rupali-codes/LinksHub
https://api.github.com/repos/rupali-codes/LinksHub
closed
[DOCS]: Improve README.md
documentation good first issue chore
### Description We can improve our readme by putting everything under a section like putting the last sentence under support section! ### Screenshots _No response_ ### Additional Information We can also add a discord icon in the community joining section and linked it to our discord server as it looks really cool :)
1.0
[DOCS]: Improve README.md - ### Description We can improve our readme by putting everything under a section like putting the last sentence under support section! ### Screenshots _No response_ ### Additional Information We can also add a discord icon in the community joining section and linked it to our discord server as it looks really cool :)
non_defect
improve readme md description we can improve our readme by putting everything under a section like putting the last sentence under support section screenshots no response additional information we can also add a discord icon in the community joining section and linked it to our discord server as it looks really cool
0
35,359
7,714,927,045
IssuesEvent
2018-05-23 05:13:44
bridgedotnet/Bridge.Newtonsoft.Json
https://api.github.com/repos/bridgedotnet/Bridge.Newtonsoft.Json
closed
DateTimeOffset deserialization fails from different time zone to deck server
defect
When I serialize and deserialize a new DateTimeOffset object on deck.net, I get the following error: `Argument_OffsetPrecision, Parameter name: offset` I am working in the UK so my timezone is set to BST which is UTC +1 hour. You can view the deck here: https://deck.net/ba083a00c886571deec195adc6497b53 And the screenshot of the issue i'm experiencing if your timezone does not match mine and does not recreate issue: ![image](https://user-images.githubusercontent.com/13521143/39571145-759409ec-4ec2-11e8-956f-300a8ccbfff4.png) Originally reported here: https://github.com/bridgedotnet/Bridge.Newtonsoft.Json/issues/89
1.0
DateTimeOffset deserialization fails from different time zone to deck server - When I serialize and deserialize a new DateTimeOffset object on deck.net, I get the following error: `Argument_OffsetPrecision, Parameter name: offset` I am working in the UK so my timezone is set to BST which is UTC +1 hour. You can view the deck here: https://deck.net/ba083a00c886571deec195adc6497b53 And the screenshot of the issue i'm experiencing if your timezone does not match mine and does not recreate issue: ![image](https://user-images.githubusercontent.com/13521143/39571145-759409ec-4ec2-11e8-956f-300a8ccbfff4.png) Originally reported here: https://github.com/bridgedotnet/Bridge.Newtonsoft.Json/issues/89
defect
datetimeoffset deserialization fails from different time zone to deck server when i serialize and deserialize a new datetimeoffset object on deck net i get the following error argument offsetprecision parameter name offset i am working in the uk so my timezone is set to bst which is utc hour you can view the deck here and the screenshot of the issue i m experiencing if your timezone does not match mine and does not recreate issue originally reported here
1
107,460
11,544,739,845
IssuesEvent
2020-02-18 12:05:29
trimox/angular-mdc-web
https://api.github.com/repos/trimox/angular-mdc-web
closed
demos: Use scrollTarget for top-app-bar elevation changes on scroll
documentation
Noticed the demo site is not changing elevation on scroll event.
1.0
demos: Use scrollTarget for top-app-bar elevation changes on scroll - Noticed the demo site is not changing elevation on scroll event.
non_defect
demos use scrolltarget for top app bar elevation changes on scroll noticed the demo site is not changing elevation on scroll event
0
48,796
13,184,743,048
IssuesEvent
2020-08-12 20:00:46
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Too many back-to-back short runs in Pdaq cause PnF processing to stall (Trac #222)
Incomplete Migration Migrated from Trac defect jeb + pnf
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/222 , reported by blaufuss and owned by tschmidt_</summary> <p> ```json { "status": "closed", "changetime": "2012-05-25T13:41:49", "description": "Too many short pdaq runs (that actually produce a few events each) cause PnF processing to slow to a crawl. \n\nRun transitions are hard at PnF, with clients needing to request new GCD, etc, server needing to close open files, etc...\n\nNeed to manage this transition better. Options:\n\n1. Improve GCDispatch, so it caches several recent runs, so a client flopping between runs isn't a problem\n\n2. Make PFServer block data from new run until it finishes current run. May slow transitons a bit, but might be more robust.", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1337953309000000", "component": "jeb + pnf", "summary": "Too many back-to-back short runs in Pdaq cause PnF processing to stall", "priority": "normal", "keywords": "", "time": "2010-12-01T16:43:51", "milestone": "", "owner": "tschmidt", "type": "defect" } ``` </p> </details>
1.0
Too many back-to-back short runs in Pdaq cause PnF processing to stall (Trac #222) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/222 , reported by blaufuss and owned by tschmidt_</summary> <p> ```json { "status": "closed", "changetime": "2012-05-25T13:41:49", "description": "Too many short pdaq runs (that actually produce a few events each) cause PnF processing to slow to a crawl. \n\nRun transitions are hard at PnF, with clients needing to request new GCD, etc, server needing to close open files, etc...\n\nNeed to manage this transition better. Options:\n\n1. Improve GCDispatch, so it caches several recent runs, so a client flopping between runs isn't a problem\n\n2. Make PFServer block data from new run until it finishes current run. May slow transitons a bit, but might be more robust.", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1337953309000000", "component": "jeb + pnf", "summary": "Too many back-to-back short runs in Pdaq cause PnF processing to stall", "priority": "normal", "keywords": "", "time": "2010-12-01T16:43:51", "milestone": "", "owner": "tschmidt", "type": "defect" } ``` </p> </details>
defect
too many back to back short runs in pdaq cause pnf processing to stall trac migrated from reported by blaufuss and owned by tschmidt json status closed changetime description too many short pdaq runs that actually produce a few events each cause pnf processing to slow to a crawl n nrun transitions are hard at pnf with clients needing to request new gcd etc server needing to close open files etc n nneed to manage this transition better options n improve gcdispatch so it caches several recent runs so a client flopping between runs isn t a problem n make pfserver block data from new run until it finishes current run may slow transitons a bit but might be more robust reporter blaufuss cc resolution fixed ts component jeb pnf summary too many back to back short runs in pdaq cause pnf processing to stall priority normal keywords time milestone owner tschmidt type defect
1
32,929
6,970,593,374
IssuesEvent
2017-12-11 10:51:46
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
p:calendar DateTimePicker error in javascript for mode="inline"
defect
I´m reopening an bug from your old issuetracker... https://code.google.com/p/primefaces/issues/detail?id=6790&thanks=6790&ts=1397477544 for more information also read http://stackoverflow.com/questions/23036356/bug-in-primefaces-calendar-timepicker issues still exists in 5.1.19 and 5.2.6
1.0
p:calendar DateTimePicker error in javascript for mode="inline" - I´m reopening an bug from your old issuetracker... https://code.google.com/p/primefaces/issues/detail?id=6790&thanks=6790&ts=1397477544 for more information also read http://stackoverflow.com/questions/23036356/bug-in-primefaces-calendar-timepicker issues still exists in 5.1.19 and 5.2.6
defect
p calendar datetimepicker error in javascript for mode inline i´m reopening an bug from your old issuetracker for more information also read issues still exists in and
1
447,709
12,892,074,755
IssuesEvent
2020-07-13 18:55:32
coyiutoc/Honda-Capstone-Salmon
https://api.github.com/repos/coyiutoc/Honda-Capstone-Salmon
opened
Order tags in search column by # of supporting evidence
high priority
Tags in search column need # of supporting evidence included + should be ordered from most to greatest. e.g. [ Salmon (12) ] [ Tuna (10) ] [ Snapper (4) ]
1.0
Order tags in search column by # of supporting evidence - Tags in search column need # of supporting evidence included + should be ordered from most to greatest. e.g. [ Salmon (12) ] [ Tuna (10) ] [ Snapper (4) ]
non_defect
order tags in search column by of supporting evidence tags in search column need of supporting evidence included should be ordered from most to greatest e g
0
67,679
21,051,814,942
IssuesEvent
2022-03-31 21:15:18
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]: InvocationTargetException while using dynamic Selenium grid 4
I-defect needs-triaging
### What happened? Getting InvocationTargetException while using dynamic Selenium grid 4. We have downloaded docker compose file (docker-compose-v3-dynamic-grid.yml) from https://github.com/SeleniumHQ/docker-selenium/blob/trunk/docker-compose-v3-dynamic-grid.yml and created config.toml file using the code from https://github.com/SeleniumHQ/docker-selenium#dynamic-grid without changing anything. The config.toml file is place in NodeDocker folder. ### How can we reproduce the issue? ```shell selenium-compose.yml: version: "3" services: node-docker: image: selenium/node-docker:latest volumes: - ./assets:/opt/selenium/assets - ./config/config.toml:/opt/bin/config.toml depends_on: - selenium-hub environment: - SE_EVENT_BUS_HOST=selenium-hub - SE_EVENT_BUS_PUBLISH_PORT=4442 - SE_EVENT_BUS_SUBSCRIBE_PORT=4443 selenium-hub: image: selenium/hub:4.1.3-20220327 container_name: selenium-hub ports: - "4442:4442" - "4443:4443" - "4444:4444" config.toml [docker] # Configs have a mapping between the Docker image to use and the capabilities that need to be matched to # start a container with the given image. configs = [ "selenium/standalone-chrome:latest", "{\"browserName\": \"chrome\"}" "selenium/standalone-firefox:latest", "{\"browserName\": \"firefox\"}" ] # URL for connecting to the docker daemon # Most simple approach, leave it as http://127.0.0.1:2375, and mount /var/run/docker.sock. # 127.0.0.1 is used because interally the container uses socat when /var/run/docker.sock is mounted # If var/run/docker.sock is not mounted: # Windows: make sure Docker Desktop exposes the daemon via tcp, and use http://host.docker.internal:2375. # macOS: install socat and run the following command, socat -4 TCP-LISTEN:2375,fork UNIX-CONNECT:/var/run/docker.sock, # then use http://host.docker.internal:2375. # Linux: varies from machine to machine, please mount /var/run/docker.sock. If this does not work, please create an issue. url = "http://host.docker.internal:2375" # Docker image used for video recording video-image = "selenium/video:latest" # Uncomment the following section if you are running the node on a separate VM # Fill out the placeholders with appropriate values #[server] #host = <ip-from-node-machine> #port = <port-from-node-machine> Running below docker compose command: docker-compose -f selenium-compose.yml up ``` ### Relevant log output ```shell selenium-hub | 2022-03-31 20:24:30,036 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-hub.conf" during parsing selenium-hub | 2022-03-31 20:24:30,039 INFO supervisord started with pid 8 selenium-hub | 2022-03-31 20:24:31,041 INFO spawned: 'selenium-grid-hub' with pid 10 selenium-hub | 20:24:31.411 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding selenium-hub | 2022-03-31 20:24:31,413 INFO success: selenium-grid-hub entered RUNNING state, process has stayed up for > than 0 seconds (startsecs) selenium-hub | 20:24:31.416 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing selenium-hub | 20:24:31.517 INFO [BoundZmqEventBus.<init>] - XPUB binding to [binding to tcp://*:4442, advertising as tcp://172.22.0.2:4442], XSUB binding to [binding to tcp://*:4443, advertising as tcp://172.22.0.2:4443] selenium-hub | 20:24:31.567 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://172.22.0.2:4442 and tcp://172.22.0.2:4443 selenium-hub | 20:24:31.587 INFO [UnboundZmqEventBus.<init>] - Sockets created selenium-hub | 20:24:32.589 INFO [UnboundZmqEventBus.<init>] - Event bus ready selenium-hub | 20:24:33.198 INFO [Hub.execute] - Started Selenium Hub 4.1.3 (revision 7b1ebf28ef): http://172.22.0.2:4444 selenium-hub | Trapped SIGTERM/SIGINT/x so shutting down supervisord... selenium-hub | 2022-03-31 20:31:13,705 WARN received SIGTERM indicating exit request selenium-hub | 2022-03-31 20:31:13,706 INFO waiting for selenium-grid-hub to die selenium-hub | 2022-03-31 20:31:14,707 INFO stopped: selenium-grid-hub (terminated by SIGTERM) selenium-hub | Shutdown complete selenium-hub | 2022-03-31 20:31:35,542 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-hub.conf" during parsing selenium-hub | 2022-03-31 20:31:35,545 INFO supervisord started with pid 9 selenium-hub | 2022-03-31 20:31:36,547 INFO spawned: 'selenium-grid-hub' with pid 11 selenium-hub | 20:31:36.964 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding selenium-hub | 2022-03-31 20:31:36,968 INFO success: selenium-grid-hub entered RUNNING state, process has stayed up for > than 0 seconds (startsecs) selenium-hub | 20:31:36.971 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing selenium-hub | 20:31:37.102 INFO [BoundZmqEventBus.<init>] - XPUB binding to [binding to tcp://*:4442, advertising as tcp://172.22.0.2:4442], XSUB binding to [binding to tcp://*:4443, advertising as tcp://172.22.0.2:4443] selenium-hub | 20:31:37.158 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://172.22.0.2:4442 and tcp://172.22.0.2:4443 selenium-hub | 20:31:37.181 INFO [UnboundZmqEventBus.<init>] - Sockets created selenium-hub | 20:31:38.183 INFO [UnboundZmqEventBus.<init>] - Event bus ready selenium-hub | 20:31:38.936 INFO [Hub.execute] - Started Selenium Hub 4.1.3 (revision 7b1ebf28ef): http://172.22.0.2:4444 node-docker_1 | 2022-03-31 20:33:12,162 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-docker.conf" during parsing node-docker_1 | 2022-03-31 20:33:12,165 INFO supervisord started with pid 8 node-docker_1 | 2022-03-31 20:33:13,167 INFO spawned: 'socat' with pid 10 node-docker_1 | 2022-03-31 20:33:13,169 INFO spawned: 'selenium-grid-docker' with pid 11 node-docker_1 | 2022-03-31 20:33:13,176 INFO success: socat entered RUNNING state, process has stayed up for > than 0 seconds (startsecs) node-docker_1 | 2022-03-31 20:33:13,177 INFO success: selenium-grid-docker entered RUNNING state, process has stayed up for > than 0 seconds (startsecs) node-docker_1 | 2022-03-31 20:33:13,177 INFO exited: socat (exit status 0; expected) node-docker_1 | Starting Selenium Grid Node Docker... node-docker_1 | 20:33:13.656 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding node-docker_1 | 20:33:13.663 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing node-docker_1 | 20:33:13.831 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://selenium-hub:4442 and tcp://selenium-hub:4443 node-docker_1 | 20:33:13.889 INFO [UnboundZmqEventBus.<init>] - Sockets created node-docker_1 | 20:33:14.892 INFO [UnboundZmqEventBus.<init>] - Event bus ready node-docker_1 | 20:33:15.016 INFO [NodeServer.createHandlers] - Reporting self as: http://172.22.0.3:5555 node-docker_1 | 20:33:15.032 INFO [NodeOptions.getSessionFactories] - Detected 12 available processors node-docker_1 | java.lang.reflect.InvocationTargetException node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) node-docker_1 | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) node-docker_1 | at java.base/java.lang.reflect.Method.invoke(Method.java:566) node-docker_1 | at org.openqa.selenium.grid.Bootstrap.runMain(Bootstrap.java:77) node-docker_1 | at org.openqa.selenium.grid.Bootstrap.main(Bootstrap.java:70) node-docker_1 | Caused by: org.openqa.selenium.grid.config.ConfigException: java.lang.reflect.InvocationTargetException node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.getClass(MemoizedConfig.java:115) node-docker_1 | at org.openqa.selenium.grid.node.config.NodeOptions.getNode(NodeOptions.java:149) node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.createHandlers(NodeServer.java:127) node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.asServer(NodeServer.java:183) node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.execute(NodeServer.java:231) node-docker_1 | at org.openqa.selenium.grid.TemplateGridCommand.lambda$configure$4(TemplateGridCommand.java:129) node-docker_1 | at org.openqa.selenium.grid.Main.launch(Main.java:83) node-docker_1 | at org.openqa.selenium.grid.Main.go(Main.java:57) node-docker_1 | at org.openqa.selenium.grid.Main.main(Main.java:42) node-docker_1 | ... 6 more node-docker_1 | Caused by: java.lang.reflect.InvocationTargetException node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) node-docker_1 | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) node-docker_1 | at java.base/java.lang.reflect.Method.invoke(Method.java:566) node-docker_1 | at org.openqa.selenium.grid.config.ClassCreation.callCreateMethod(ClassCreation.java:50) node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.lambda$getClass$4(MemoizedConfig.java:100) node-docker_1 | at java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1737) node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.getClass(MemoizedConfig.java:95) node-docker_1 | ... 14 more node-docker_1 | Caused by: org.openqa.selenium.docker.DockerException: Unable to reach the Docker daemon at http://127.0.0.1:2375 node-docker_1 | at org.openqa.selenium.grid.node.docker.DockerOptions.getDockerSessionFactories(DockerOptions.java:125) node-docker_1 | at org.openqa.selenium.grid.node.local.LocalNodeFactory.create(LocalNodeFactory.java:80) node-docker_1 | ... 22 more node-docker_1 | Exception in thread "Thread-0" java.lang.NullPointerException node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.lambda$new$0(NodeServer.java:79) node-docker_1 | at java.base/java.lang.Thread.run(Thread.java:829) node-docker_1 | 2022-03-31 20:33:15,858 INFO exited: selenium-grid-docker (exit status 1; not expected) ``` ### Operating System Windows 10 ### Selenium version Java 4.1.3 ### What are the browser(s) and version(s) where you see this issue? dynamic grid using docker ### What are the browser driver(s) and version(s) where you see this issue? selenium/node-docker:latest, selenium/standalone-chrome:latest, selenium/standalone-firefox:latest ### Are you using Selenium Grid? 4
1.0
[🐛 Bug]: InvocationTargetException while using dynamic Selenium grid 4 - ### What happened? Getting InvocationTargetException while using dynamic Selenium grid 4. We have downloaded docker compose file (docker-compose-v3-dynamic-grid.yml) from https://github.com/SeleniumHQ/docker-selenium/blob/trunk/docker-compose-v3-dynamic-grid.yml and created config.toml file using the code from https://github.com/SeleniumHQ/docker-selenium#dynamic-grid without changing anything. The config.toml file is place in NodeDocker folder. ### How can we reproduce the issue? ```shell selenium-compose.yml: version: "3" services: node-docker: image: selenium/node-docker:latest volumes: - ./assets:/opt/selenium/assets - ./config/config.toml:/opt/bin/config.toml depends_on: - selenium-hub environment: - SE_EVENT_BUS_HOST=selenium-hub - SE_EVENT_BUS_PUBLISH_PORT=4442 - SE_EVENT_BUS_SUBSCRIBE_PORT=4443 selenium-hub: image: selenium/hub:4.1.3-20220327 container_name: selenium-hub ports: - "4442:4442" - "4443:4443" - "4444:4444" config.toml [docker] # Configs have a mapping between the Docker image to use and the capabilities that need to be matched to # start a container with the given image. configs = [ "selenium/standalone-chrome:latest", "{\"browserName\": \"chrome\"}" "selenium/standalone-firefox:latest", "{\"browserName\": \"firefox\"}" ] # URL for connecting to the docker daemon # Most simple approach, leave it as http://127.0.0.1:2375, and mount /var/run/docker.sock. # 127.0.0.1 is used because interally the container uses socat when /var/run/docker.sock is mounted # If var/run/docker.sock is not mounted: # Windows: make sure Docker Desktop exposes the daemon via tcp, and use http://host.docker.internal:2375. # macOS: install socat and run the following command, socat -4 TCP-LISTEN:2375,fork UNIX-CONNECT:/var/run/docker.sock, # then use http://host.docker.internal:2375. # Linux: varies from machine to machine, please mount /var/run/docker.sock. If this does not work, please create an issue. url = "http://host.docker.internal:2375" # Docker image used for video recording video-image = "selenium/video:latest" # Uncomment the following section if you are running the node on a separate VM # Fill out the placeholders with appropriate values #[server] #host = <ip-from-node-machine> #port = <port-from-node-machine> Running below docker compose command: docker-compose -f selenium-compose.yml up ``` ### Relevant log output ```shell selenium-hub | 2022-03-31 20:24:30,036 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-hub.conf" during parsing selenium-hub | 2022-03-31 20:24:30,039 INFO supervisord started with pid 8 selenium-hub | 2022-03-31 20:24:31,041 INFO spawned: 'selenium-grid-hub' with pid 10 selenium-hub | 20:24:31.411 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding selenium-hub | 2022-03-31 20:24:31,413 INFO success: selenium-grid-hub entered RUNNING state, process has stayed up for > than 0 seconds (startsecs) selenium-hub | 20:24:31.416 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing selenium-hub | 20:24:31.517 INFO [BoundZmqEventBus.<init>] - XPUB binding to [binding to tcp://*:4442, advertising as tcp://172.22.0.2:4442], XSUB binding to [binding to tcp://*:4443, advertising as tcp://172.22.0.2:4443] selenium-hub | 20:24:31.567 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://172.22.0.2:4442 and tcp://172.22.0.2:4443 selenium-hub | 20:24:31.587 INFO [UnboundZmqEventBus.<init>] - Sockets created selenium-hub | 20:24:32.589 INFO [UnboundZmqEventBus.<init>] - Event bus ready selenium-hub | 20:24:33.198 INFO [Hub.execute] - Started Selenium Hub 4.1.3 (revision 7b1ebf28ef): http://172.22.0.2:4444 selenium-hub | Trapped SIGTERM/SIGINT/x so shutting down supervisord... selenium-hub | 2022-03-31 20:31:13,705 WARN received SIGTERM indicating exit request selenium-hub | 2022-03-31 20:31:13,706 INFO waiting for selenium-grid-hub to die selenium-hub | 2022-03-31 20:31:14,707 INFO stopped: selenium-grid-hub (terminated by SIGTERM) selenium-hub | Shutdown complete selenium-hub | 2022-03-31 20:31:35,542 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-hub.conf" during parsing selenium-hub | 2022-03-31 20:31:35,545 INFO supervisord started with pid 9 selenium-hub | 2022-03-31 20:31:36,547 INFO spawned: 'selenium-grid-hub' with pid 11 selenium-hub | 20:31:36.964 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding selenium-hub | 2022-03-31 20:31:36,968 INFO success: selenium-grid-hub entered RUNNING state, process has stayed up for > than 0 seconds (startsecs) selenium-hub | 20:31:36.971 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing selenium-hub | 20:31:37.102 INFO [BoundZmqEventBus.<init>] - XPUB binding to [binding to tcp://*:4442, advertising as tcp://172.22.0.2:4442], XSUB binding to [binding to tcp://*:4443, advertising as tcp://172.22.0.2:4443] selenium-hub | 20:31:37.158 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://172.22.0.2:4442 and tcp://172.22.0.2:4443 selenium-hub | 20:31:37.181 INFO [UnboundZmqEventBus.<init>] - Sockets created selenium-hub | 20:31:38.183 INFO [UnboundZmqEventBus.<init>] - Event bus ready selenium-hub | 20:31:38.936 INFO [Hub.execute] - Started Selenium Hub 4.1.3 (revision 7b1ebf28ef): http://172.22.0.2:4444 node-docker_1 | 2022-03-31 20:33:12,162 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-docker.conf" during parsing node-docker_1 | 2022-03-31 20:33:12,165 INFO supervisord started with pid 8 node-docker_1 | 2022-03-31 20:33:13,167 INFO spawned: 'socat' with pid 10 node-docker_1 | 2022-03-31 20:33:13,169 INFO spawned: 'selenium-grid-docker' with pid 11 node-docker_1 | 2022-03-31 20:33:13,176 INFO success: socat entered RUNNING state, process has stayed up for > than 0 seconds (startsecs) node-docker_1 | 2022-03-31 20:33:13,177 INFO success: selenium-grid-docker entered RUNNING state, process has stayed up for > than 0 seconds (startsecs) node-docker_1 | 2022-03-31 20:33:13,177 INFO exited: socat (exit status 0; expected) node-docker_1 | Starting Selenium Grid Node Docker... node-docker_1 | 20:33:13.656 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding node-docker_1 | 20:33:13.663 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing node-docker_1 | 20:33:13.831 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://selenium-hub:4442 and tcp://selenium-hub:4443 node-docker_1 | 20:33:13.889 INFO [UnboundZmqEventBus.<init>] - Sockets created node-docker_1 | 20:33:14.892 INFO [UnboundZmqEventBus.<init>] - Event bus ready node-docker_1 | 20:33:15.016 INFO [NodeServer.createHandlers] - Reporting self as: http://172.22.0.3:5555 node-docker_1 | 20:33:15.032 INFO [NodeOptions.getSessionFactories] - Detected 12 available processors node-docker_1 | java.lang.reflect.InvocationTargetException node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) node-docker_1 | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) node-docker_1 | at java.base/java.lang.reflect.Method.invoke(Method.java:566) node-docker_1 | at org.openqa.selenium.grid.Bootstrap.runMain(Bootstrap.java:77) node-docker_1 | at org.openqa.selenium.grid.Bootstrap.main(Bootstrap.java:70) node-docker_1 | Caused by: org.openqa.selenium.grid.config.ConfigException: java.lang.reflect.InvocationTargetException node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.getClass(MemoizedConfig.java:115) node-docker_1 | at org.openqa.selenium.grid.node.config.NodeOptions.getNode(NodeOptions.java:149) node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.createHandlers(NodeServer.java:127) node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.asServer(NodeServer.java:183) node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.execute(NodeServer.java:231) node-docker_1 | at org.openqa.selenium.grid.TemplateGridCommand.lambda$configure$4(TemplateGridCommand.java:129) node-docker_1 | at org.openqa.selenium.grid.Main.launch(Main.java:83) node-docker_1 | at org.openqa.selenium.grid.Main.go(Main.java:57) node-docker_1 | at org.openqa.selenium.grid.Main.main(Main.java:42) node-docker_1 | ... 6 more node-docker_1 | Caused by: java.lang.reflect.InvocationTargetException node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) node-docker_1 | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) node-docker_1 | at java.base/java.lang.reflect.Method.invoke(Method.java:566) node-docker_1 | at org.openqa.selenium.grid.config.ClassCreation.callCreateMethod(ClassCreation.java:50) node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.lambda$getClass$4(MemoizedConfig.java:100) node-docker_1 | at java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1737) node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.getClass(MemoizedConfig.java:95) node-docker_1 | ... 14 more node-docker_1 | Caused by: org.openqa.selenium.docker.DockerException: Unable to reach the Docker daemon at http://127.0.0.1:2375 node-docker_1 | at org.openqa.selenium.grid.node.docker.DockerOptions.getDockerSessionFactories(DockerOptions.java:125) node-docker_1 | at org.openqa.selenium.grid.node.local.LocalNodeFactory.create(LocalNodeFactory.java:80) node-docker_1 | ... 22 more node-docker_1 | Exception in thread "Thread-0" java.lang.NullPointerException node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.lambda$new$0(NodeServer.java:79) node-docker_1 | at java.base/java.lang.Thread.run(Thread.java:829) node-docker_1 | 2022-03-31 20:33:15,858 INFO exited: selenium-grid-docker (exit status 1; not expected) ``` ### Operating System Windows 10 ### Selenium version Java 4.1.3 ### What are the browser(s) and version(s) where you see this issue? dynamic grid using docker ### What are the browser driver(s) and version(s) where you see this issue? selenium/node-docker:latest, selenium/standalone-chrome:latest, selenium/standalone-firefox:latest ### Are you using Selenium Grid? 4
defect
invocationtargetexception while using dynamic selenium grid what happened getting invocationtargetexception while using dynamic selenium grid we have downloaded docker compose file docker compose dynamic grid yml from and created config toml file using the code from without changing anything the config toml file is place in nodedocker folder how can we reproduce the issue shell selenium compose yml version services node docker image selenium node docker latest volumes assets opt selenium assets config config toml opt bin config toml depends on selenium hub environment se event bus host selenium hub se event bus publish port se event bus subscribe port selenium hub image selenium hub container name selenium hub ports config toml configs have a mapping between the docker image to use and the capabilities that need to be matched to start a container with the given image configs selenium standalone chrome latest browsername chrome selenium standalone firefox latest browsername firefox url for connecting to the docker daemon most simple approach leave it as and mount var run docker sock is used because interally the container uses socat when var run docker sock is mounted if var run docker sock is not mounted windows make sure docker desktop exposes the daemon via tcp and use macos install socat and run the following command socat tcp listen fork unix connect var run docker sock then use linux varies from machine to machine please mount var run docker sock if this does not work please create an issue url docker image used for video recording video image selenium video latest uncomment the following section if you are running the node on a separate vm fill out the placeholders with appropriate values host port running below docker compose command docker compose f selenium compose yml up relevant log output shell selenium hub info included extra file etc supervisor conf d selenium grid hub conf during parsing selenium hub info supervisord started with pid selenium hub info spawned selenium grid hub with pid selenium hub info using the system default encoding selenium hub info success selenium grid hub entered running state process has stayed up for than seconds startsecs selenium hub info using opentelemetry for tracing selenium hub info xpub binding to xsub binding to selenium hub info connecting to tcp and tcp selenium hub info sockets created selenium hub info event bus ready selenium hub info started selenium hub revision selenium hub trapped sigterm sigint x so shutting down supervisord selenium hub warn received sigterm indicating exit request selenium hub info waiting for selenium grid hub to die selenium hub info stopped selenium grid hub terminated by sigterm selenium hub shutdown complete selenium hub info included extra file etc supervisor conf d selenium grid hub conf during parsing selenium hub info supervisord started with pid selenium hub info spawned selenium grid hub with pid selenium hub info using the system default encoding selenium hub info success selenium grid hub entered running state process has stayed up for than seconds startsecs selenium hub info using opentelemetry for tracing selenium hub info xpub binding to xsub binding to selenium hub info connecting to tcp and tcp selenium hub info sockets created selenium hub info event bus ready selenium hub info started selenium hub revision node docker info included extra file etc supervisor conf d selenium grid docker conf during parsing node docker info supervisord started with pid node docker info spawned socat with pid node docker info spawned selenium grid docker with pid node docker info success socat entered running state process has stayed up for than seconds startsecs node docker info success selenium grid docker entered running state process has stayed up for than seconds startsecs node docker info exited socat exit status expected node docker starting selenium grid node docker node docker info using the system default encoding node docker info using opentelemetry for tracing node docker info connecting to tcp selenium hub and tcp selenium hub node docker info sockets created node docker info event bus ready node docker info reporting self as node docker info detected available processors node docker java lang reflect invocationtargetexception node docker at java base jdk internal reflect nativemethodaccessorimpl native method node docker at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java node docker at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java node docker at java base java lang reflect method invoke method java node docker at org openqa selenium grid bootstrap runmain bootstrap java node docker at org openqa selenium grid bootstrap main bootstrap java node docker caused by org openqa selenium grid config configexception java lang reflect invocationtargetexception node docker at org openqa selenium grid config memoizedconfig getclass memoizedconfig java node docker at org openqa selenium grid node config nodeoptions getnode nodeoptions java node docker at org openqa selenium grid node httpd nodeserver createhandlers nodeserver java node docker at org openqa selenium grid node httpd nodeserver asserver nodeserver java node docker at org openqa selenium grid node httpd nodeserver execute nodeserver java node docker at org openqa selenium grid templategridcommand lambda configure templategridcommand java node docker at org openqa selenium grid main launch main java node docker at org openqa selenium grid main go main java node docker at org openqa selenium grid main main main java node docker more node docker caused by java lang reflect invocationtargetexception node docker at java base jdk internal reflect nativemethodaccessorimpl native method node docker at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java node docker at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java node docker at java base java lang reflect method invoke method java node docker at org openqa selenium grid config classcreation callcreatemethod classcreation java node docker at org openqa selenium grid config memoizedconfig lambda getclass memoizedconfig java node docker at java base java util concurrent concurrenthashmap computeifabsent concurrenthashmap java node docker at org openqa selenium grid config memoizedconfig getclass memoizedconfig java node docker more node docker caused by org openqa selenium docker dockerexception unable to reach the docker daemon at node docker at org openqa selenium grid node docker dockeroptions getdockersessionfactories dockeroptions java node docker at org openqa selenium grid node local localnodefactory create localnodefactory java node docker more node docker exception in thread thread java lang nullpointerexception node docker at org openqa selenium grid node httpd nodeserver lambda new nodeserver java node docker at java base java lang thread run thread java node docker info exited selenium grid docker exit status not expected operating system windows selenium version java what are the browser s and version s where you see this issue dynamic grid using docker what are the browser driver s and version s where you see this issue selenium node docker latest selenium standalone chrome latest selenium standalone firefox latest are you using selenium grid
1
106,414
13,284,191,262
IssuesEvent
2020-08-24 05:40:35
Qiskit/qiskit.org
https://api.github.com/repos/Qiskit/qiskit.org
closed
Compare Marina and Human design work
Design
Design kickoff session with human to compare and align on work to be done
1.0
Compare Marina and Human design work - Design kickoff session with human to compare and align on work to be done
non_defect
compare marina and human design work design kickoff session with human to compare and align on work to be done
0
6,587
2,610,257,312
IssuesEvent
2015-02-26 19:22:10
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳激光去痤疮有副作用吗
auto-migrated Priority-Medium Type-Defect
``` 深圳激光去痤疮有副作用吗【深圳韩方科颜全国热线400-869-181 8,24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构�� �韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳� ��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不 反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创�� �内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客� ��上的痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:40
1.0
深圳激光去痤疮有副作用吗 - ``` 深圳激光去痤疮有副作用吗【深圳韩方科颜全国热线400-869-181 8,24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构�� �韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳� ��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不 反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创�� �内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客� ��上的痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:40
defect
深圳激光去痤疮有副作用吗 深圳激光去痤疮有副作用吗【 , 】深圳韩方科颜专业祛痘连锁机构,机构�� �韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳� ��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不 反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创�� �内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客� ��上的痘痘。 original issue reported on code google com by szft com on may at
1
133,191
28,761,071,044
IssuesEvent
2023-05-01 00:44:39
muradshahsuvarov/employee-commute-route-optimizer
https://api.github.com/repos/muradshahsuvarov/employee-commute-route-optimizer
closed
Implement multiple location route finder algorithm
code
Let N be the number of commute locations. Using the route finder, find the shortest route considering traffic, and etc.
1.0
Implement multiple location route finder algorithm - Let N be the number of commute locations. Using the route finder, find the shortest route considering traffic, and etc.
non_defect
implement multiple location route finder algorithm let n be the number of commute locations using the route finder find the shortest route considering traffic and etc
0
50,348
13,187,454,304
IssuesEvent
2020-08-13 03:28:00
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters (Trac #565)
Migrated from Trac defect tools/ports
There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters. A patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached. <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/565 , reported by kislat and owned by cgils</em></summary> <p> ```json { "status": "closed", "changetime": "2009-06-30T14:16:42", "description": "There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.\n\nA patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.", "reporter": "kislat", "cc": "", "resolution": "fixed", "_ts": "1246371402000000", "component": "tools/ports", "summary": "rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters", "priority": "normal", "keywords": "", "time": "2009-06-30T12:29:54", "milestone": "", "owner": "cgils", "type": "defect" } ``` </p> </details>
1.0
rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters (Trac #565) - There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters. A patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached. <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/565 , reported by kislat and owned by cgils</em></summary> <p> ```json { "status": "closed", "changetime": "2009-06-30T14:16:42", "description": "There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.\n\nA patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.", "reporter": "kislat", "cc": "", "resolution": "fixed", "_ts": "1246371402000000", "component": "tools/ports", "summary": "rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters", "priority": "normal", "keywords": "", "time": "2009-06-30T12:29:54", "milestone": "", "owner": "cgils", "type": "defect" } ``` </p> </details>
defect
rootcint won t create dictionaries if rootsys is longer than characters trac there is a bug in rootcint root and later versions that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable after resolving symlinks is longer than characters a patch for the file root source file core utils src rootcint cxx for root is attached migrated from reported by kislat and owned by cgils json status closed changetime description there is a bug in rootcint root and later versions that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable after resolving symlinks is longer than characters n na patch for the file root source file core utils src rootcint cxx for root is attached reporter kislat cc resolution fixed ts component tools ports summary rootcint won t create dictionaries if rootsys is longer than characters priority normal keywords time milestone owner cgils type defect
1
114,045
9,672,687,759
IssuesEvent
2019-05-22 04:46:13
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
mysql_user should flush privileges when changed
affects_2.3 bug database module mysql support:community test waiting_on_contributor
From @lordnynex on 2015-10-26T22:48:07Z ##### Issue Type: Bug Report ##### Component Name: mysql_user module ##### Ansible Version: Any version using mysql_user ##### Ansible Configuration: N/A ##### Environment: N/A ##### Summary: mysql_user should flush privileges when changed. Currently requires extra task to reload mysql or manually flush privileges. ##### Steps To Reproduce: ``` yaml - mysql_user: name=root password=root priv=*.*:ALL state=present check_implicit_admin=yes ``` ##### Expected Results: Changed credentials should be usable in subsequent tasks without a mysql reload. Ex ``` yaml - shell: mysqladmin -u root password root ``` ##### Actual Results: Tasks/commands attempting to use the modified credentials will fail execution. Copied from original issue: ansible/ansible-modules-core#2357
1.0
mysql_user should flush privileges when changed - From @lordnynex on 2015-10-26T22:48:07Z ##### Issue Type: Bug Report ##### Component Name: mysql_user module ##### Ansible Version: Any version using mysql_user ##### Ansible Configuration: N/A ##### Environment: N/A ##### Summary: mysql_user should flush privileges when changed. Currently requires extra task to reload mysql or manually flush privileges. ##### Steps To Reproduce: ``` yaml - mysql_user: name=root password=root priv=*.*:ALL state=present check_implicit_admin=yes ``` ##### Expected Results: Changed credentials should be usable in subsequent tasks without a mysql reload. Ex ``` yaml - shell: mysqladmin -u root password root ``` ##### Actual Results: Tasks/commands attempting to use the modified credentials will fail execution. Copied from original issue: ansible/ansible-modules-core#2357
non_defect
mysql user should flush privileges when changed from lordnynex on issue type bug report component name mysql user module ansible version any version using mysql user ansible configuration n a environment n a summary mysql user should flush privileges when changed currently requires extra task to reload mysql or manually flush privileges steps to reproduce yaml mysql user name root password root priv all state present check implicit admin yes expected results changed credentials should be usable in subsequent tasks without a mysql reload ex yaml shell mysqladmin u root password root actual results tasks commands attempting to use the modified credentials will fail execution copied from original issue ansible ansible modules core
0
65,974
19,842,995,580
IssuesEvent
2022-01-21 00:45:24
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Mentioning a user with emojis in the display name causes Element Android to 'mention' a few times
T-Defect
### Steps to reproduce 1. Mention a person with emojis in their username 2. Send it ### Outcome #### What did you expect? Mentions once #### What happened instead? It 'mentions' a few times, though in other Matrix clients it appears as one ![Element Android](https://user-images.githubusercontent.com/47475676/150444960-ce373164-6413-47dc-bd0a-6f12ef9ae276.jpg) ![FluffyChat](https://user-images.githubusercontent.com/47475676/150445054-aa6d057d-8e88-4ae1-a52d-8b9dab54187a.jpg) ### Your phone model Mito W1 ### Operating system version Android 9 ### Application version and app store 1.3.15 [40103152] (G-b5583) ### Homeserver _No response_ ### Will you send logs? No
1.0
Mentioning a user with emojis in the display name causes Element Android to 'mention' a few times - ### Steps to reproduce 1. Mention a person with emojis in their username 2. Send it ### Outcome #### What did you expect? Mentions once #### What happened instead? It 'mentions' a few times, though in other Matrix clients it appears as one ![Element Android](https://user-images.githubusercontent.com/47475676/150444960-ce373164-6413-47dc-bd0a-6f12ef9ae276.jpg) ![FluffyChat](https://user-images.githubusercontent.com/47475676/150445054-aa6d057d-8e88-4ae1-a52d-8b9dab54187a.jpg) ### Your phone model Mito W1 ### Operating system version Android 9 ### Application version and app store 1.3.15 [40103152] (G-b5583) ### Homeserver _No response_ ### Will you send logs? No
defect
mentioning a user with emojis in the display name causes element android to mention a few times steps to reproduce mention a person with emojis in their username send it outcome what did you expect mentions once what happened instead it mentions a few times though in other matrix clients it appears as one your phone model mito operating system version android application version and app store g homeserver no response will you send logs no
1
79,396
28,147,604,437
IssuesEvent
2023-04-02 17:09:39
scipy/scipy
https://api.github.com/repos/scipy/scipy
opened
BUG: <Please write a comprehensive title after the 'BUG: ' prefix>
defect
### Describe your issue. Issue: 'keep_feasible=True' in the 'scipy.optimize.NonlinearConstraint' requires 'x0' in 'scipy.optimize.minimize' to meet the constraints. This is so unreasonable for users. Besides, sometimes, it also can't converge to the right results, while ''keep_feasible=False' can. However, if I set ''keep_feasible=False', it sometimes converges to the point of minimum objective but violating constraints. Analysis: The reason of the failure is quite obvious, because once we set 'keep_feasible=True' , the optimization process doesn't accept the "infeasible" trials to compute the finte difference to let the process go further. Improvement suggestion: Why can't we make the process go as "keep_feasible=False", so that it can use the "infeasible" trials to converge to the final feasible result, but only accept the "feasible" trials as the valid steps towards the final result. We should also set a number limit to the **consecutive** "infeasible" trials so that it can come back to the last valid step and make it as the final result or continue trials in different directions or step-sizes. ### Reproducing Code Example ```python cons=NonlinearConstraint(fun_con, 0+1e-6, np.inf, keep_feasible=True) #require x0 to fulfil cons res = opt.minimize(obj, x0, #method = 'SLSQP', #'COBYLA'(fail), #'SLSQP'(default) constraints=cons, bounds=bounds, tol=1e-4) ``` ### Error message ```shell Exception has occurred: ValueError `x0` is infeasible with respect to some inequality constraint with `keep_feasible` set to True. ``` ### SciPy/NumPy/Python version and system information ```shell { "Compilers": { "c": { "name": "gcc", "linker": "ld.bfd", "version": "10.3.0", "commands": "cc" }, "cython": { "name": "cython", "linker": "cython", "version": "0.29.33", "commands": "cython" }, "c++": { "name": "gcc", "linker": "ld.bfd", "version": "10.3.0", "commands": "c++" }, "fortran": { "name": "gcc", "linker": "ld.bfd", "version": "10.3.0", "commands": "gfortran" }, "pythran": { "version": "0.12.1", "include directory": "C:\\Users\\runneradmin\\AppData\\Local\\Temp\\pip-build-env-u63ta2f1\\overlay\\Lib\\site-packages\\pythran" } }, "Machine Information": { "host": { "cpu": "x86_64", "family": "x86_64", "endian": "little", "system": "windows" }, "build": { "cpu": "x86_64", "family": "x86_64", "endian": "little", "system": "windows" }, "cross-compiled": false }, "Build Dependencies": { "blas": { "name": "openblas", "found": true, "version": "0.3.18", "detection method": "pkgconfig", "include directory": "c:/opt/openblas/if_32/64/include", "lib directory": "c:/opt/openblas/if_32/64/lib", "openblas configuration": "USE_64BITINT= DYNAMIC_ARCH=1 DYNAMIC_OLDER= NO_CBLAS= NO_LAPACK= NO_LAPACKE= NO_AFFINITY=1 USE_OPENMP= PRESCOTT MAX_THREADS=4", "pc file directory": "c:/opt/openblas/if_32/64/lib/pkgconfig" }, "lapack": { "name": "openblas", "found": true, "version": "0.3.18", "detection method": "pkgconfig", "include directory": "c:/opt/openblas/if_32/64/include", "lib directory": "c:/opt/openblas/if_32/64/lib", "openblas configuration": "USE_64BITINT= DYNAMIC_ARCH=1 DYNAMIC_OLDER= NO_CBLAS= NO_LAPACK= NO_LAPACKE= NO_AFFINITY=1 USE_OPENMP= PRESCOTT MAX_THREADS=4", "pc file directory": "c:/opt/openblas/if_32/64/lib/pkgconfig" } }, "Python Information": { "path": "C:\\Users\\runneradmin\\AppData\\Local\\Temp\\cibw-run-a1px0t3e\\cp39-win_amd64\\build\\venv\\Scripts\\python.exe", "version": "3.9" } } ```
1.0
BUG: <Please write a comprehensive title after the 'BUG: ' prefix> - ### Describe your issue. Issue: 'keep_feasible=True' in the 'scipy.optimize.NonlinearConstraint' requires 'x0' in 'scipy.optimize.minimize' to meet the constraints. This is so unreasonable for users. Besides, sometimes, it also can't converge to the right results, while ''keep_feasible=False' can. However, if I set ''keep_feasible=False', it sometimes converges to the point of minimum objective but violating constraints. Analysis: The reason of the failure is quite obvious, because once we set 'keep_feasible=True' , the optimization process doesn't accept the "infeasible" trials to compute the finte difference to let the process go further. Improvement suggestion: Why can't we make the process go as "keep_feasible=False", so that it can use the "infeasible" trials to converge to the final feasible result, but only accept the "feasible" trials as the valid steps towards the final result. We should also set a number limit to the **consecutive** "infeasible" trials so that it can come back to the last valid step and make it as the final result or continue trials in different directions or step-sizes. ### Reproducing Code Example ```python cons=NonlinearConstraint(fun_con, 0+1e-6, np.inf, keep_feasible=True) #require x0 to fulfil cons res = opt.minimize(obj, x0, #method = 'SLSQP', #'COBYLA'(fail), #'SLSQP'(default) constraints=cons, bounds=bounds, tol=1e-4) ``` ### Error message ```shell Exception has occurred: ValueError `x0` is infeasible with respect to some inequality constraint with `keep_feasible` set to True. ``` ### SciPy/NumPy/Python version and system information ```shell { "Compilers": { "c": { "name": "gcc", "linker": "ld.bfd", "version": "10.3.0", "commands": "cc" }, "cython": { "name": "cython", "linker": "cython", "version": "0.29.33", "commands": "cython" }, "c++": { "name": "gcc", "linker": "ld.bfd", "version": "10.3.0", "commands": "c++" }, "fortran": { "name": "gcc", "linker": "ld.bfd", "version": "10.3.0", "commands": "gfortran" }, "pythran": { "version": "0.12.1", "include directory": "C:\\Users\\runneradmin\\AppData\\Local\\Temp\\pip-build-env-u63ta2f1\\overlay\\Lib\\site-packages\\pythran" } }, "Machine Information": { "host": { "cpu": "x86_64", "family": "x86_64", "endian": "little", "system": "windows" }, "build": { "cpu": "x86_64", "family": "x86_64", "endian": "little", "system": "windows" }, "cross-compiled": false }, "Build Dependencies": { "blas": { "name": "openblas", "found": true, "version": "0.3.18", "detection method": "pkgconfig", "include directory": "c:/opt/openblas/if_32/64/include", "lib directory": "c:/opt/openblas/if_32/64/lib", "openblas configuration": "USE_64BITINT= DYNAMIC_ARCH=1 DYNAMIC_OLDER= NO_CBLAS= NO_LAPACK= NO_LAPACKE= NO_AFFINITY=1 USE_OPENMP= PRESCOTT MAX_THREADS=4", "pc file directory": "c:/opt/openblas/if_32/64/lib/pkgconfig" }, "lapack": { "name": "openblas", "found": true, "version": "0.3.18", "detection method": "pkgconfig", "include directory": "c:/opt/openblas/if_32/64/include", "lib directory": "c:/opt/openblas/if_32/64/lib", "openblas configuration": "USE_64BITINT= DYNAMIC_ARCH=1 DYNAMIC_OLDER= NO_CBLAS= NO_LAPACK= NO_LAPACKE= NO_AFFINITY=1 USE_OPENMP= PRESCOTT MAX_THREADS=4", "pc file directory": "c:/opt/openblas/if_32/64/lib/pkgconfig" } }, "Python Information": { "path": "C:\\Users\\runneradmin\\AppData\\Local\\Temp\\cibw-run-a1px0t3e\\cp39-win_amd64\\build\\venv\\Scripts\\python.exe", "version": "3.9" } } ```
defect
bug describe your issue issue keep feasible true in the scipy optimize nonlinearconstraint requires in scipy optimize minimize to meet the constraints this is so unreasonable for users besides sometimes it also can t converge to the right results while keep feasible false can however if i set keep feasible false it sometimes converges to the point of minimum objective but violating constraints analysis the reason of the failure is quite obvious because once we set keep feasible true the optimization process doesn t accept the infeasible trials to compute the finte difference to let the process go further improvement suggestion why can t we make the process go as keep feasible false so that it can use the infeasible trials to converge to the final feasible result but only accept the feasible trials as the valid steps towards the final result we should also set a number limit to the consecutive infeasible trials so that it can come back to the last valid step and make it as the final result or continue trials in different directions or step sizes reproducing code example python cons nonlinearconstraint fun con np inf keep feasible true require to fulfil cons res opt minimize obj method slsqp cobyla fail slsqp default constraints cons bounds bounds tol error message shell exception has occurred valueerror is infeasible with respect to some inequality constraint with keep feasible set to true scipy numpy python version and system information shell compilers c name gcc linker ld bfd version commands cc cython name cython linker cython version commands cython c name gcc linker ld bfd version commands c fortran name gcc linker ld bfd version commands gfortran pythran version include directory c users runneradmin appdata local temp pip build env overlay lib site packages pythran machine information host cpu family endian little system windows build cpu family endian little system windows cross compiled false build dependencies blas name openblas found true version detection method pkgconfig include directory c opt openblas if include lib directory c opt openblas if lib openblas configuration use dynamic arch dynamic older no cblas no lapack no lapacke no affinity use openmp prescott max threads pc file directory c opt openblas if lib pkgconfig lapack name openblas found true version detection method pkgconfig include directory c opt openblas if include lib directory c opt openblas if lib openblas configuration use dynamic arch dynamic older no cblas no lapack no lapacke no affinity use openmp prescott max threads pc file directory c opt openblas if lib pkgconfig python information path c users runneradmin appdata local temp cibw run win build venv scripts python exe version
1
432,953
12,500,434,559
IssuesEvent
2020-06-01 22:16:25
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
[Android] User wallets not permitted up to a threshold of BAT earner
OS/Android priority/P1
Issue: Currently a lot of users are being pushed in to the user wallet funnel w/o having a lot of earned BAT creating a load for the wallet partner. Fix: Prevent any UI to trigger 'Verify Wallet' until XX BAT has been earned as a virtual grant before user is allowed to verify their user wallet. Please block user wallet release on to Android until fix for this issue is made. We will lead with Android and make fix on Desktop after that. cc @aekeus @LaurenWags @kjozwiak @jamesmudgett
1.0
[Android] User wallets not permitted up to a threshold of BAT earner - Issue: Currently a lot of users are being pushed in to the user wallet funnel w/o having a lot of earned BAT creating a load for the wallet partner. Fix: Prevent any UI to trigger 'Verify Wallet' until XX BAT has been earned as a virtual grant before user is allowed to verify their user wallet. Please block user wallet release on to Android until fix for this issue is made. We will lead with Android and make fix on Desktop after that. cc @aekeus @LaurenWags @kjozwiak @jamesmudgett
non_defect
user wallets not permitted up to a threshold of bat earner issue currently a lot of users are being pushed in to the user wallet funnel w o having a lot of earned bat creating a load for the wallet partner fix prevent any ui to trigger verify wallet until xx bat has been earned as a virtual grant before user is allowed to verify their user wallet please block user wallet release on to android until fix for this issue is made we will lead with android and make fix on desktop after that cc aekeus laurenwags kjozwiak jamesmudgett
0
72,926
13,937,102,554
IssuesEvent
2020-10-22 13:46:04
microsoft/Azure-Kinect-Sensor-SDK
https://api.github.com/repos/microsoft/Azure-Kinect-Sensor-SDK
opened
Unity | Can't display color or depth image in the Body Tracking example
Bug Code Sample Triage Needed
**Description** I am trying to create a script to display Depth and Color images on a `RawImage` object. I added to the `DeviceConfiguration` : `ColorResolution.R720p, ImageFormat.ColorBGRA32` Then assigned it to the `currentFrameData` by modifying `BackgroundData` and `SkeletalTrackingProvider`. But I'm having trouble on retrieving the bytes in a good format. I can't find any help online, and I hope someone here knows how to deal with it. Also I don't understand the "// Repack data and store image data" code. I tried multiple workaround but nothing seems to work. Any help would be appreciated
1.0
Unity | Can't display color or depth image in the Body Tracking example - **Description** I am trying to create a script to display Depth and Color images on a `RawImage` object. I added to the `DeviceConfiguration` : `ColorResolution.R720p, ImageFormat.ColorBGRA32` Then assigned it to the `currentFrameData` by modifying `BackgroundData` and `SkeletalTrackingProvider`. But I'm having trouble on retrieving the bytes in a good format. I can't find any help online, and I hope someone here knows how to deal with it. Also I don't understand the "// Repack data and store image data" code. I tried multiple workaround but nothing seems to work. Any help would be appreciated
non_defect
unity can t display color or depth image in the body tracking example description i am trying to create a script to display depth and color images on a rawimage object i added to the deviceconfiguration colorresolution imageformat then assigned it to the currentframedata by modifying backgrounddata and skeletaltrackingprovider but i m having trouble on retrieving the bytes in a good format i can t find any help online and i hope someone here knows how to deal with it also i don t understand the repack data and store image data code i tried multiple workaround but nothing seems to work any help would be appreciated
0
209,982
23,731,003,802
IssuesEvent
2022-08-31 01:41:41
zettatips/wedding-website
https://api.github.com/repos/zettatips/wedding-website
closed
CVE-2021-35065 (High) detected in glob-parent-3.1.0.tgz - autoclosed
security vulnerability
## CVE-2021-35065 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-3.1.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent directory path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - gulp-4.0.2.tgz (Root Library) - vinyl-fs-3.0.3.tgz - glob-stream-6.1.0.tgz - :x: **glob-parent-3.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/zettatips/wedding-website/commit/a31757c82d2be8eafc3d6f819f065ae51339ae2f">a31757c82d2be8eafc3d6f819f065ae51339ae2f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution: glob-parent - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-35065 (High) detected in glob-parent-3.1.0.tgz - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-3.1.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent directory path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - gulp-4.0.2.tgz (Root Library) - vinyl-fs-3.0.3.tgz - glob-stream-6.1.0.tgz - :x: **glob-parent-3.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/zettatips/wedding-website/commit/a31757c82d2be8eafc3d6f819f065ae51339ae2f">a31757c82d2be8eafc3d6f819f065ae51339ae2f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution: glob-parent - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in glob parent tgz autoclosed cve high severity vulnerability vulnerable library glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy gulp tgz root library vinyl fs tgz glob stream tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with mend
0
65,045
19,060,142,204
IssuesEvent
2021-11-26 06:10:46
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
ztest crashes with default zfs_abd_scatter_min_size
Type: Defect
### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | 11 Kernel Version | not relevant Architecture | x86_64 OpenZFS Version | ded851b2 ### Describe the problem you're observing ztest crashes an awful lot. Most of the crashes, IME, look something like: ``` (gdb) bt #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50 #1 0x00007ffff7737537 in __GI_abort () at abort.c:79 #2 0x00007ffff7ae6923 in vpanic (fmt=0x7ffff7d74540 "Got SIGSEGV at address: 0x%lx\n", adx=adx@entry=0x7fffffffce58) at kernel.c:612 #3 0x00007ffff7ae69bb in panic (fmt=fmt@entry=0x7ffff7d74540 "Got SIGSEGV at address: 0x%lx\n") at kernel.c:621 #4 0x00007ffff7afaeb6 in arc_buf_sigsegv (sig=<optimized out>, si=<optimized out>, unused=<optimized out>) at ../../module/zfs/arc.c:1515 #5 <signal handler called> #6 __memmove_avx_unaligned_erms () at ../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:437 #7 0x00007ffff7aef9be in abd_copy_to_buf_off_cb (buf=<optimized out>, size=size@entry=4096, private=private@entry=0x7fffffffd5a8) at ../../module/zfs/abd.c:828 #8 0x00007ffff7af1044 in abd_iterate_func (private=0x7fffffffd5a8, func=0x7ffff7aef9a0 <abd_copy_to_buf_off_cb>, size=4096, off=<optimized out>, abd=0x7fffd8021c90) at ../../module/zfs/abd.c:805 #9 abd_iterate_func (abd=0x7fffd8021c90, off=<optimized out>, size=<optimized out>, func=0x7ffff7aef9a0 <abd_copy_to_buf_off_cb>, private=0x7fffffffd5a8) at ../../module/zfs/abd.c:780 #10 0x00007ffff7af1278 in abd_copy_to_buf_off (buf=<optimized out>, abd=<optimized out>, off=off@entry=0, size=<optimized out>) at ../../module/zfs/abd.c:842 #11 0x00007ffff7b0239e in abd_copy_to_buf (size=<optimized out>, abd=<optimized out>, buf=<optimized out>) at ../../include/sys/abd.h:159 #12 arc_buf_fill (buf=0x555555b35a90, spa=spa@entry=0x5555556569a0, zb=zb@entry=0x7fffffffd6a0, flags=flags@entry=0) at ../../module/zfs/arc.c:2067 #13 0x00007ffff7b0327d in arc_untransform (buf=<optimized out>, spa=0x5555556569a0, zb=zb@entry=0x7fffffffd6a0, in_place=in_place@entry=B_FALSE) at ../../module/zfs/arc.c:2171 #14 0x00007ffff7b366f6 in dmu_objset_own_impl (ds=ds@entry=0x5555556af3b0, type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, decrypt=decrypt@entry=B_TRUE, osp=osp@entry=0x7fffffffd8c8, tag=<optimized out>) at ../../module/zfs/dmu_objset.c:774 #15 0x00007ffff7b3abed in dmu_objset_own_impl (tag=0x555555573cf0 <__func__.9>, osp=0x7fffffffd8c8, decrypt=B_TRUE, readonly=B_TRUE, type=DMU_OST_ANY, ds=0x5555556af3b0) at ../../module/zfs/dmu_objset.c:757 #16 dmu_objset_own (name=name@entry=0x5555565cbfa0 "ztest/ds_4", type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, decrypt=decrypt@entry=B_TRUE, tag=tag@entry=0x555555573cf0 <__func__.9>, osp=osp@entry=0x7fffffffd8c8) at ../../module/zfs/dmu_objset.c:808 #17 0x0000555555563fcb in ztest_dmu_objset_own (name=name@entry=0x5555565cbfa0 "ztest/ds_4", type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, tag=tag@entry=0x555555573cf0 <__func__.9>, osp=osp@entry=0x7fffffffd8c8, decrypt=B_TRUE) at ztest.c:1602 #18 0x0000555555567d9c in ztest_replay_zil_cb (name=name@entry=0x5555565cbfa0 "ztest/ds_4", arg=arg@entry=0x0) at ztest.c:7255 #19 0x00007ffff7b35d0d in dmu_objset_find_impl (spa=spa@entry=0x5555556569a0, name=name@entry=0x5555565cbfa0 "ztest/ds_4", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2) at ../../module/zfs/dmu_objset.c:2951 #20 0x00007ffff7b35e40 in dmu_objset_find_impl (spa=0x5555556569a0, name=name@entry=0x55555557a960 <ztest_opts> "ztest", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2) at ../../module/zfs/dmu_objset.c:2894 #21 0x00007ffff7b3b303 in dmu_objset_find (name=name@entry=0x55555557a960 <ztest_opts> "ztest", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2) at ../../module/zfs/dmu_objset.c:2967 #22 0x000055555555cdc4 in ztest_run (zs=0x7ffff7ffb738) at ztest.c:7563 #23 main (argc=<optimized out>, argv=<optimized out>) at ztest.c:8062 ``` After a round of bisecting, I ended up at 87c25d56, which I would not have guessed, but here we are. And lo, if you [extend ztest to set zfs_abd_scatter_min_size to 4097](https://github.com/rincebrain/zfs/commit/74828db4139ba4a690a99c348c4ada86830660ee) on x86_64, it goes from crashing practically always to crashing never so far. If we ask valgrind, first it complains a lot about [uninitialized values in the crypto code being read a bunch](https://github.com/openzfs/zfs/blob/6217656da33c0920cb9f213742fd51dd215bc455/module/zfs/dsl_crypt.c#L1197-L1200), but if you zero those, it becomes limited to eventually spitting out: ``` ==129414== Thread 109: ==129414== Conditional jump or move depends on uninitialised value(s) ==129414== at 0x483EEEE: bcmp (vg_replace_strmem.c:1111) ==129414== by 0x48ABA33: abd_cmp_buf_off_cb (abd.c:852) ==129414== by 0x48AD043: abd_iterate_func (abd.c:805) ==129414== by 0x48AD043: abd_iterate_func (abd.c:780) ==129414== by 0x48AD304: abd_cmp_buf_off (abd.c:866) ==129414== by 0x48AD377: abd_cmp_buf (abd.h:165) ==129414== by 0x48AD377: abd_return_buf (abd.c:673) ==129414== by 0x48C0F42: arc_read_done (arc.c:5692) ==129414== by 0x4A606A3: zio_done (zio.c:4835) ==129414== by 0x4A54968: __zio_execute (zio.c:2209) ==129414== by 0x4A54968: zio_execute (zio.c:2122) ==129414== by 0x48A3961: taskq_thread (taskq.c:237) ==129414== by 0x4EFFEA6: start_thread (pthread_create.c:477) ==129414== by 0x5018DEE: clone (clone.S:95) ==129414== Uninitialised value was created by a heap allocation ==129414== at 0x483AEB8: memalign (vg_replace_malloc.c:906) ==129414== by 0x483AFCE: posix_memalign (vg_replace_malloc.c:1070) ==129414== by 0x48AF0BF: umem_alloc_aligned (umem.h:105) ==129414== by 0x48AF0BF: abd_alloc_chunks (abd_os.c:579) ==129414== by 0x48ABF8C: abd_alloc (abd.c:192) ==129414== by 0x48BC838: arc_hdr_alloc_abd (arc.c:3191) ==129414== by 0x48C24DD: arc_read (arc.c:6188) ==129414== by 0x4A4C5BF: zil_read_log_block (zil.c:241) ==129414== by 0x4A4C5BF: zil_parse (zil.c:398) ==129414== by 0x4A4D659: zil_check_log_chain (zil.c:975) ==129414== by 0x48F225A: dmu_objset_find_dp_impl (dmu_objset.c:2725) ==129414== by 0x48F2625: dmu_objset_find_dp_cb (dmu_objset.c:2758) ==129414== by 0x48A3961: taskq_thread (taskq.c:237) ==129414== by 0x4EFFEA6: start_thread (pthread_create.c:477) ==129414== ==129414== ==129414== Process terminating with default action of signal 6 (SIGABRT): dumping core ==129414== at 0x4F56CE1: raise (raise.c:51) ==129414== by 0x4F40536: abort (abort.c:79) ==129414== by 0x48A2922: vpanic (kernel.c:612) ==129414== by 0x48A29BA: panic (kernel.c:621) ==129414== by 0x48B6EB5: arc_buf_sigsegv (arc.c:1515) ==129414== by 0x4F0B13F: ??? (in /usr/lib/x86_64-linux-gnu/libpthread-2.31.so) ==129414== by 0x483F7F2: memmove (vg_replace_strmem.c:1270) ==129414== by 0x48AB9BD: abd_copy_to_buf_off_cb (abd.c:828) ==129414== by 0x48AD043: abd_iterate_func (abd.c:805) ==129414== by 0x48AD043: abd_iterate_func (abd.c:780) ==129414== by 0x48AD277: abd_copy_to_buf_off (abd.c:842) ==129414== by 0x48BE39D: abd_copy_to_buf (abd.h:159) ==129414== by 0x48BE39D: arc_buf_fill (arc.c:2067) ==129414== by 0x48BF27C: arc_untransform (arc.c:2171) ``` Obviously we could just...make ztest do that for now, but that seems problematic, and it's not presently clear to me whether the logical flaw is in the umem implementations of things or elsewhere? (Will continue looking, of course, but.) ### Describe how to reproduce the problem Above. ### Include any warning/errors/backtraces from the system logs Above.
1.0
ztest crashes with default zfs_abd_scatter_min_size - ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | 11 Kernel Version | not relevant Architecture | x86_64 OpenZFS Version | ded851b2 ### Describe the problem you're observing ztest crashes an awful lot. Most of the crashes, IME, look something like: ``` (gdb) bt #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50 #1 0x00007ffff7737537 in __GI_abort () at abort.c:79 #2 0x00007ffff7ae6923 in vpanic (fmt=0x7ffff7d74540 "Got SIGSEGV at address: 0x%lx\n", adx=adx@entry=0x7fffffffce58) at kernel.c:612 #3 0x00007ffff7ae69bb in panic (fmt=fmt@entry=0x7ffff7d74540 "Got SIGSEGV at address: 0x%lx\n") at kernel.c:621 #4 0x00007ffff7afaeb6 in arc_buf_sigsegv (sig=<optimized out>, si=<optimized out>, unused=<optimized out>) at ../../module/zfs/arc.c:1515 #5 <signal handler called> #6 __memmove_avx_unaligned_erms () at ../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:437 #7 0x00007ffff7aef9be in abd_copy_to_buf_off_cb (buf=<optimized out>, size=size@entry=4096, private=private@entry=0x7fffffffd5a8) at ../../module/zfs/abd.c:828 #8 0x00007ffff7af1044 in abd_iterate_func (private=0x7fffffffd5a8, func=0x7ffff7aef9a0 <abd_copy_to_buf_off_cb>, size=4096, off=<optimized out>, abd=0x7fffd8021c90) at ../../module/zfs/abd.c:805 #9 abd_iterate_func (abd=0x7fffd8021c90, off=<optimized out>, size=<optimized out>, func=0x7ffff7aef9a0 <abd_copy_to_buf_off_cb>, private=0x7fffffffd5a8) at ../../module/zfs/abd.c:780 #10 0x00007ffff7af1278 in abd_copy_to_buf_off (buf=<optimized out>, abd=<optimized out>, off=off@entry=0, size=<optimized out>) at ../../module/zfs/abd.c:842 #11 0x00007ffff7b0239e in abd_copy_to_buf (size=<optimized out>, abd=<optimized out>, buf=<optimized out>) at ../../include/sys/abd.h:159 #12 arc_buf_fill (buf=0x555555b35a90, spa=spa@entry=0x5555556569a0, zb=zb@entry=0x7fffffffd6a0, flags=flags@entry=0) at ../../module/zfs/arc.c:2067 #13 0x00007ffff7b0327d in arc_untransform (buf=<optimized out>, spa=0x5555556569a0, zb=zb@entry=0x7fffffffd6a0, in_place=in_place@entry=B_FALSE) at ../../module/zfs/arc.c:2171 #14 0x00007ffff7b366f6 in dmu_objset_own_impl (ds=ds@entry=0x5555556af3b0, type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, decrypt=decrypt@entry=B_TRUE, osp=osp@entry=0x7fffffffd8c8, tag=<optimized out>) at ../../module/zfs/dmu_objset.c:774 #15 0x00007ffff7b3abed in dmu_objset_own_impl (tag=0x555555573cf0 <__func__.9>, osp=0x7fffffffd8c8, decrypt=B_TRUE, readonly=B_TRUE, type=DMU_OST_ANY, ds=0x5555556af3b0) at ../../module/zfs/dmu_objset.c:757 #16 dmu_objset_own (name=name@entry=0x5555565cbfa0 "ztest/ds_4", type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, decrypt=decrypt@entry=B_TRUE, tag=tag@entry=0x555555573cf0 <__func__.9>, osp=osp@entry=0x7fffffffd8c8) at ../../module/zfs/dmu_objset.c:808 #17 0x0000555555563fcb in ztest_dmu_objset_own (name=name@entry=0x5555565cbfa0 "ztest/ds_4", type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, tag=tag@entry=0x555555573cf0 <__func__.9>, osp=osp@entry=0x7fffffffd8c8, decrypt=B_TRUE) at ztest.c:1602 #18 0x0000555555567d9c in ztest_replay_zil_cb (name=name@entry=0x5555565cbfa0 "ztest/ds_4", arg=arg@entry=0x0) at ztest.c:7255 #19 0x00007ffff7b35d0d in dmu_objset_find_impl (spa=spa@entry=0x5555556569a0, name=name@entry=0x5555565cbfa0 "ztest/ds_4", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2) at ../../module/zfs/dmu_objset.c:2951 #20 0x00007ffff7b35e40 in dmu_objset_find_impl (spa=0x5555556569a0, name=name@entry=0x55555557a960 <ztest_opts> "ztest", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2) at ../../module/zfs/dmu_objset.c:2894 #21 0x00007ffff7b3b303 in dmu_objset_find (name=name@entry=0x55555557a960 <ztest_opts> "ztest", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2) at ../../module/zfs/dmu_objset.c:2967 #22 0x000055555555cdc4 in ztest_run (zs=0x7ffff7ffb738) at ztest.c:7563 #23 main (argc=<optimized out>, argv=<optimized out>) at ztest.c:8062 ``` After a round of bisecting, I ended up at 87c25d56, which I would not have guessed, but here we are. And lo, if you [extend ztest to set zfs_abd_scatter_min_size to 4097](https://github.com/rincebrain/zfs/commit/74828db4139ba4a690a99c348c4ada86830660ee) on x86_64, it goes from crashing practically always to crashing never so far. If we ask valgrind, first it complains a lot about [uninitialized values in the crypto code being read a bunch](https://github.com/openzfs/zfs/blob/6217656da33c0920cb9f213742fd51dd215bc455/module/zfs/dsl_crypt.c#L1197-L1200), but if you zero those, it becomes limited to eventually spitting out: ``` ==129414== Thread 109: ==129414== Conditional jump or move depends on uninitialised value(s) ==129414== at 0x483EEEE: bcmp (vg_replace_strmem.c:1111) ==129414== by 0x48ABA33: abd_cmp_buf_off_cb (abd.c:852) ==129414== by 0x48AD043: abd_iterate_func (abd.c:805) ==129414== by 0x48AD043: abd_iterate_func (abd.c:780) ==129414== by 0x48AD304: abd_cmp_buf_off (abd.c:866) ==129414== by 0x48AD377: abd_cmp_buf (abd.h:165) ==129414== by 0x48AD377: abd_return_buf (abd.c:673) ==129414== by 0x48C0F42: arc_read_done (arc.c:5692) ==129414== by 0x4A606A3: zio_done (zio.c:4835) ==129414== by 0x4A54968: __zio_execute (zio.c:2209) ==129414== by 0x4A54968: zio_execute (zio.c:2122) ==129414== by 0x48A3961: taskq_thread (taskq.c:237) ==129414== by 0x4EFFEA6: start_thread (pthread_create.c:477) ==129414== by 0x5018DEE: clone (clone.S:95) ==129414== Uninitialised value was created by a heap allocation ==129414== at 0x483AEB8: memalign (vg_replace_malloc.c:906) ==129414== by 0x483AFCE: posix_memalign (vg_replace_malloc.c:1070) ==129414== by 0x48AF0BF: umem_alloc_aligned (umem.h:105) ==129414== by 0x48AF0BF: abd_alloc_chunks (abd_os.c:579) ==129414== by 0x48ABF8C: abd_alloc (abd.c:192) ==129414== by 0x48BC838: arc_hdr_alloc_abd (arc.c:3191) ==129414== by 0x48C24DD: arc_read (arc.c:6188) ==129414== by 0x4A4C5BF: zil_read_log_block (zil.c:241) ==129414== by 0x4A4C5BF: zil_parse (zil.c:398) ==129414== by 0x4A4D659: zil_check_log_chain (zil.c:975) ==129414== by 0x48F225A: dmu_objset_find_dp_impl (dmu_objset.c:2725) ==129414== by 0x48F2625: dmu_objset_find_dp_cb (dmu_objset.c:2758) ==129414== by 0x48A3961: taskq_thread (taskq.c:237) ==129414== by 0x4EFFEA6: start_thread (pthread_create.c:477) ==129414== ==129414== ==129414== Process terminating with default action of signal 6 (SIGABRT): dumping core ==129414== at 0x4F56CE1: raise (raise.c:51) ==129414== by 0x4F40536: abort (abort.c:79) ==129414== by 0x48A2922: vpanic (kernel.c:612) ==129414== by 0x48A29BA: panic (kernel.c:621) ==129414== by 0x48B6EB5: arc_buf_sigsegv (arc.c:1515) ==129414== by 0x4F0B13F: ??? (in /usr/lib/x86_64-linux-gnu/libpthread-2.31.so) ==129414== by 0x483F7F2: memmove (vg_replace_strmem.c:1270) ==129414== by 0x48AB9BD: abd_copy_to_buf_off_cb (abd.c:828) ==129414== by 0x48AD043: abd_iterate_func (abd.c:805) ==129414== by 0x48AD043: abd_iterate_func (abd.c:780) ==129414== by 0x48AD277: abd_copy_to_buf_off (abd.c:842) ==129414== by 0x48BE39D: abd_copy_to_buf (abd.h:159) ==129414== by 0x48BE39D: arc_buf_fill (arc.c:2067) ==129414== by 0x48BF27C: arc_untransform (arc.c:2171) ``` Obviously we could just...make ztest do that for now, but that seems problematic, and it's not presently clear to me whether the logical flaw is in the umem implementations of things or elsewhere? (Will continue looking, of course, but.) ### Describe how to reproduce the problem Above. ### Include any warning/errors/backtraces from the system logs Above.
defect
ztest crashes with default zfs abd scatter min size system information type version name distribution name debian distribution version kernel version not relevant architecture openzfs version describe the problem you re observing ztest crashes an awful lot most of the crashes ime look something like gdb bt gi raise sig sig entry at sysdeps unix sysv linux raise c in gi abort at abort c in vpanic fmt got sigsegv at address lx n adx adx entry at kernel c in panic fmt fmt entry got sigsegv at address lx n at kernel c in arc buf sigsegv sig si unused at module zfs arc c memmove avx unaligned erms at sysdeps multiarch memmove vec unaligned erms s in abd copy to buf off cb buf size size entry private private entry at module zfs abd c in abd iterate func private func size off abd at module zfs abd c abd iterate func abd off size func private at module zfs abd c in abd copy to buf off buf abd off off entry size at module zfs abd c in abd copy to buf size abd buf at include sys abd h arc buf fill buf spa spa entry zb zb entry flags flags entry at module zfs arc c in arc untransform buf spa zb zb entry in place in place entry b false at module zfs arc c in dmu objset own impl ds ds entry type type entry dmu ost any readonly readonly entry b true decrypt decrypt entry b true osp osp entry tag at module zfs dmu objset c in dmu objset own impl tag osp decrypt b true readonly b true type dmu ost any ds at module zfs dmu objset c dmu objset own name name entry ztest ds type type entry dmu ost any readonly readonly entry b true decrypt decrypt entry b true tag tag entry osp osp entry at module zfs dmu objset c in ztest dmu objset own name name entry ztest ds type type entry dmu ost any readonly readonly entry b true tag tag entry osp osp entry decrypt b true at ztest c in ztest replay zil cb name name entry ztest ds arg arg entry at ztest c in dmu objset find impl spa spa entry name name entry ztest ds func func entry arg arg entry flags flags entry at module zfs dmu objset c in dmu objset find impl spa name name entry ztest func func entry arg arg entry flags flags entry at module zfs dmu objset c in dmu objset find name name entry ztest func func entry arg arg entry flags flags entry at module zfs dmu objset c in ztest run zs at ztest c main argc argv at ztest c after a round of bisecting i ended up at which i would not have guessed but here we are and lo if you extend ztest to set zfs abd scatter min size to on it goes from crashing practically always to crashing never so far if we ask valgrind first it complains a lot about but if you zero those it becomes limited to eventually spitting out thread conditional jump or move depends on uninitialised value s at bcmp vg replace strmem c by abd cmp buf off cb abd c by abd iterate func abd c by abd iterate func abd c by abd cmp buf off abd c by abd cmp buf abd h by abd return buf abd c by arc read done arc c by zio done zio c by zio execute zio c by zio execute zio c by taskq thread taskq c by start thread pthread create c by clone clone s uninitialised value was created by a heap allocation at memalign vg replace malloc c by posix memalign vg replace malloc c by umem alloc aligned umem h by abd alloc chunks abd os c by abd alloc abd c by arc hdr alloc abd arc c by arc read arc c by zil read log block zil c by zil parse zil c by zil check log chain zil c by dmu objset find dp impl dmu objset c by dmu objset find dp cb dmu objset c by taskq thread taskq c by start thread pthread create c process terminating with default action of signal sigabrt dumping core at raise raise c by abort abort c by vpanic kernel c by panic kernel c by arc buf sigsegv arc c by in usr lib linux gnu libpthread so by memmove vg replace strmem c by abd copy to buf off cb abd c by abd iterate func abd c by abd iterate func abd c by abd copy to buf off abd c by abd copy to buf abd h by arc buf fill arc c by arc untransform arc c obviously we could just make ztest do that for now but that seems problematic and it s not presently clear to me whether the logical flaw is in the umem implementations of things or elsewhere will continue looking of course but describe how to reproduce the problem above include any warning errors backtraces from the system logs above
1
49,108
13,185,233,982
IssuesEvent
2020-08-12 20:59:33
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
test ticket for templates (Trac #755)
Incomplete Migration Migrated from Trac booking defect
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/755 , reported by nega and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2014-09-05T22:04:25", "description": "this is the ticket body\n\n{{{\nsome\n formatted\n text\n}}}", "reporter": "nega", "cc": "negapluck@gmail.com", "resolution": "fixed", "_ts": "1409954665676165", "component": "booking", "summary": "test ticket for templates", "priority": "normal", "keywords": "", "time": "2014-09-05T21:57:55", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
test ticket for templates (Trac #755) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/755 , reported by nega and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2014-09-05T22:04:25", "description": "this is the ticket body\n\n{{{\nsome\n formatted\n text\n}}}", "reporter": "nega", "cc": "negapluck@gmail.com", "resolution": "fixed", "_ts": "1409954665676165", "component": "booking", "summary": "test ticket for templates", "priority": "normal", "keywords": "", "time": "2014-09-05T21:57:55", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
test ticket for templates trac migrated from reported by nega and owned by nega json status closed changetime description this is the ticket body n n nsome n formatted n text n reporter nega cc negapluck gmail com resolution fixed ts component booking summary test ticket for templates priority normal keywords time milestone owner nega type defect
1
47,157
13,056,043,323
IssuesEvent
2020-07-30 03:28:57
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
svn version upgrade needed (Trac #73)
Migrated from Trac defect infrastructure
DART nodes need svn version 1.4.x or better for latest and greatest dartboard voodoo. Migrated from https://code.icecube.wisc.edu/ticket/73 ```json { "status": "closed", "changetime": "2007-06-26T16:16:23", "description": "DART nodes need svn version 1.4.x or better for latest\nand greatest dartboard voodoo.\n\n", "reporter": "blaufuss", "cc": "", "resolution": "wont or cant fix", "_ts": "1182874583000000", "component": "infrastructure", "summary": "svn version upgrade needed", "priority": "normal", "keywords": "", "time": "2007-06-25T21:20:12", "milestone": "", "owner": "cgils", "type": "defect" } ```
1.0
svn version upgrade needed (Trac #73) - DART nodes need svn version 1.4.x or better for latest and greatest dartboard voodoo. Migrated from https://code.icecube.wisc.edu/ticket/73 ```json { "status": "closed", "changetime": "2007-06-26T16:16:23", "description": "DART nodes need svn version 1.4.x or better for latest\nand greatest dartboard voodoo.\n\n", "reporter": "blaufuss", "cc": "", "resolution": "wont or cant fix", "_ts": "1182874583000000", "component": "infrastructure", "summary": "svn version upgrade needed", "priority": "normal", "keywords": "", "time": "2007-06-25T21:20:12", "milestone": "", "owner": "cgils", "type": "defect" } ```
defect
svn version upgrade needed trac dart nodes need svn version x or better for latest and greatest dartboard voodoo migrated from json status closed changetime description dart nodes need svn version x or better for latest nand greatest dartboard voodoo n n reporter blaufuss cc resolution wont or cant fix ts component infrastructure summary svn version upgrade needed priority normal keywords time milestone owner cgils type defect
1
15,675
9,593,549,970
IssuesEvent
2019-05-09 11:52:13
molgenis/molgenis
https://api.github.com/repos/molgenis/molgenis
closed
Calls with token plus session change the session's authentication
8.0.0-SNAPSHOT mod:security mod:web team-platform
### How to Reproduce Create a token `blah` for user A. Log in on molgenis as user B. (You get a session cookie for B.) Make any call in the browser with `?molgenis-token=blah` (So with a session cookie for A and a token for B) Now go back to making calls with the session cookie. ### Expected behavior Calls with the session cookie still give results for user A. ### Observed behavior Calls with the session cookie now give results for user B. The session got changed into one for B.
True
Calls with token plus session change the session's authentication - ### How to Reproduce Create a token `blah` for user A. Log in on molgenis as user B. (You get a session cookie for B.) Make any call in the browser with `?molgenis-token=blah` (So with a session cookie for A and a token for B) Now go back to making calls with the session cookie. ### Expected behavior Calls with the session cookie still give results for user A. ### Observed behavior Calls with the session cookie now give results for user B. The session got changed into one for B.
non_defect
calls with token plus session change the session s authentication how to reproduce create a token blah for user a log in on molgenis as user b you get a session cookie for b make any call in the browser with molgenis token blah so with a session cookie for a and a token for b now go back to making calls with the session cookie expected behavior calls with the session cookie still give results for user a observed behavior calls with the session cookie now give results for user b the session got changed into one for b
0
128,223
5,051,062,837
IssuesEvent
2016-12-20 20:41:12
scottbaggett/lexus-int
https://api.github.com/repos/scottbaggett/lexus-int
closed
(Article) Input final copy/images
Content Missing High Priority
I downloaded and put the images into the proper folders. - I already updated the images for crafted for lexus & all-lit-up. Copy needs to be checked though. - Copydeck is inside dropbox in copydeck folder.
1.0
(Article) Input final copy/images - I downloaded and put the images into the proper folders. - I already updated the images for crafted for lexus & all-lit-up. Copy needs to be checked though. - Copydeck is inside dropbox in copydeck folder.
non_defect
article input final copy images i downloaded and put the images into the proper folders i already updated the images for crafted for lexus all lit up copy needs to be checked though copydeck is inside dropbox in copydeck folder
0
38,629
8,950,552,041
IssuesEvent
2019-01-25 11:07:15
svigerske/ipopt-donotuse
https://api.github.com/repos/svigerske/ipopt-donotuse
closed
Release of Ipopt 3.0
Initial Release Ipopt defect highest
Issue created by migration from Trac. Original creator: claird Original creation time: 2005-08-26 22:08:10 Assignee: ipopt-team Version: 3.0 This milestone marks the initial release of Ipopt 3.0.0.
1.0
Release of Ipopt 3.0 - Issue created by migration from Trac. Original creator: claird Original creation time: 2005-08-26 22:08:10 Assignee: ipopt-team Version: 3.0 This milestone marks the initial release of Ipopt 3.0.0.
defect
release of ipopt issue created by migration from trac original creator claird original creation time assignee ipopt team version this milestone marks the initial release of ipopt
1
54,305
13,543,362,155
IssuesEvent
2020-09-16 18:50:22
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
closed
Site-source conversion factor for steam is wrong
Defect NotIDDChange
Issue overview -------------- The default site-source conversion factor for steam in the ABUPS report might be wrong. With all defaults, the value is 0.300 for steam. This seems more like the conversion efficiency of steam. Other sources put the site-source factor at 1.20-1.35. ![image](https://user-images.githubusercontent.com/17144687/92638201-2ac8af00-f297-11ea-9f0c-72316bf74715.png) ### Details Some additional details for this issue (if relevant): - Version of EnergyPlus v9.4 in develop ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
Site-source conversion factor for steam is wrong - Issue overview -------------- The default site-source conversion factor for steam in the ABUPS report might be wrong. With all defaults, the value is 0.300 for steam. This seems more like the conversion efficiency of steam. Other sources put the site-source factor at 1.20-1.35. ![image](https://user-images.githubusercontent.com/17144687/92638201-2ac8af00-f297-11ea-9f0c-72316bf74715.png) ### Details Some additional details for this issue (if relevant): - Version of EnergyPlus v9.4 in develop ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
defect
site source conversion factor for steam is wrong issue overview the default site source conversion factor for steam in the abups report might be wrong with all defaults the value is for steam this seems more like the conversion efficiency of steam other sources put the site source factor at details some additional details for this issue if relevant version of energyplus in develop checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
1
3,639
4,462,331,542
IssuesEvent
2016-08-24 09:31:46
opensheetmusicdisplay/opensheetmusicdisplay
https://api.github.com/repos/opensheetmusicdisplay/opensheetmusicdisplay
closed
Get rid of `dist` in repo
infrastructure
Its annoying. Take my PR #16 for instance. you can't properly browse changed files, everything is covered by this random dist stuff. So I thought about having a npm prepublish hook that builds stuff before publish so it gets included inside the tgz for npm. The only problem here is that npm respects `.gitignore` for packaging, so I have to figure out how to do this.
1.0
Get rid of `dist` in repo - Its annoying. Take my PR #16 for instance. you can't properly browse changed files, everything is covered by this random dist stuff. So I thought about having a npm prepublish hook that builds stuff before publish so it gets included inside the tgz for npm. The only problem here is that npm respects `.gitignore` for packaging, so I have to figure out how to do this.
non_defect
get rid of dist in repo its annoying take my pr for instance you can t properly browse changed files everything is covered by this random dist stuff so i thought about having a npm prepublish hook that builds stuff before publish so it gets included inside the tgz for npm the only problem here is that npm respects gitignore for packaging so i have to figure out how to do this
0
26,097
4,586,712,116
IssuesEvent
2016-09-20 00:11:50
Openki/Openki
https://api.github.com/repos/Openki/Openki
closed
Answers to comments are only removed server-side
defect
if you click the closing-x to a sub-comment / to an answer to a comment, it will not be removed on the client. Server seems to delete the comment correctly
1.0
Answers to comments are only removed server-side - if you click the closing-x to a sub-comment / to an answer to a comment, it will not be removed on the client. Server seems to delete the comment correctly
defect
answers to comments are only removed server side if you click the closing x to a sub comment to an answer to a comment it will not be removed on the client server seems to delete the comment correctly
1
54,711
13,894,597,682
IssuesEvent
2020-10-19 14:52:19
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
"Show rooms with unread messages first" checkbox is buggy
bug defect type:new-room-list
Element version: 1.7.9 (app.element.io) Pressing the "Show rooms with unread messages first" checkbox produces weird but deterministic behaviour. Clicking it does not toggle the checkbox unless pressing more times or reopening the dialog. When pressing multiple times it always shows the wrong state. Video: ![untitled](https://user-images.githubusercontent.com/25297359/96377598-ed551c80-1186-11eb-9056-5399f28625d0.GIF)
1.0
"Show rooms with unread messages first" checkbox is buggy - Element version: 1.7.9 (app.element.io) Pressing the "Show rooms with unread messages first" checkbox produces weird but deterministic behaviour. Clicking it does not toggle the checkbox unless pressing more times or reopening the dialog. When pressing multiple times it always shows the wrong state. Video: ![untitled](https://user-images.githubusercontent.com/25297359/96377598-ed551c80-1186-11eb-9056-5399f28625d0.GIF)
defect
show rooms with unread messages first checkbox is buggy element version app element io pressing the show rooms with unread messages first checkbox produces weird but deterministic behaviour clicking it does not toggle the checkbox unless pressing more times or reopening the dialog when pressing multiple times it always shows the wrong state video
1
60,847
17,023,538,095
IssuesEvent
2021-07-03 02:32:12
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Data browser breaks on many-versioned Relation
Component: website Priority: major Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 4.02pm, Tuesday, 12th January 2010]** Relation #16377 (Downs Link path) has 81 versions (as of date of filing this bug). It also has loads and loads of members. Trying to view the edit history using http://www.openstreetmap.org/browse/relation/16377/history always results in a 500. Out of memory, maybe?
1.0
Data browser breaks on many-versioned Relation - **[Submitted to the original trac issue database at 4.02pm, Tuesday, 12th January 2010]** Relation #16377 (Downs Link path) has 81 versions (as of date of filing this bug). It also has loads and loads of members. Trying to view the edit history using http://www.openstreetmap.org/browse/relation/16377/history always results in a 500. Out of memory, maybe?
defect
data browser breaks on many versioned relation relation downs link path has versions as of date of filing this bug it also has loads and loads of members trying to view the edit history using always results in a out of memory maybe
1
389,592
11,504,281,155
IssuesEvent
2020-02-12 22:55:02
garden-io/garden
https://api.github.com/repos/garden-io/garden
closed
Unexpected error when templating with ternary operator and ||
bug priority:low
## Bug Noticed two issues when using a ternary operator and a `||` in template strings. ### Current Behavior Firstly, the following template string: ```yaml TEST_OUTPUT: "${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}" ``` will fail with: ```console Invalid template string ${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}: Circular reference detected when resolving key (runtime -> runtime.tasks.echo-hello) ``` Secondly, the following template string: ```yaml TEST_OUTPUT: "${runtime.tasks.no-task.outputs.log || 'default'}" ``` fails with: ```console Unable to resolve one or more runtime template values for service 'backend': ${runtime.tasks.no-task.outputs.log} ``` **if** `no-task` doesn't exist. Finally the error isn't very helpful if the `dependencies` field is not set. It's the same as above. ### Expected behavior Should work. ### Reproducible example ```yaml kind: Module name: frontend type: container tasks: - name: echo-hello args: ["echo", "hello"] ... --- kind: Module name: backend type: container services: - name: backend ... env: TEST_OUTPUT: "${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}" # FIXME TEST_OUTPUT: "${runtime.tasks.echo-hello.outputs.log || 'default'}" # And FIXME dependencies: ["echo-hello"] ``` ### Workaround N/A
1.0
Unexpected error when templating with ternary operator and || - ## Bug Noticed two issues when using a ternary operator and a `||` in template strings. ### Current Behavior Firstly, the following template string: ```yaml TEST_OUTPUT: "${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}" ``` will fail with: ```console Invalid template string ${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}: Circular reference detected when resolving key (runtime -> runtime.tasks.echo-hello) ``` Secondly, the following template string: ```yaml TEST_OUTPUT: "${runtime.tasks.no-task.outputs.log || 'default'}" ``` fails with: ```console Unable to resolve one or more runtime template values for service 'backend': ${runtime.tasks.no-task.outputs.log} ``` **if** `no-task` doesn't exist. Finally the error isn't very helpful if the `dependencies` field is not set. It's the same as above. ### Expected behavior Should work. ### Reproducible example ```yaml kind: Module name: frontend type: container tasks: - name: echo-hello args: ["echo", "hello"] ... --- kind: Module name: backend type: container services: - name: backend ... env: TEST_OUTPUT: "${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}" # FIXME TEST_OUTPUT: "${runtime.tasks.echo-hello.outputs.log || 'default'}" # And FIXME dependencies: ["echo-hello"] ``` ### Workaround N/A
non_defect
unexpected error when templating with ternary operator and bug noticed two issues when using a ternary operator and a in template strings current behavior firstly the following template string yaml test output runtime tasks echo hello runtime tasks echo hello outputs log default will fail with console invalid template string runtime tasks echo hello runtime tasks echo hello outputs log default circular reference detected when resolving key runtime runtime tasks echo hello secondly the following template string yaml test output runtime tasks no task outputs log default fails with console unable to resolve one or more runtime template values for service backend runtime tasks no task outputs log if no task doesn t exist finally the error isn t very helpful if the dependencies field is not set it s the same as above expected behavior should work reproducible example yaml kind module name frontend type container tasks name echo hello args kind module name backend type container services name backend env test output runtime tasks echo hello runtime tasks echo hello outputs log default fixme test output runtime tasks echo hello outputs log default and fixme dependencies workaround n a
0
176,076
21,365,696,439
IssuesEvent
2022-04-20 01:11:05
phytomichael/KSA
https://api.github.com/repos/phytomichael/KSA
closed
CVE-2018-11039 (Medium) detected in spring-web-3.1.1.RELEASE.jar - autoclosed
security vulnerability
## CVE-2018-11039 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-3.1.1.RELEASE.jar</b></p></summary> <p>Spring Framework Parent</p> <p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_df7b546f-7ca1-4faf-a7cf-4b2796ca07e7/20190718203425_56332/git_depth_0/KSA/ksa/ksa/ksa-web-root/ksa-web/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/KSA/ksa/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-web-3.1.1.RELEASE.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Framework (versions 5.0.x prior to 5.0.7, versions 4.3.x prior to 4.3.18, and older unsupported versions) allow web applications to change the HTTP request method to any HTTP method (including TRACE) using the HiddenHttpMethodFilter in Spring MVC. If an application has a pre-existing XSS vulnerability, a malicious user (or attacker) can use this filter to escalate to an XST (Cross Site Tracing) attack. <p>Publish Date: 2018-06-25 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11039>CVE-2018-11039</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-11039">https://nvd.nist.gov/vuln/detail/CVE-2018-11039</a></p> <p>Release Date: 2018-06-25</p> <p>Fix Resolution: 5.0.7,4.3.18</p> </p> </details> <p></p>
True
CVE-2018-11039 (Medium) detected in spring-web-3.1.1.RELEASE.jar - autoclosed - ## CVE-2018-11039 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-3.1.1.RELEASE.jar</b></p></summary> <p>Spring Framework Parent</p> <p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_df7b546f-7ca1-4faf-a7cf-4b2796ca07e7/20190718203425_56332/git_depth_0/KSA/ksa/ksa/ksa-web-root/ksa-web/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/KSA/ksa/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-web-3.1.1.RELEASE.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Framework (versions 5.0.x prior to 5.0.7, versions 4.3.x prior to 4.3.18, and older unsupported versions) allow web applications to change the HTTP request method to any HTTP method (including TRACE) using the HiddenHttpMethodFilter in Spring MVC. If an application has a pre-existing XSS vulnerability, a malicious user (or attacker) can use this filter to escalate to an XST (Cross Site Tracing) attack. <p>Publish Date: 2018-06-25 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11039>CVE-2018-11039</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-11039">https://nvd.nist.gov/vuln/detail/CVE-2018-11039</a></p> <p>Release Date: 2018-06-25</p> <p>Fix Resolution: 5.0.7,4.3.18</p> </p> </details> <p></p>
non_defect
cve medium detected in spring web release jar autoclosed cve medium severity vulnerability vulnerable library spring web release jar spring framework parent path to dependency file tmp whitesource archiveextractor git depth ksa ksa ksa ksa web root ksa web pom xml path to vulnerable library root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar repository org springframework spring web release spring web release jar ksa ksa web root ksa web target root web inf lib spring web release jar root repository org springframework spring web release spring web release jar repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar ksa ksa ksa ksa web root ksa web target root web inf lib spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar dependency hierarchy x spring web release jar vulnerable library vulnerability details spring framework versions x prior to versions x prior to and older unsupported versions allow web applications to change the http request method to any http method including trace using the hiddenhttpmethodfilter in spring mvc if an application has a pre existing xss vulnerability a malicious user or attacker can use this filter to escalate to an xst cross site tracing attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
30,221
6,046,530,636
IssuesEvent
2017-06-12 12:23:09
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Dropdown should scroll to selected item when overlay gets opened
defect
Hi, I use dropdown component with bindings. When dropdown is opened for the first time, the scroll is not on selected item, I've to scroll manually to see it. Perhaps I'm missing something ? Thanks
1.0
Dropdown should scroll to selected item when overlay gets opened - Hi, I use dropdown component with bindings. When dropdown is opened for the first time, the scroll is not on selected item, I've to scroll manually to see it. Perhaps I'm missing something ? Thanks
defect
dropdown should scroll to selected item when overlay gets opened hi i use dropdown component with bindings when dropdown is opened for the first time the scroll is not on selected item i ve to scroll manually to see it perhaps i m missing something thanks
1
56,133
14,940,921,747
IssuesEvent
2021-01-25 18:59:12
AeroScripts/QuestieDev
https://api.github.com/repos/AeroScripts/QuestieDev
closed
Auto Turn In Bugs on Argent Dawn Valor Token quests
Questie - Quest Issue Type - Defect
<!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 15M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @BreakBB / TheCrux#1702 (Discord) - Development * @drejjmit / Drejjmit#8241 (Discord) - Testing/Research If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description <!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. --> Auto Turn In gets hung up when turning in multiple AD rep items. It'll fail on the first quest, and just keep repeating it. If you hold shift you can sort of un-stick it. * Corruptor's Scourgestones * Invader's Scourgestones * Minion's Scourgestones * Argent Dawn Valor Token - Item - World of Warcraft || https://classic.wowhead.com/item=12844/argent-dawn-valor-token#reward-from-q:0+1+12 Happy to say this is a conflict with another addon, what can I do to help verify that? I don't think I have Auto Turn In enabled elsewhere, but hard to know for sure. Hey @AeroScripts or @drejjmit, next time I have a bunch of Scourgestones do I just run profiler while I try and turn them in? What can I do to help pinpoint the issue? ## Screenshots <!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. --> ## Questie version <!-- Which version of Questie are you using? You can find it by: - 1. Hovering over the Questie Minimap Icon - 2. looking at your Questie.toc file (open it with any text editor). It looks something like this: "v5.9.0" or "## Version: 5.9.0". --> 6.0.4
1.0
Auto Turn In Bugs on Argent Dawn Valor Token quests - <!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 15M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @BreakBB / TheCrux#1702 (Discord) - Development * @drejjmit / Drejjmit#8241 (Discord) - Testing/Research If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description <!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. --> Auto Turn In gets hung up when turning in multiple AD rep items. It'll fail on the first quest, and just keep repeating it. If you hold shift you can sort of un-stick it. * Corruptor's Scourgestones * Invader's Scourgestones * Minion's Scourgestones * Argent Dawn Valor Token - Item - World of Warcraft || https://classic.wowhead.com/item=12844/argent-dawn-valor-token#reward-from-q:0+1+12 Happy to say this is a conflict with another addon, what can I do to help verify that? I don't think I have Auto Turn In enabled elsewhere, but hard to know for sure. Hey @AeroScripts or @drejjmit, next time I have a bunch of Scourgestones do I just run profiler while I try and turn them in? What can I do to help pinpoint the issue? ## Screenshots <!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. --> ## Questie version <!-- Which version of Questie are you using? You can find it by: - 1. Hovering over the Questie Minimap Icon - 2. looking at your Questie.toc file (open it with any text editor). It looks something like this: "v5.9.0" or "## Version: 5.9.0". --> 6.0.4
defect
auto turn in bugs on argent dawn valor token quests read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is breakbb thecrux discord development drejjmit drejjmit discord testing research if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description auto turn in gets hung up when turning in multiple ad rep items it ll fail on the first quest and just keep repeating it if you hold shift you can sort of un stick it corruptor s scourgestones invader s scourgestones minion s scourgestones argent dawn valor token item world of warcraft happy to say this is a conflict with another addon what can i do to help verify that i don t think i have auto turn in enabled elsewhere but hard to know for sure hey aeroscripts or drejjmit next time i have a bunch of scourgestones do i just run profiler while i try and turn them in what can i do to help pinpoint the issue screenshots questie version which version of questie are you using you can find it by hovering over the questie minimap icon looking at your questie toc file open it with any text editor it looks something like this or version
1
41,017
10,265,853,154
IssuesEvent
2019-08-22 19:55:27
carbon-design-system/ibm-security
https://api.github.com/repos/carbon-design-system/ibm-security
closed
Tearsheet - wrong type size for Cancel button
Defect
@cameroncalder -- - [ ] Helper text on Cancel button `Information will not be saved upon exit` should be label-01 (size 12 not 16) - [ ] Remove period after `exit` in helper text
1.0
Tearsheet - wrong type size for Cancel button - @cameroncalder -- - [ ] Helper text on Cancel button `Information will not be saved upon exit` should be label-01 (size 12 not 16) - [ ] Remove period after `exit` in helper text
defect
tearsheet wrong type size for cancel button cameroncalder helper text on cancel button information will not be saved upon exit should be label size not remove period after exit in helper text
1
71,025
15,169,658,601
IssuesEvent
2021-02-12 21:34:19
gavarasana/ps-state
https://api.github.com/repos/gavarasana/ps-state
closed
CVE-2020-7693 (Medium) detected in sockjs-0.3.19.tgz
security vulnerability
## CVE-2020-7693 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sockjs-0.3.19.tgz</b></p></summary> <p>SockJS-node is a server counterpart of SockJS-client a JavaScript library that provides a WebSocket-like object in the browser. SockJS gives you a coherent, cross-browser, Javascript API which creates a low latency, full duplex, cross-domain communication</p> <p>Library home page: <a href="https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz">https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz</a></p> <p>Path to dependency file: ps-state/package.json</p> <p>Path to vulnerable library: ps-state/node_modules/sockjs/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.1.tgz (Root Library) - webpack-dev-server-3.10.3.tgz - :x: **sockjs-0.3.19.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/gavarasana/ps-state/commit/0ff4648715ecf764b163d4c6a4b8ab91144b9bd0">0ff4648715ecf764b163d4c6a4b8ab91144b9bd0</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Incorrect handling of Upgrade header with the value websocket leads in crashing of containers hosting sockjs apps. This affects the package sockjs before 0.3.20. <p>Publish Date: 2020-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693>CVE-2020-7693</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sockjs/sockjs-node/pull/265">https://github.com/sockjs/sockjs-node/pull/265</a></p> <p>Release Date: 2020-07-09</p> <p>Fix Resolution: sockjs - 0.3.20</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7693 (Medium) detected in sockjs-0.3.19.tgz - ## CVE-2020-7693 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sockjs-0.3.19.tgz</b></p></summary> <p>SockJS-node is a server counterpart of SockJS-client a JavaScript library that provides a WebSocket-like object in the browser. SockJS gives you a coherent, cross-browser, Javascript API which creates a low latency, full duplex, cross-domain communication</p> <p>Library home page: <a href="https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz">https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz</a></p> <p>Path to dependency file: ps-state/package.json</p> <p>Path to vulnerable library: ps-state/node_modules/sockjs/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.1.tgz (Root Library) - webpack-dev-server-3.10.3.tgz - :x: **sockjs-0.3.19.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/gavarasana/ps-state/commit/0ff4648715ecf764b163d4c6a4b8ab91144b9bd0">0ff4648715ecf764b163d4c6a4b8ab91144b9bd0</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Incorrect handling of Upgrade header with the value websocket leads in crashing of containers hosting sockjs apps. This affects the package sockjs before 0.3.20. <p>Publish Date: 2020-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693>CVE-2020-7693</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sockjs/sockjs-node/pull/265">https://github.com/sockjs/sockjs-node/pull/265</a></p> <p>Release Date: 2020-07-09</p> <p>Fix Resolution: sockjs - 0.3.20</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in sockjs tgz cve medium severity vulnerability vulnerable library sockjs tgz sockjs node is a server counterpart of sockjs client a javascript library that provides a websocket like object in the browser sockjs gives you a coherent cross browser javascript api which creates a low latency full duplex cross domain communication library home page a href path to dependency file ps state package json path to vulnerable library ps state node modules sockjs package json dependency hierarchy react scripts tgz root library webpack dev server tgz x sockjs tgz vulnerable library found in head commit a href found in base branch main vulnerability details incorrect handling of upgrade header with the value websocket leads in crashing of containers hosting sockjs apps this affects the package sockjs before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution sockjs step up your open source security game with whitesource
0
30,094
6,021,089,195
IssuesEvent
2017-06-07 17:55:01
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
Cache eviction not fail fast.
Team: Core Type: Defect
When using the following hz xml only on retrieving the cache, I get a validation error. But the XML file could be validated when loaded. ```` <hazelcast xsi:schemaLocation="http://www.hazelcast.com/schema/config http://www.hazelcast.com/schema/config/hazelcast-config-3.8.xsd" xmlns="http://www.hazelcast.com/schema/config" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"> <group> <name>workers</name> </group> <lite-member enabled="true"/> <network> <port port-count="200" auto-increment="true">5701</port> <join> <multicast enabled="false"/> <tcp-ip enabled="true"> <member>10.212.40.101:5701</member> <member>10.212.40.102:5701</member> <member>10.212.40.103:5701</member> <member>10.212.40.104:5701</member> <member>10.212.40.105:5701</member> <member>10.212.40.106:5701</member> <member>10.212.40.107:5701</member> <member>10.212.40.108:5701</member> <member>10.212.40.109:5701</member> <member>10.212.40.110:5701</member> </tcp-ip> </join> <ssl enabled="false"/> </network> <properties> <property name="hazelcast.phone.home.enabled">false</property> </properties> <license-key>...</license-key> <!--MANAGEMENT_CENTER_CONFIG--> <native-memory allocator-type="POOLED" enabled="true"> <size unit="GIGABYTES" value="2" /> <metadata-space-percentage>20</metadata-space-percentage> </native-memory> <cache name="cache"> <eviction size="10000000" max-size-policy="ENTRY_COUNT" eviction-policy="LFU"/> <backup-count>1</backup-count> <async-backup-count>0</async-backup-count> <in-memory-format>NATIVE</in-memory-format> </cache> </hazelcast> `` ``` ERROR 2017-06-07 19:28:12,284 [Thread-5] com.hazelcast.simulator.worker.testcontainer.TestManager: --------------------------- global prepare of LongStringCacheTest FAILED --------------------------- java.lang.IllegalArgumentException: Invalid max-size policy (ENTRY_COUNT) for com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore! Only USED_NATIVE_MEMORY_SIZE, USED_NATIVE_MEMORY_PERCENTAGE, FREE_NATIVE_MEMORY_SIZE, FREE_NATIVE_MEMORY_PERCENTAGE are supported. at com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore.createCacheEvictionChecker(HiDensityNativeMemoryCacheRecordStore.java:121) at com.hazelcast.cache.impl.AbstractCacheRecordStore.<init>(AbstractCacheRecordStore.java:150) at com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore.<init>(HiDensityNativeMemoryCacheRecordStore.java:59) at com.hazelcast.cache.EnterpriseCacheService.newNativeRecordStore(EnterpriseCacheService.java:243) at com.hazelcast.cache.EnterpriseCacheService.createNewRecordStore(EnterpriseCacheService.java:217) at com.hazelcast.cache.impl.CachePartitionSegment.createNew(CachePartitionSegment.java:51) at com.hazelcast.cache.impl.CachePartitionSegment.createNew(CachePartitionSegment.java:37) at com.hazelcast.util.ConcurrencyUtil.getOrPutSynchronized(ConcurrencyUtil.java:73) at com.hazelcast.cache.impl.CachePartitionSegment.getOrCreateRecordStore(CachePartitionSegment.java:67) at com.hazelcast.cache.impl.AbstractCacheService.getOrCreateRecordStore(AbstractCacheService.java:278) ```
1.0
Cache eviction not fail fast. - When using the following hz xml only on retrieving the cache, I get a validation error. But the XML file could be validated when loaded. ```` <hazelcast xsi:schemaLocation="http://www.hazelcast.com/schema/config http://www.hazelcast.com/schema/config/hazelcast-config-3.8.xsd" xmlns="http://www.hazelcast.com/schema/config" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"> <group> <name>workers</name> </group> <lite-member enabled="true"/> <network> <port port-count="200" auto-increment="true">5701</port> <join> <multicast enabled="false"/> <tcp-ip enabled="true"> <member>10.212.40.101:5701</member> <member>10.212.40.102:5701</member> <member>10.212.40.103:5701</member> <member>10.212.40.104:5701</member> <member>10.212.40.105:5701</member> <member>10.212.40.106:5701</member> <member>10.212.40.107:5701</member> <member>10.212.40.108:5701</member> <member>10.212.40.109:5701</member> <member>10.212.40.110:5701</member> </tcp-ip> </join> <ssl enabled="false"/> </network> <properties> <property name="hazelcast.phone.home.enabled">false</property> </properties> <license-key>...</license-key> <!--MANAGEMENT_CENTER_CONFIG--> <native-memory allocator-type="POOLED" enabled="true"> <size unit="GIGABYTES" value="2" /> <metadata-space-percentage>20</metadata-space-percentage> </native-memory> <cache name="cache"> <eviction size="10000000" max-size-policy="ENTRY_COUNT" eviction-policy="LFU"/> <backup-count>1</backup-count> <async-backup-count>0</async-backup-count> <in-memory-format>NATIVE</in-memory-format> </cache> </hazelcast> `` ``` ERROR 2017-06-07 19:28:12,284 [Thread-5] com.hazelcast.simulator.worker.testcontainer.TestManager: --------------------------- global prepare of LongStringCacheTest FAILED --------------------------- java.lang.IllegalArgumentException: Invalid max-size policy (ENTRY_COUNT) for com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore! Only USED_NATIVE_MEMORY_SIZE, USED_NATIVE_MEMORY_PERCENTAGE, FREE_NATIVE_MEMORY_SIZE, FREE_NATIVE_MEMORY_PERCENTAGE are supported. at com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore.createCacheEvictionChecker(HiDensityNativeMemoryCacheRecordStore.java:121) at com.hazelcast.cache.impl.AbstractCacheRecordStore.<init>(AbstractCacheRecordStore.java:150) at com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore.<init>(HiDensityNativeMemoryCacheRecordStore.java:59) at com.hazelcast.cache.EnterpriseCacheService.newNativeRecordStore(EnterpriseCacheService.java:243) at com.hazelcast.cache.EnterpriseCacheService.createNewRecordStore(EnterpriseCacheService.java:217) at com.hazelcast.cache.impl.CachePartitionSegment.createNew(CachePartitionSegment.java:51) at com.hazelcast.cache.impl.CachePartitionSegment.createNew(CachePartitionSegment.java:37) at com.hazelcast.util.ConcurrencyUtil.getOrPutSynchronized(ConcurrencyUtil.java:73) at com.hazelcast.cache.impl.CachePartitionSegment.getOrCreateRecordStore(CachePartitionSegment.java:67) at com.hazelcast.cache.impl.AbstractCacheService.getOrCreateRecordStore(AbstractCacheService.java:278) ```
defect
cache eviction not fail fast when using the following hz xml only on retrieving the cache i get a validation error but the xml file could be validated when loaded hazelcast xsi schemalocation xmlns xmlns xsi workers false native error com hazelcast simulator worker testcontainer testmanager global prepare of longstringcachetest failed java lang illegalargumentexception invalid max size policy entry count for com hazelcast cache hidensity impl nativememory hidensitynativememorycacherecordstore only used native memory size used native memory percentage free native memory size free native memory percentage are supported at com hazelcast cache hidensity impl nativememory hidensitynativememorycacherecordstore createcacheevictionchecker hidensitynativememorycacherecordstore java at com hazelcast cache impl abstractcacherecordstore abstractcacherecordstore java at com hazelcast cache hidensity impl nativememory hidensitynativememorycacherecordstore hidensitynativememorycacherecordstore java at com hazelcast cache enterprisecacheservice newnativerecordstore enterprisecacheservice java at com hazelcast cache enterprisecacheservice createnewrecordstore enterprisecacheservice java at com hazelcast cache impl cachepartitionsegment createnew cachepartitionsegment java at com hazelcast cache impl cachepartitionsegment createnew cachepartitionsegment java at com hazelcast util concurrencyutil getorputsynchronized concurrencyutil java at com hazelcast cache impl cachepartitionsegment getorcreaterecordstore cachepartitionsegment java at com hazelcast cache impl abstractcacheservice getorcreaterecordstore abstractcacheservice java
1
64,739
18,854,942,171
IssuesEvent
2021-11-12 04:11:21
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
TNC does not return optimal parameters
defect scipy.optimize
The truncated Newton (TNC) optimizer `scipy.optimize.minimize(method='TNC')` returns apparently not the _optimal_ parameter vector and function value, but the _last_ one, which often (depending on the dimension of the problem in 10-30% of cases in our applications) has a higher function value. **Expected behavior**: It should return the optimal found parameters and function value. This behavior was not observed for any other scipy optimizer considered. #### Reproducing code example: We encountered this problem in our parameter estimation tool [github.com/icb-dcm/pypesto](https://github.com/icb-dcm/pypesto), which employs an objective function which records a history of all function evaluations and can thus report if the best value in the trace is better than the value reported by the optimizer. See here for a discussion: https://github.com/ICB-DCM/pyPESTO/issues/327. The error can be reproduced in cell 5 of the notebook https://github.com/ICB-DCM/pyPESTO/blob/master/doc/example/rosenbrock.ipynb. The problem only has box constraints, which are not violated. #### Error message: ``` Function values from history and optimizer do not match: 1.3168387678656086, 2.0179911928533514 Parameters obtained from history and optimizer do not match: [0.99553374 0.98995083 0.97544456 0.94768476 0.89550744 0.8019584 0.64185456 0.40180156 0.14794723 0.00844711], [0.9940831 0.98497745 0.96635783 0.93780681 0.87380595 0.74153588 0.53395112 0.25523609 0.03957006 0.00100739] Function values from history and optimizer do not match: 1.6349167203078765, 2.336760892779954 ``` #### Scipy/Numpy/Python version information: ``` 1.4.1 / 1.18.4 / 3.7.4 ```
1.0
TNC does not return optimal parameters - The truncated Newton (TNC) optimizer `scipy.optimize.minimize(method='TNC')` returns apparently not the _optimal_ parameter vector and function value, but the _last_ one, which often (depending on the dimension of the problem in 10-30% of cases in our applications) has a higher function value. **Expected behavior**: It should return the optimal found parameters and function value. This behavior was not observed for any other scipy optimizer considered. #### Reproducing code example: We encountered this problem in our parameter estimation tool [github.com/icb-dcm/pypesto](https://github.com/icb-dcm/pypesto), which employs an objective function which records a history of all function evaluations and can thus report if the best value in the trace is better than the value reported by the optimizer. See here for a discussion: https://github.com/ICB-DCM/pyPESTO/issues/327. The error can be reproduced in cell 5 of the notebook https://github.com/ICB-DCM/pyPESTO/blob/master/doc/example/rosenbrock.ipynb. The problem only has box constraints, which are not violated. #### Error message: ``` Function values from history and optimizer do not match: 1.3168387678656086, 2.0179911928533514 Parameters obtained from history and optimizer do not match: [0.99553374 0.98995083 0.97544456 0.94768476 0.89550744 0.8019584 0.64185456 0.40180156 0.14794723 0.00844711], [0.9940831 0.98497745 0.96635783 0.93780681 0.87380595 0.74153588 0.53395112 0.25523609 0.03957006 0.00100739] Function values from history and optimizer do not match: 1.6349167203078765, 2.336760892779954 ``` #### Scipy/Numpy/Python version information: ``` 1.4.1 / 1.18.4 / 3.7.4 ```
defect
tnc does not return optimal parameters the truncated newton tnc optimizer scipy optimize minimize method tnc returns apparently not the optimal parameter vector and function value but the last one which often depending on the dimension of the problem in of cases in our applications has a higher function value expected behavior it should return the optimal found parameters and function value this behavior was not observed for any other scipy optimizer considered reproducing code example we encountered this problem in our parameter estimation tool which employs an objective function which records a history of all function evaluations and can thus report if the best value in the trace is better than the value reported by the optimizer see here for a discussion the error can be reproduced in cell of the notebook the problem only has box constraints which are not violated error message function values from history and optimizer do not match parameters obtained from history and optimizer do not match function values from history and optimizer do not match scipy numpy python version information
1
536,618
15,711,415,117
IssuesEvent
2021-03-27 07:07:10
nhcarrigan/code-snippets
https://api.github.com/repos/nhcarrigan/code-snippets
closed
[FEAT] - Add external integration notes
⭐ goal: addition 🏁 status: ready for dev 📄 aspect: text 🔒 staff only 🟧 priority: high
# Feature Request ## Is your feature request related to a problem? Please describe: Need to document the external integrations used in repos. <!--A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]--> ## Describe the solution you'd like - LGTM - - Repo is automatically added - - Need to re-run build after initial project setup - Code Climate - - Need to add repo manually - - Enable PR comments and status updates - - Add `**/.spec.ts` to ignore settings (for Angular - because `tests` directory is already ignored) <!--A clear and concise description of what you want to happen.--> ## Additional context <!--Add any other context or screenshots about the feature request here.-->
1.0
[FEAT] - Add external integration notes - # Feature Request ## Is your feature request related to a problem? Please describe: Need to document the external integrations used in repos. <!--A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]--> ## Describe the solution you'd like - LGTM - - Repo is automatically added - - Need to re-run build after initial project setup - Code Climate - - Need to add repo manually - - Enable PR comments and status updates - - Add `**/.spec.ts` to ignore settings (for Angular - because `tests` directory is already ignored) <!--A clear and concise description of what you want to happen.--> ## Additional context <!--Add any other context or screenshots about the feature request here.-->
non_defect
add external integration notes feature request is your feature request related to a problem please describe need to document the external integrations used in repos describe the solution you d like lgtm repo is automatically added need to re run build after initial project setup code climate need to add repo manually enable pr comments and status updates add spec ts to ignore settings for angular because tests directory is already ignored additional context
0
72,583
24,189,595,489
IssuesEvent
2022-09-23 16:09:50
vector-im/element-call
https://api.github.com/repos/vector-im/element-call
closed
Camera is left on after hanging up
T-Defect S-Minor O-Frequent X-Regression Privacy
### Steps to reproduce 1. Have a call 2. Press hangup button 3. Observe that camera light is still on ### Outcome Looks like https://github.com/vector-im/element-call/issues/149 has regressed ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Will you send logs? No
1.0
Camera is left on after hanging up - ### Steps to reproduce 1. Have a call 2. Press hangup button 3. Observe that camera light is still on ### Outcome Looks like https://github.com/vector-im/element-call/issues/149 has regressed ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Will you send logs? No
defect
camera is left on after hanging up steps to reproduce have a call press hangup button observe that camera light is still on outcome looks like has regressed operating system no response browser information no response url for webapp no response will you send logs no
1
103,052
16,601,965,029
IssuesEvent
2021-06-01 20:50:49
samq-ghdemo/SEARCH-NCJIS-nibrs
https://api.github.com/repos/samq-ghdemo/SEARCH-NCJIS-nibrs
opened
CVE-2020-26939 (Medium) detected in bcprov-jdk15on-1.54.jar
security vulnerability
## CVE-2020-26939 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.54.jar</b></p></summary> <p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p> <p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-route/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar</p> <p> Dependency Hierarchy: - :x: **bcprov-jdk15on-1.54.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/SEARCH-NCJIS-nibrs/commit/2643373aa9a184ff4ea81e98caf4009bf2ee8e91">2643373aa9a184ff4ea81e98caf4009bf2ee8e91</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Legion of the Bouncy Castle BC before 1.61 and BC-FJA before 1.0.1.2, attackers can obtain sensitive information about a private exponent because of Observable Differences in Behavior to Error Inputs. This occurs in org.bouncycastle.crypto.encodings.OAEPEncoding. Sending invalid ciphertext that decrypts to a short payload in the OAEP Decoder could result in the throwing of an early exception, potentially leaking some information about the private exponent of the RSA private key performing the encryption. <p>Publish Date: 2020-11-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26939>CVE-2020-26939</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/bcgit/bc-java/wiki/CVE-2020-26939">https://github.com/bcgit/bc-java/wiki/CVE-2020-26939</a></p> <p>Release Date: 2020-10-11</p> <p>Fix Resolution: org.bouncycastle:bcprov-jdk14:1.61,org.bouncycastle:bcprov-ext-debug-jdk15on:1.61,org.bouncycastle:bcprov-debug-jdk15on:1.61,org.bouncycastle:bcprov-ext-jdk15on:1.61,org.bouncycastle:bcprov-jdk15on:1.61</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.bouncycastle","packageName":"bcprov-jdk15on","packageVersion":"1.54","packageFilePaths":["/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-validate-common/pom.xml","/tools/nibrs-fbi-service/pom.xml","/tools/nibrs-common/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-summary-report-common/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-flatfile/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-validation/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.bouncycastle:bcprov-jdk15on:1.54","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.bouncycastle:bcprov-jdk14:1.61,org.bouncycastle:bcprov-ext-debug-jdk15on:1.61,org.bouncycastle:bcprov-debug-jdk15on:1.61,org.bouncycastle:bcprov-ext-jdk15on:1.61,org.bouncycastle:bcprov-jdk15on:1.61"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-26939","vulnerabilityDetails":"In Legion of the Bouncy Castle BC before 1.61 and BC-FJA before 1.0.1.2, attackers can obtain sensitive information about a private exponent because of Observable Differences in Behavior to Error Inputs. This occurs in org.bouncycastle.crypto.encodings.OAEPEncoding. Sending invalid ciphertext that decrypts to a short payload in the OAEP Decoder could result in the throwing of an early exception, potentially leaking some information about the private exponent of the RSA private key performing the encryption.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26939","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-26939 (Medium) detected in bcprov-jdk15on-1.54.jar - ## CVE-2020-26939 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.54.jar</b></p></summary> <p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p> <p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-route/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar</p> <p> Dependency Hierarchy: - :x: **bcprov-jdk15on-1.54.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/SEARCH-NCJIS-nibrs/commit/2643373aa9a184ff4ea81e98caf4009bf2ee8e91">2643373aa9a184ff4ea81e98caf4009bf2ee8e91</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Legion of the Bouncy Castle BC before 1.61 and BC-FJA before 1.0.1.2, attackers can obtain sensitive information about a private exponent because of Observable Differences in Behavior to Error Inputs. This occurs in org.bouncycastle.crypto.encodings.OAEPEncoding. Sending invalid ciphertext that decrypts to a short payload in the OAEP Decoder could result in the throwing of an early exception, potentially leaking some information about the private exponent of the RSA private key performing the encryption. <p>Publish Date: 2020-11-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26939>CVE-2020-26939</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/bcgit/bc-java/wiki/CVE-2020-26939">https://github.com/bcgit/bc-java/wiki/CVE-2020-26939</a></p> <p>Release Date: 2020-10-11</p> <p>Fix Resolution: org.bouncycastle:bcprov-jdk14:1.61,org.bouncycastle:bcprov-ext-debug-jdk15on:1.61,org.bouncycastle:bcprov-debug-jdk15on:1.61,org.bouncycastle:bcprov-ext-jdk15on:1.61,org.bouncycastle:bcprov-jdk15on:1.61</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.bouncycastle","packageName":"bcprov-jdk15on","packageVersion":"1.54","packageFilePaths":["/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-validate-common/pom.xml","/tools/nibrs-fbi-service/pom.xml","/tools/nibrs-common/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-summary-report-common/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-flatfile/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-validation/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.bouncycastle:bcprov-jdk15on:1.54","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.bouncycastle:bcprov-jdk14:1.61,org.bouncycastle:bcprov-ext-debug-jdk15on:1.61,org.bouncycastle:bcprov-debug-jdk15on:1.61,org.bouncycastle:bcprov-ext-jdk15on:1.61,org.bouncycastle:bcprov-jdk15on:1.61"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-26939","vulnerabilityDetails":"In Legion of the Bouncy Castle BC before 1.61 and BC-FJA before 1.0.1.2, attackers can obtain sensitive information about a private exponent because of Observable Differences in Behavior to Error Inputs. This occurs in org.bouncycastle.crypto.encodings.OAEPEncoding. Sending invalid ciphertext that decrypts to a short payload in the OAEP Decoder could result in the throwing of an early exception, potentially leaking some information about the private exponent of the RSA private key performing the encryption.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26939","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
cve medium detected in bcprov jar cve medium severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for jdk to jdk library home page a href path to dependency file search ncjis nibrs tools nibrs route pom xml path to vulnerable library home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar search ncjis nibrs web nibrs web target nibrs web web inf lib bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar dependency hierarchy x bcprov jar vulnerable library found in head commit a href found in base branch master vulnerability details in legion of the bouncy castle bc before and bc fja before attackers can obtain sensitive information about a private exponent because of observable differences in behavior to error inputs this occurs in org bouncycastle crypto encodings oaepencoding sending invalid ciphertext that decrypts to a short payload in the oaep decoder could result in the throwing of an early exception potentially leaking some information about the private exponent of the rsa private key performing the encryption publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org bouncycastle bcprov org bouncycastle bcprov ext debug org bouncycastle bcprov debug org bouncycastle bcprov ext org bouncycastle bcprov check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org bouncycastle bcprov isminimumfixversionavailable true minimumfixversion org bouncycastle bcprov org bouncycastle bcprov ext debug org bouncycastle bcprov debug org bouncycastle bcprov ext org bouncycastle bcprov basebranches vulnerabilityidentifier cve vulnerabilitydetails in legion of the bouncy castle bc before and bc fja before attackers can obtain sensitive information about a private exponent because of observable differences in behavior to error inputs this occurs in org bouncycastle crypto encodings oaepencoding sending invalid ciphertext that decrypts to a short payload in the oaep decoder could result in the throwing of an early exception potentially leaking some information about the private exponent of the rsa private key performing the encryption vulnerabilityurl
0
1,113
2,595,144,187
IssuesEvent
2015-02-20 11:52:31
keyboardsurfer/blinkendroid
https://api.github.com/repos/keyboardsurfer/blinkendroid
opened
kein ticketupdate&Wakelock nach back von playerview
auto-migrated Priority-Critical Type-Defect
``` wenn der server crashed oder beendet wird und man drückt den backbutton, dann wird die loginactivity mit der ticketliste nicht mehr geupdated und auch das wakelock geht nicht mehr ``` ----- Original issue reported on code.google.com by `lischke@gmail.com` on 31 Oct 2010 at 5:19
1.0
kein ticketupdate&Wakelock nach back von playerview - ``` wenn der server crashed oder beendet wird und man drückt den backbutton, dann wird die loginactivity mit der ticketliste nicht mehr geupdated und auch das wakelock geht nicht mehr ``` ----- Original issue reported on code.google.com by `lischke@gmail.com` on 31 Oct 2010 at 5:19
defect
kein ticketupdate wakelock nach back von playerview wenn der server crashed oder beendet wird und man drückt den backbutton dann wird die loginactivity mit der ticketliste nicht mehr geupdated und auch das wakelock geht nicht mehr original issue reported on code google com by lischke gmail com on oct at
1
55,537
6,485,247,796
IssuesEvent
2017-08-19 08:20:53
haskell-tools/haskell-tools
https://api.github.com/repos/haskell-tools/haskell-tools
opened
Flag configuration can prevent refactoring
category:bug origin:stackage-testing package:daemon type:project-handling
In the package NineP, there is flag that should be specified but somehow it seems it isn't. The dependencies look like: ``` if flag(bytestring-in-base) -- bytestring was in base-2.0 and 2.1.1 build-depends: base >= 2.0 && < 2.2, binary >= 0.5.0.2 else -- inbase 1.0 and 3.0 bytestring is a separate package build-depends: base < 2.0 || >= 3, bytestring >= 0.9, binary >= 0.5.0.2 ```
1.0
Flag configuration can prevent refactoring - In the package NineP, there is flag that should be specified but somehow it seems it isn't. The dependencies look like: ``` if flag(bytestring-in-base) -- bytestring was in base-2.0 and 2.1.1 build-depends: base >= 2.0 && < 2.2, binary >= 0.5.0.2 else -- inbase 1.0 and 3.0 bytestring is a separate package build-depends: base < 2.0 || >= 3, bytestring >= 0.9, binary >= 0.5.0.2 ```
non_defect
flag configuration can prevent refactoring in the package ninep there is flag that should be specified but somehow it seems it isn t the dependencies look like if flag bytestring in base bytestring was in base and build depends base else inbase and bytestring is a separate package build depends base bytestring binary
0
37,562
8,434,463,891
IssuesEvent
2018-10-17 10:15:24
scalameta/metals
https://api.github.com/repos/scalameta/metals
closed
Presentation compiler does not pick up new changes in dependent module
defect
* clean compile project with two modules: a and b * open editor in module a * add new `class A` in module a * re-compile module a so that classfiles for `A` are generated * open buffer in module b that depends on a * write `new A`, compilation error I tried doing `Global.askReset()` and it seems to fix this problem. The presentation compiler picked up the change, but I think we should try to figure out when is the best time to run `askReset`.
1.0
Presentation compiler does not pick up new changes in dependent module - * clean compile project with two modules: a and b * open editor in module a * add new `class A` in module a * re-compile module a so that classfiles for `A` are generated * open buffer in module b that depends on a * write `new A`, compilation error I tried doing `Global.askReset()` and it seems to fix this problem. The presentation compiler picked up the change, but I think we should try to figure out when is the best time to run `askReset`.
defect
presentation compiler does not pick up new changes in dependent module clean compile project with two modules a and b open editor in module a add new class a in module a re compile module a so that classfiles for a are generated open buffer in module b that depends on a write new a compilation error i tried doing global askreset and it seems to fix this problem the presentation compiler picked up the change but i think we should try to figure out when is the best time to run askreset
1
9,013
2,615,120,769
IssuesEvent
2015-03-01 05:47:09
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
samples repo has obsolete maven repo inside it with stall buzz library
auto-migrated Component-Google-APIs Priority-Medium Type-Defect
``` The mavenrepo tree should be pruned out of google-api-java-client-samples. ``` Original issue reported on code.google.com by `ai...@google.com` on 1 Jun 2011 at 9:18
1.0
samples repo has obsolete maven repo inside it with stall buzz library - ``` The mavenrepo tree should be pruned out of google-api-java-client-samples. ``` Original issue reported on code.google.com by `ai...@google.com` on 1 Jun 2011 at 9:18
defect
samples repo has obsolete maven repo inside it with stall buzz library the mavenrepo tree should be pruned out of google api java client samples original issue reported on code google com by ai google com on jun at
1
39,422
9,449,926,318
IssuesEvent
2019-04-16 04:10:57
supertuxkart/stk-code
https://api.github.com/repos/supertuxkart/stk-code
closed
Minor issues in Black Forest
C: 3D Modelling P4: minor R: fixed T: defect
A few very minor issues - [x] Wrong zipper arrow direction in reverse. Apply to the reverse-only jump pad and to the texture in the bottom of the road leading to the village on the hill. ~~**Fixed for the jump pad in SVN but NOT for the zipper in the bottom of the road from the hill**~~ **EDIT: now fixed there too** ![bf_issue1](https://user-images.githubusercontent.com/25536748/53312039-b34d4d80-38b3-11e9-99a8-53f594e35634.png) - [x] The houses of the village are not solid for physics, so you can go through their walls and see grass... - [ ] Texture transitions are notoriously difficult, but ideally when the road split into two the "border" part of the texture should not go in the middle of the side road. ![bf_issue2](https://user-images.githubusercontent.com/25536748/53312038-b34d4d80-38b3-11e9-82ce-553d2565cc79.png) - [x] The names on the placards are not there. I think this is a texturing issue. See the screenshot above and look at the signs. - [x] See the picture below. ![bf_issue3](https://user-images.githubusercontent.com/25536748/53312037-b34d4d80-38b3-11e9-86a9-5dbbfda37ab7.png) - [x] At the end of the monastery path on the right, the tree with the light-colored trunk has a buggy hitbox which extends well beyond the visible trunk. - [x] There is a tree floating in the air near the monastery path (drive in reverse to spot it, just right of the "stairs")
1.0
Minor issues in Black Forest - A few very minor issues - [x] Wrong zipper arrow direction in reverse. Apply to the reverse-only jump pad and to the texture in the bottom of the road leading to the village on the hill. ~~**Fixed for the jump pad in SVN but NOT for the zipper in the bottom of the road from the hill**~~ **EDIT: now fixed there too** ![bf_issue1](https://user-images.githubusercontent.com/25536748/53312039-b34d4d80-38b3-11e9-99a8-53f594e35634.png) - [x] The houses of the village are not solid for physics, so you can go through their walls and see grass... - [ ] Texture transitions are notoriously difficult, but ideally when the road split into two the "border" part of the texture should not go in the middle of the side road. ![bf_issue2](https://user-images.githubusercontent.com/25536748/53312038-b34d4d80-38b3-11e9-82ce-553d2565cc79.png) - [x] The names on the placards are not there. I think this is a texturing issue. See the screenshot above and look at the signs. - [x] See the picture below. ![bf_issue3](https://user-images.githubusercontent.com/25536748/53312037-b34d4d80-38b3-11e9-86a9-5dbbfda37ab7.png) - [x] At the end of the monastery path on the right, the tree with the light-colored trunk has a buggy hitbox which extends well beyond the visible trunk. - [x] There is a tree floating in the air near the monastery path (drive in reverse to spot it, just right of the "stairs")
defect
minor issues in black forest a few very minor issues wrong zipper arrow direction in reverse apply to the reverse only jump pad and to the texture in the bottom of the road leading to the village on the hill fixed for the jump pad in svn but not for the zipper in the bottom of the road from the hill edit now fixed there too the houses of the village are not solid for physics so you can go through their walls and see grass texture transitions are notoriously difficult but ideally when the road split into two the border part of the texture should not go in the middle of the side road the names on the placards are not there i think this is a texturing issue see the screenshot above and look at the signs see the picture below at the end of the monastery path on the right the tree with the light colored trunk has a buggy hitbox which extends well beyond the visible trunk there is a tree floating in the air near the monastery path drive in reverse to spot it just right of the stairs
1
6,435
2,610,243,111
IssuesEvent
2015-02-26 19:17:20
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
台州割包皮包茎手术哪家医院专业
auto-migrated Priority-Medium Type-Defect
``` 台州割包皮包茎手术哪家医院专业【台州五洲生殖医院】24小 时健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院� ��址:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘� ��104、108、118、198及椒江一金清公交车直达枫南小区,乘坐107 、105、109、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 1:34
1.0
台州割包皮包茎手术哪家医院专业 - ``` 台州割包皮包茎手术哪家医院专业【台州五洲生殖医院】24小 时健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院� ��址:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘� ��104、108、118、198及椒江一金清公交车直达枫南小区,乘坐107 、105、109、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 1:34
defect
台州割包皮包茎手术哪家医院专业 台州割包皮包茎手术哪家医院专业【台州五洲生殖医院】 时健康咨询热线 微信号tzwzszyy 医院� ��址 (枫南大转盘旁)乘车线路 乘� �� 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
1
47,709
13,066,127,238
IssuesEvent
2020-07-30 21:02:57
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade (Trac #1030)
Migrated from Trac combo core defect
When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade. An example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3 ```text In [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy Out[43]: 9226.722570765558 In [44]: dataclasses.get_most_energetic_cascade(tree).energy Out[44]: 16.27508624107153 ``` This issue occured in icerec trunk, r133025. Migrated from https://code.icecube.wisc.edu/ticket/1030 ```json { "status": "closed", "changetime": "2015-08-10T20:31:17", "description": "When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.\n\nAn example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3\n\n\n{{{\nIn [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy\nOut[43]: 9226.722570765558\n\nIn [44]: dataclasses.get_most_energetic_cascade(tree).energy\nOut[44]: 16.27508624107153\n}}}\n\nThis issue occured in icerec trunk, r133025.\n", "reporter": "hbretz", "cc": "david.schultz@icecube.wisc.edu", "resolution": "fixed", "_ts": "1439238677802127", "component": "combo core", "summary": "dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade", "priority": "blocker", "keywords": "", "time": "2015-06-24T15:41:58", "milestone": "", "owner": "olivas", "type": "defect" } ```
1.0
dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade (Trac #1030) - When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade. An example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3 ```text In [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy Out[43]: 9226.722570765558 In [44]: dataclasses.get_most_energetic_cascade(tree).energy Out[44]: 16.27508624107153 ``` This issue occured in icerec trunk, r133025. Migrated from https://code.icecube.wisc.edu/ticket/1030 ```json { "status": "closed", "changetime": "2015-08-10T20:31:17", "description": "When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.\n\nAn example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3\n\n\n{{{\nIn [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy\nOut[43]: 9226.722570765558\n\nIn [44]: dataclasses.get_most_energetic_cascade(tree).energy\nOut[44]: 16.27508624107153\n}}}\n\nThis issue occured in icerec trunk, r133025.\n", "reporter": "hbretz", "cc": "david.schultz@icecube.wisc.edu", "resolution": "fixed", "_ts": "1439238677802127", "component": "combo core", "summary": "dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade", "priority": "blocker", "keywords": "", "time": "2015-06-24T15:41:58", "milestone": "", "owner": "olivas", "type": "defect" } ```
defect
dataclasses get most energetic cascade returns cascade with lower energy than dataclasses get most energetic inice cascade trac when using dataclasses get most energetic cascade it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses get most energetic inice cascade an example file where this happens e g for the first event can be found at data user hbretz cascades vs inice cascades text in dataclasses get most energetic inice cascade tree energy out in dataclasses get most energetic cascade tree energy out this issue occured in icerec trunk migrated from json status closed changetime description when using dataclasses get most energetic cascade it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses get most energetic inice cascade n nan example file where this happens e g for the first event can be found at data user hbretz cascades vs inice cascades n n n nin dataclasses get most energetic inice cascade tree energy nout n nin dataclasses get most energetic cascade tree energy nout n n nthis issue occured in icerec trunk n reporter hbretz cc david schultz icecube wisc edu resolution fixed ts component combo core summary dataclasses get most energetic cascade returns cascade with lower energy than dataclasses get most energetic inice cascade priority blocker keywords time milestone owner olivas type defect
1
9,674
2,615,165,107
IssuesEvent
2015-03-01 06:45:15
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
Non accurate progression pourcentage
auto-migrated Priority-Triage Type-Defect
``` 0. What version of Reaver are you using? Reaver v1.4 1. What operating system are you using? Backtrack 5r3 - Linux 3.2.6 2. Is your wireless card in monitor mode? Chipset: Atheros AR927 Driver: ath9k Monitor mode: YES 3. What is the signal strength of the Access Point you are trying to crack? According to airodump-ng : -72 4. What is the manufacturer and model # of the device you are trying to crack? Sagem F@st 3304 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b 7C:03:4C:49:CB:FA -e SAGEM_CBF9 -c 6 -N -S -A -v (I use aireplay-ng for the association process and manualy set the channel using:) iwconfig mon0 channel 6 aireplay-ng -1 5000 -a 7C:03:4C:49:CB:FA -h 54:e6:fc:xx:xx:xx -e SAGEM_CBF9 mon0 6. Please describe what you think the issue is. The progression pourcentage is incorrect 39.94% 39.98% 90.93% 90.95% (see the output below) 7. Paste the output from Reaver below. Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [?] Restore previous session for 7C:03:4C:49:CB:FA? [n/Y] y [+] Restored previous session [+] Waiting for beacon from 7C:03:4C:49:CB:FA [+] Associated with 7C:03:4C:49:CB:FA (ESSID: SAGEM_CBF9) [+] 39.35% complete @ 2013-01-08 14:58:46 (2 seconds/pin) [+] Trying pin 43225675 [+] Trying pin 43235674 [+] Trying pin 43245673 [+] Trying pin 43255672 [+] Trying pin 43265671 [+] 39.39% complete @ 2013-01-08 14:58:58 (2 seconds/pin) [+] Trying pin 43275670 [+] Trying pin 43285679 [+] Trying pin 43295678 [+] Trying pin 43305674 [+] Trying pin 43315673 [+] 39.44% complete @ 2013-01-08 14:59:10 (2 seconds/pin) [+] Trying pin 43325672 [+] Trying pin 43335671 [+] Trying pin 43345670 [+] Trying pin 43355679 [+] Trying pin 43365678 [+] 39.48% complete @ 2013-01-08 14:59:21 (2 seconds/pin) [+] Trying pin 43375677 [+] Trying pin 43385676 [+] Trying pin 43395675 [+] Trying pin 43405671 [+] Trying pin 43415670 [+] 39.53% complete @ 2013-01-08 14:59:33 (2 seconds/pin) [+] Trying pin 43425679 [+] Trying pin 43435678 [+] Trying pin 43445677 [+] Trying pin 43455676 [+] Trying pin 43465675 [+] 39.57% complete @ 2013-01-08 14:59:44 (2 seconds/pin) [+] Trying pin 43475674 [+] Trying pin 43485673 [+] Trying pin 43495672 [+] Trying pin 43505678 [+] Trying pin 43515677 [+] 39.62% complete @ 2013-01-08 14:59:56 (2 seconds/pin) [+] Trying pin 43525676 [+] Trying pin 43535675 [+] Trying pin 43545674 [+] Trying pin 43555673 [+] Trying pin 43565672 [+] 39.66% complete @ 2013-01-08 15:00:09 (2 seconds/pin) [+] Trying pin 43575671 [+] Trying pin 43585670 [+] Trying pin 43595679 [+] Trying pin 43605675 [+] Trying pin 43615674 [+] 39.71% complete @ 2013-01-08 15:00:20 (2 seconds/pin) [+] Trying pin 43625673 [+] Trying pin 43635672 [+] Trying pin 43645671 [+] Trying pin 43655670 [+] Trying pin 43665679 [+] 39.75% complete @ 2013-01-08 15:00:32 (2 seconds/pin) [+] Trying pin 43675678 [+] Trying pin 43685677 [+] Trying pin 43695676 [+] Trying pin 43705672 [+] Trying pin 43715671 [+] 39.80% complete @ 2013-01-08 15:00:43 (2 seconds/pin) [+] Trying pin 43725670 [+] Trying pin 43735679 [+] Trying pin 43745678 [+] Trying pin 43755677 [+] Trying pin 43765676 [+] 39.85% complete @ 2013-01-08 15:00:55 (2 seconds/pin) [+] Trying pin 43775675 [+] Trying pin 43785674 [+] Trying pin 43795673 [+] Trying pin 43805679 [+] Trying pin 43815678 [+] 39.89% complete @ 2013-01-08 15:01:06 (2 seconds/pin) [+] Trying pin 43825677 [+] Trying pin 43835676 [+] Trying pin 43845675 [+] Trying pin 43855674 [+] Trying pin 43865673 [+] 39.94% complete @ 2013-01-08 15:01:18 (2 seconds/pin) [+] Trying pin 43875672 [+] Trying pin 43885671 [+] Trying pin 43895670 [+] Trying pin 43905676 [+] Trying pin 43915675 [+] 39.98% complete @ 2013-01-08 15:01:30 (2 seconds/pin) [+] Trying pin 43925674 [+] Trying pin 43935673 [+] Trying pin 43945672 [+] Trying pin 43940004 [+] Trying pin 43940004 [+] 90.93% complete @ 2013-01-08 15:01:45 (2 seconds/pin) [+] Trying pin 43942220 [+] Trying pin 43942220 [+] Trying pin 43943333 [+] Trying pin 43943333 [+] Trying pin 43944446 [+] 90.95% complete @ 2013-01-08 15:02:06 (2 seconds/pin) [+] Trying pin 43944446 [+] Trying pin 43945559 [+] Trying pin 43945559 [+] Trying pin 43946662 [+] Trying pin 43946662 [+] 90.97% complete @ 2013-01-08 15:02:25 (2 seconds/pin) [+] Trying pin 43947775 [+] Trying pin 43947775 [+] Trying pin 43948888 [+] Trying pin 43948888 [+] Trying pin 43949991 [+] 90.99% complete @ 2013-01-08 15:02:46 (2 seconds/pin) [+] Trying pin 43949991 [+] Trying pin 43940011 [+] Trying pin 43940011 [+] Trying pin 43940028 [+] Trying pin 43940028 [+] 91.02% complete @ 2013-01-08 15:03:05 (2 seconds/pin) [+] Trying pin 43940035 [+] Trying pin 43940035 [+] Trying pin 43940042 [+] Trying pin 43940042 [+] Trying pin 43940059 [+] 91.04% complete @ 2013-01-08 15:03:26 (2 seconds/pin) [+] Trying pin 43940059 [+] Trying pin 43940066 [+] Trying pin 43940066 [+] Trying pin 43940073 [+] Trying pin 43940073 [+] 91.06% complete @ 2013-01-08 15:03:45 (2 seconds/pin) [+] Trying pin 43940080 [+] Trying pin 43940080 [+] Trying pin 43940097 [+] Trying pin 43940097 [+] Trying pin 43940103 [+] 91.08% complete @ 2013-01-08 15:04:06 (2 seconds/pin) [+] Trying pin 43940103 [+] Trying pin 43940110 [+] Trying pin 43940110 [+] Trying pin 43940127 [+] Trying pin 43940127 [+] 91.11% complete @ 2013-01-08 15:04:25 (2 seconds/pin) [+] Trying pin 43940134 [+] Trying pin 43940134 [+] Trying pin 43940141 [+] Trying pin 43940141 [+] Trying pin 43940158 [+] 91.13% complete @ 2013-01-08 15:04:46 (3 seconds/pin) [+] Trying pin 43940158 [+] Trying pin 439401 ``` Original issue reported on code.google.com by `i.boud...@gmail.com` on 8 Jan 2013 at 8:35
1.0
Non accurate progression pourcentage - ``` 0. What version of Reaver are you using? Reaver v1.4 1. What operating system are you using? Backtrack 5r3 - Linux 3.2.6 2. Is your wireless card in monitor mode? Chipset: Atheros AR927 Driver: ath9k Monitor mode: YES 3. What is the signal strength of the Access Point you are trying to crack? According to airodump-ng : -72 4. What is the manufacturer and model # of the device you are trying to crack? Sagem F@st 3304 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b 7C:03:4C:49:CB:FA -e SAGEM_CBF9 -c 6 -N -S -A -v (I use aireplay-ng for the association process and manualy set the channel using:) iwconfig mon0 channel 6 aireplay-ng -1 5000 -a 7C:03:4C:49:CB:FA -h 54:e6:fc:xx:xx:xx -e SAGEM_CBF9 mon0 6. Please describe what you think the issue is. The progression pourcentage is incorrect 39.94% 39.98% 90.93% 90.95% (see the output below) 7. Paste the output from Reaver below. Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [?] Restore previous session for 7C:03:4C:49:CB:FA? [n/Y] y [+] Restored previous session [+] Waiting for beacon from 7C:03:4C:49:CB:FA [+] Associated with 7C:03:4C:49:CB:FA (ESSID: SAGEM_CBF9) [+] 39.35% complete @ 2013-01-08 14:58:46 (2 seconds/pin) [+] Trying pin 43225675 [+] Trying pin 43235674 [+] Trying pin 43245673 [+] Trying pin 43255672 [+] Trying pin 43265671 [+] 39.39% complete @ 2013-01-08 14:58:58 (2 seconds/pin) [+] Trying pin 43275670 [+] Trying pin 43285679 [+] Trying pin 43295678 [+] Trying pin 43305674 [+] Trying pin 43315673 [+] 39.44% complete @ 2013-01-08 14:59:10 (2 seconds/pin) [+] Trying pin 43325672 [+] Trying pin 43335671 [+] Trying pin 43345670 [+] Trying pin 43355679 [+] Trying pin 43365678 [+] 39.48% complete @ 2013-01-08 14:59:21 (2 seconds/pin) [+] Trying pin 43375677 [+] Trying pin 43385676 [+] Trying pin 43395675 [+] Trying pin 43405671 [+] Trying pin 43415670 [+] 39.53% complete @ 2013-01-08 14:59:33 (2 seconds/pin) [+] Trying pin 43425679 [+] Trying pin 43435678 [+] Trying pin 43445677 [+] Trying pin 43455676 [+] Trying pin 43465675 [+] 39.57% complete @ 2013-01-08 14:59:44 (2 seconds/pin) [+] Trying pin 43475674 [+] Trying pin 43485673 [+] Trying pin 43495672 [+] Trying pin 43505678 [+] Trying pin 43515677 [+] 39.62% complete @ 2013-01-08 14:59:56 (2 seconds/pin) [+] Trying pin 43525676 [+] Trying pin 43535675 [+] Trying pin 43545674 [+] Trying pin 43555673 [+] Trying pin 43565672 [+] 39.66% complete @ 2013-01-08 15:00:09 (2 seconds/pin) [+] Trying pin 43575671 [+] Trying pin 43585670 [+] Trying pin 43595679 [+] Trying pin 43605675 [+] Trying pin 43615674 [+] 39.71% complete @ 2013-01-08 15:00:20 (2 seconds/pin) [+] Trying pin 43625673 [+] Trying pin 43635672 [+] Trying pin 43645671 [+] Trying pin 43655670 [+] Trying pin 43665679 [+] 39.75% complete @ 2013-01-08 15:00:32 (2 seconds/pin) [+] Trying pin 43675678 [+] Trying pin 43685677 [+] Trying pin 43695676 [+] Trying pin 43705672 [+] Trying pin 43715671 [+] 39.80% complete @ 2013-01-08 15:00:43 (2 seconds/pin) [+] Trying pin 43725670 [+] Trying pin 43735679 [+] Trying pin 43745678 [+] Trying pin 43755677 [+] Trying pin 43765676 [+] 39.85% complete @ 2013-01-08 15:00:55 (2 seconds/pin) [+] Trying pin 43775675 [+] Trying pin 43785674 [+] Trying pin 43795673 [+] Trying pin 43805679 [+] Trying pin 43815678 [+] 39.89% complete @ 2013-01-08 15:01:06 (2 seconds/pin) [+] Trying pin 43825677 [+] Trying pin 43835676 [+] Trying pin 43845675 [+] Trying pin 43855674 [+] Trying pin 43865673 [+] 39.94% complete @ 2013-01-08 15:01:18 (2 seconds/pin) [+] Trying pin 43875672 [+] Trying pin 43885671 [+] Trying pin 43895670 [+] Trying pin 43905676 [+] Trying pin 43915675 [+] 39.98% complete @ 2013-01-08 15:01:30 (2 seconds/pin) [+] Trying pin 43925674 [+] Trying pin 43935673 [+] Trying pin 43945672 [+] Trying pin 43940004 [+] Trying pin 43940004 [+] 90.93% complete @ 2013-01-08 15:01:45 (2 seconds/pin) [+] Trying pin 43942220 [+] Trying pin 43942220 [+] Trying pin 43943333 [+] Trying pin 43943333 [+] Trying pin 43944446 [+] 90.95% complete @ 2013-01-08 15:02:06 (2 seconds/pin) [+] Trying pin 43944446 [+] Trying pin 43945559 [+] Trying pin 43945559 [+] Trying pin 43946662 [+] Trying pin 43946662 [+] 90.97% complete @ 2013-01-08 15:02:25 (2 seconds/pin) [+] Trying pin 43947775 [+] Trying pin 43947775 [+] Trying pin 43948888 [+] Trying pin 43948888 [+] Trying pin 43949991 [+] 90.99% complete @ 2013-01-08 15:02:46 (2 seconds/pin) [+] Trying pin 43949991 [+] Trying pin 43940011 [+] Trying pin 43940011 [+] Trying pin 43940028 [+] Trying pin 43940028 [+] 91.02% complete @ 2013-01-08 15:03:05 (2 seconds/pin) [+] Trying pin 43940035 [+] Trying pin 43940035 [+] Trying pin 43940042 [+] Trying pin 43940042 [+] Trying pin 43940059 [+] 91.04% complete @ 2013-01-08 15:03:26 (2 seconds/pin) [+] Trying pin 43940059 [+] Trying pin 43940066 [+] Trying pin 43940066 [+] Trying pin 43940073 [+] Trying pin 43940073 [+] 91.06% complete @ 2013-01-08 15:03:45 (2 seconds/pin) [+] Trying pin 43940080 [+] Trying pin 43940080 [+] Trying pin 43940097 [+] Trying pin 43940097 [+] Trying pin 43940103 [+] 91.08% complete @ 2013-01-08 15:04:06 (2 seconds/pin) [+] Trying pin 43940103 [+] Trying pin 43940110 [+] Trying pin 43940110 [+] Trying pin 43940127 [+] Trying pin 43940127 [+] 91.11% complete @ 2013-01-08 15:04:25 (2 seconds/pin) [+] Trying pin 43940134 [+] Trying pin 43940134 [+] Trying pin 43940141 [+] Trying pin 43940141 [+] Trying pin 43940158 [+] 91.13% complete @ 2013-01-08 15:04:46 (3 seconds/pin) [+] Trying pin 43940158 [+] Trying pin 439401 ``` Original issue reported on code.google.com by `i.boud...@gmail.com` on 8 Jan 2013 at 8:35
defect
non accurate progression pourcentage what version of reaver are you using reaver what operating system are you using backtrack linux is your wireless card in monitor mode chipset atheros driver monitor mode yes what is the signal strength of the access point you are trying to crack according to airodump ng what is the manufacturer and model of the device you are trying to crack sagem f st what is the entire command line string you are supplying to reaver reaver i b cb fa e sagem c n s a v i use aireplay ng for the association process and manualy set the channel using iwconfig channel aireplay ng a cb fa h fc xx xx xx e sagem please describe what you think the issue is the progression pourcentage is incorrect see the output below paste the output from reaver below reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner restore previous session for cb fa y restored previous session waiting for beacon from cb fa associated with cb fa essid sagem complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin original issue reported on code google com by i boud gmail com on jan at
1
82,054
10,219,374,311
IssuesEvent
2019-08-15 18:24:59
unlock-protocol/unlock
https://api.github.com/repos/unlock-protocol/unlock
closed
We Need a Consistent Grid
design
## Description Currently different pages use different grids and rules, which don't follow the design and creating unsightly inconsistencies. Let's discuss and figure out a way forward. A consistent grid will help us look better but also make dev easier. ![image](https://user-images.githubusercontent.com/1377389/55981621-1ed15b80-5c65-11e9-9471-9ecb276c3c88.png)
1.0
We Need a Consistent Grid - ## Description Currently different pages use different grids and rules, which don't follow the design and creating unsightly inconsistencies. Let's discuss and figure out a way forward. A consistent grid will help us look better but also make dev easier. ![image](https://user-images.githubusercontent.com/1377389/55981621-1ed15b80-5c65-11e9-9471-9ecb276c3c88.png)
non_defect
we need a consistent grid description currently different pages use different grids and rules which don t follow the design and creating unsightly inconsistencies let s discuss and figure out a way forward a consistent grid will help us look better but also make dev easier
0