Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
26,136
| 4,593,627,316
|
IssuesEvent
|
2016-09-21 02:06:33
|
afisher1/GridLAB-D
|
https://api.github.com/repos/afisher1/GridLAB-D
|
opened
|
#63 add_module and add_class do not work properly,
|
defect
|
The new module needs to be listed in SRCDIRS
,
|
1.0
|
#63 add_module and add_class do not work properly,
- The new module needs to be listed in SRCDIRS
,
|
defect
|
add module and add class do not work properly the new module needs to be listed in srcdirs
| 1
|
5,826
| 2,610,216,204
|
IssuesEvent
|
2015-02-26 19:08:55
|
chrsmith/somefinders
|
https://api.github.com/repos/chrsmith/somefinders
|
opened
|
сенина ответы 2011
|
auto-migrated Priority-Medium Type-Defect
|
```
'''Валентин Романов'''
Привет всем не подскажите где можно найти
.сенина ответы 2011. где то видел уже
'''Валерий Беляев'''
Вот держи линк http://bit.ly/1csOg26
'''Владислав Рогов'''
Спасибо вроде то но просит телефон вводить
'''Алан Козлов'''
Не это не влияет на баланс
'''Георгий Меркушев'''
Не это не влияет на баланс
Информация о файле: сенина ответы 2011
Загружен: В этом месяце
Скачан раз: 838
Рейтинг: 1111
Средняя скорость скачивания: 839
Похожих файлов: 19
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 6:31
|
1.0
|
сенина ответы 2011 - ```
'''Валентин Романов'''
Привет всем не подскажите где можно найти
.сенина ответы 2011. где то видел уже
'''Валерий Беляев'''
Вот держи линк http://bit.ly/1csOg26
'''Владислав Рогов'''
Спасибо вроде то но просит телефон вводить
'''Алан Козлов'''
Не это не влияет на баланс
'''Георгий Меркушев'''
Не это не влияет на баланс
Информация о файле: сенина ответы 2011
Загружен: В этом месяце
Скачан раз: 838
Рейтинг: 1111
Средняя скорость скачивания: 839
Похожих файлов: 19
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 6:31
|
defect
|
сенина ответы валентин романов привет всем не подскажите где можно найти сенина ответы где то видел уже валерий беляев вот держи линк владислав рогов спасибо вроде то но просит телефон вводить алан козлов не это не влияет на баланс георгий меркушев не это не влияет на баланс информация о файле сенина ответы загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
| 1
|
97,452
| 20,263,954,860
|
IssuesEvent
|
2022-02-15 10:15:09
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
[4.0] com_contact email
|
No Code Attached Yet
|
The new form-control css -classes are really great. They give us the possibility to improve the UX significantly.
Unfortunately there is a small problem with the email-field.
Even if the input is not a valid email address the field gets the css-classes
`class="form-control validate-email required valid form-control-success"`
|
1.0
|
[4.0] com_contact email - The new form-control css -classes are really great. They give us the possibility to improve the UX significantly.
Unfortunately there is a small problem with the email-field.
Even if the input is not a valid email address the field gets the css-classes
`class="form-control validate-email required valid form-control-success"`
|
non_defect
|
com contact email the new form control css classes are really great they give us the possibility to improve the ux significantly unfortunately there is a small problem with the email field even if the input is not a valid email address the field gets the css classes class form control validate email required valid form control success
| 0
|
282,207
| 30,889,222,804
|
IssuesEvent
|
2023-08-04 02:25:02
|
madhans23/linux-4.1.15
|
https://api.github.com/repos/madhans23/linux-4.1.15
|
reopened
|
CVE-2016-6480 (Medium) detected in linux-stable-rtv4.1.33
|
Mend: dependency security vulnerability
|
## CVE-2016-6480 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/aacraid/commctrl.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/aacraid/commctrl.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Race condition in the ioctl_send_fib function in drivers/scsi/aacraid/commctrl.c in the Linux kernel through 4.7 allows local users to cause a denial of service (out-of-bounds access or system crash) by changing a certain size value, aka a "double fetch" vulnerability.
<p>Publish Date: 2016-08-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6480>CVE-2016-6480</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6480">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6480</a></p>
<p>Release Date: 2016-08-06</p>
<p>Fix Resolution: v4.8-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-6480 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2016-6480 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/aacraid/commctrl.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/aacraid/commctrl.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Race condition in the ioctl_send_fib function in drivers/scsi/aacraid/commctrl.c in the Linux kernel through 4.7 allows local users to cause a denial of service (out-of-bounds access or system crash) by changing a certain size value, aka a "double fetch" vulnerability.
<p>Publish Date: 2016-08-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6480>CVE-2016-6480</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6480">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6480</a></p>
<p>Release Date: 2016-08-06</p>
<p>Fix Resolution: v4.8-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files drivers scsi aacraid commctrl c drivers scsi aacraid commctrl c vulnerability details race condition in the ioctl send fib function in drivers scsi aacraid commctrl c in the linux kernel through allows local users to cause a denial of service out of bounds access or system crash by changing a certain size value aka a double fetch vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
111,600
| 14,112,381,316
|
IssuesEvent
|
2020-11-07 05:00:10
|
shibafuku/fukulog
|
https://api.github.com/repos/shibafuku/fukulog
|
opened
|
コンテンツページのワイヤーフレーム作成
|
design
|
各コンテンツ(mornibg、lunch、denner、yado)ページのワイヤーフレームは同じになるので
・コンテンツクリック後のページ
・店名クリック後のページ
の2種類を手書きしたものをスキャンして作成する
|
1.0
|
コンテンツページのワイヤーフレーム作成 - 各コンテンツ(mornibg、lunch、denner、yado)ページのワイヤーフレームは同じになるので
・コンテンツクリック後のページ
・店名クリック後のページ
の2種類を手書きしたものをスキャンして作成する
|
non_defect
|
コンテンツページのワイヤーフレーム作成 各コンテンツ(mornibg、lunch、denner、yado)ページのワイヤーフレームは同じになるので ・コンテンツクリック後のページ ・店名クリック後のページ
| 0
|
47,530
| 13,056,224,244
|
IssuesEvent
|
2020-07-30 04:02:38
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
port photonics_1.73 does not build on SL5 64 bit (Trac #686)
|
Migrated from Trac booking defect
|
Scientific Linux release 5.8 (Boron)
```text
$ gcc --version
gcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)
Copyright (C) 2006 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
```
```text
$ ./bin/port -vd install photonics_1.73
DEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73
DEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73
DEBUG: Requested variant x86_64 is not provided by port photonics_1.73.
DEBUG: Executing variant linux provides linux
DEBUG: Executing com.apple.main (photonics_1.73)
DEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz
DEBUG: Skipping unarchive (photonics_1.73) since no archive found
DEBUG: Skipping completed com.apple.unarchive (photonics_1.73)
---> Fetching photonics_1.73
DEBUG: Executing com.apple.fetch (photonics_1.73)
---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73
---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M
---> Verifying checksum(s) for photonics_1.73
DEBUG: Executing com.apple.checksum (photonics_1.73)
---> Checksumming photonics-1.73.tar.bz2
DEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2
DEBUG: setting option extract.cmd to /usr/bin/bzip2
---> Extracting photonics_1.73
DEBUG: Executing com.apple.extract (photonics_1.73)
---> Extracting photonics-1.73.tar.bz2
DEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'
DEBUG: Executing com.apple.patch (photonics_1.73)
---> Configuring photonics_1.73
DEBUG: Executing com.apple.configure (photonics_1.73)
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'
checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... yes
checking for a thread-safe mkdir -p... /bin/mkdir -p
checking for gawk... gawk
checking whether make sets $(MAKE)... yes
checking for gawk... (cached) gawk
checking for gcc... gcc
checking for C compiler default output file name... a.out
checking whether the C compiler works... yes
checking whether we are cross compiling... no
checking for suffix of executables...
checking for suffix of object files... o
checking whether we are using the GNU C compiler... yes
checking whether gcc accepts -g... yes
checking for gcc option to accept ISO C89... none needed
checking for style of include used by make... GNU
checking dependency style of gcc... gcc3
checking how to run the C preprocessor... gcc -E
checking for g++... g++
checking whether we are using the GNU C++ compiler... yes
checking whether g++ accepts -g... yes
checking dependency style of g++... gcc3
checking how to run the C++ preprocessor... g++ -E
checking for a BSD-compatible install... /usr/bin/install -c
checking whether ln -s works... yes
checking whether make sets $(MAKE)... (cached) yes
checking build system type... x86_64-unknown-linux-gnu
checking host system type... x86_64-unknown-linux-gnu
checking for a sed that does not truncate output... /bin/sed
checking for grep that handles long lines and -e... /bin/grep
checking for egrep... /bin/grep -E
checking for fgrep... /bin/grep -F
checking for ld used by gcc... /usr/bin/ld
checking if the linker (/usr/bin/ld) is GNU ld... yes
checking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B
checking the name lister (/usr/bin/nm -B) interface... BSD nm
checking the maximum length of command line arguments... 98304
checking whether the shell understands some XSI constructs... yes
checking whether the shell understands "+="... yes
checking for /usr/bin/ld option to reload object files... -r
checking how to recognize dependent libraries... pass_all
checking for ar... ar
checking for strip... strip
checking for ranlib... ranlib
checking command to parse /usr/bin/nm -B output from gcc object... ok
checking for ANSI C header files... yes
checking for sys/types.h... yes
checking for sys/stat.h... yes
checking for stdlib.h... yes
checking for string.h... yes
checking for memory.h... yes
checking for strings.h... yes
checking for inttypes.h... yes
checking for stdint.h... yes
checking for unistd.h... yes
checking for dlfcn.h... yes
checking whether we are using the GNU C++ compiler... (cached) yes
checking whether g++ accepts -g... (cached) yes
checking dependency style of g++... (cached) gcc3
checking how to run the C++ preprocessor... g++ -E
checking for objdir... .libs
checking if gcc supports -fno-rtti -fno-exceptions... no
checking for gcc option to produce PIC... -fPIC -DPIC
checking if gcc PIC flag -fPIC -DPIC works... yes
checking if gcc static flag -static works... yes
checking if gcc supports -c -o file.o... yes
checking if gcc supports -c -o file.o... (cached) yes
checking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking whether -lc should be explicitly linked in... no
checking dynamic linker characteristics... GNU/Linux ld.so
checking how to hardcode library paths into programs... immediate
checking whether stripping libraries is possible... yes
checking if libtool supports shared libraries... yes
checking whether to build shared libraries... yes
checking whether to build static libraries... yes
checking for ld used by g++... /usr/bin/ld -m elf_x86_64
checking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes
checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking for g++ option to produce PIC... -fPIC -DPIC
checking if g++ PIC flag -fPIC -DPIC works... yes
checking if g++ static flag -static works... yes
checking if g++ supports -c -o file.o... yes
checking if g++ supports -c -o file.o... (cached) yes
checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking dynamic linker characteristics... GNU/Linux ld.so
checking how to hardcode library paths into programs... immediate
checking for ANSI C header files... (cached) yes
checking for stdbool.h that conforms to C99... yes
checking for _Bool... yes
checking limits.h usability... yes
checking limits.h presence... yes
checking for limits.h... yes
checking malloc.h usability... yes
checking malloc.h presence... yes
checking for malloc.h... yes
checking for an ANSI C-conforming const... yes
checking for size_t... yes
checking for int8_t... yes
checking for int16_t... yes
checking for int32_t... yes
checking for int64_t... yes
checking for off_t... yes
checking for stdlib.h... (cached) yes
checking for GNU libc compatible malloc... yes
checking for stdlib.h... (cached) yes
checking for unistd.h... (cached) yes
checking for getpagesize... yes
checking for working mmap... yes
checking for stdlib.h... (cached) yes
checking for GNU libc compatible realloc... yes
checking for working strtod... yes
checking for strstr... yes
checking for strtod... (cached) yes
checking for strtol... yes
checking for strerror... yes
checking for memset... yes
checking for floor... no
checking for library containing floor... -lm
checking for pow... yes
checking for sqrt... yes
checking whether to enable debug mode... yes
disabled cernlib dependent code... yes
configure: creating ./config.status
config.status: creating Makefile
config.status: creating lib/Makefile
config.status: creating src/Makefile
config.status: creating ice/Makefile
config.status: creating scripts/Makefile
config.status: creating amasim/Makefile
config.status: creating level2/Makefile
config.status: creating config.h
config.status: executing depfiles commands
config.status: executing libtool commands
-------------------------------------------------------------------
Photonics: "1.73: pyrosoma r4"
please refer to the 'INSTALL' file for further instructions.
Hints:
Building photonics......................'make'
Compiling tool directory................'make tool'
Performing post compile test............'make tests'
All of the above........................'make everything'
Clean objects and binaries..............'make clean'
Clean tool directory....................'make toolclean'
Remove traces of previous configure.....'make distclean'
Also consider trying 'scripts/install_in_icetray.sh --help'
---> Building photonics_1.73 with target all
DEBUG: Executing com.apple.build (photonics_1.73)
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all'
make all-recursive
make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
Making all in lib
make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c
libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o
boundary.c:1: error: bad value (native) for -march= switch
boundary.c:1: error: bad value (native) for -mtune= switch
make[2]: *** [boundary.lo] Error 1
make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
make: *** [all] Error 2
Error: Target com.apple.build returned: shell command "cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all" returned error 2
Command output: make all-recursive
make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
Making all in lib
make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c
libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o
boundary.c:1: error: bad value (native) for -march= switch
boundary.c:1: error: bad value (native) for -mtune= switch
make[2]: *** [boundary.lo] Error 1
make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
make: *** [all] Error 2
Warning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install
icecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS
```
Migrated from https://code.icecube.wisc.edu/ticket/686
```json
{
"status": "closed",
"changetime": "2012-06-22T16:05:07",
"description": "Scientific Linux release 5.8 (Boron)\n\n{{{\n$ gcc --version\ngcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)\nCopyright (C) 2006 Free Software Foundation, Inc.\nThis is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n}}}\n\n{{{\n$ ./bin/port -vd install photonics_1.73\nDEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Requested variant x86_64 is not provided by port photonics_1.73.\nDEBUG: Executing variant linux provides linux\nDEBUG: Executing com.apple.main (photonics_1.73)\nDEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz\nDEBUG: Skipping unarchive (photonics_1.73) since no archive found\nDEBUG: Skipping completed com.apple.unarchive (photonics_1.73)\n---> Fetching photonics_1.73\nDEBUG: Executing com.apple.fetch (photonics_1.73)\n---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\n---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M\n---> Verifying checksum(s) for photonics_1.73\nDEBUG: Executing com.apple.checksum (photonics_1.73)\n---> Checksumming photonics-1.73.tar.bz2\nDEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2\nDEBUG: setting option extract.cmd to /usr/bin/bzip2\n---> Extracting photonics_1.73\nDEBUG: Executing com.apple.extract (photonics_1.73)\n---> Extracting photonics-1.73.tar.bz2\nDEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work\" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'\nDEBUG: Executing com.apple.patch (photonics_1.73)\n---> Configuring photonics_1.73\nDEBUG: Executing com.apple.configure (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether build environment is sane... yes\nchecking for a thread-safe mkdir -p... /bin/mkdir -p\nchecking for gawk... gawk\nchecking whether make sets $(MAKE)... yes\nchecking for gawk... (cached) gawk\nchecking for gcc... gcc\nchecking for C compiler default output file name... a.out\nchecking whether the C compiler works... yes\nchecking whether we are cross compiling... no\nchecking for suffix of executables... \nchecking for suffix of object files... o\nchecking whether we are using the GNU C compiler... yes\nchecking whether gcc accepts -g... yes\nchecking for gcc option to accept ISO C89... none needed\nchecking for style of include used by make... GNU\nchecking dependency style of gcc... gcc3\nchecking how to run the C preprocessor... gcc -E\nchecking for g++... g++\nchecking whether we are using the GNU C++ compiler... yes\nchecking whether g++ accepts -g... yes\nchecking dependency style of g++... gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether ln -s works... yes\nchecking whether make sets $(MAKE)... (cached) yes\nchecking build system type... x86_64-unknown-linux-gnu\nchecking host system type... x86_64-unknown-linux-gnu\nchecking for a sed that does not truncate output... /bin/sed\nchecking for grep that handles long lines and -e... /bin/grep\nchecking for egrep... /bin/grep -E\nchecking for fgrep... /bin/grep -F\nchecking for ld used by gcc... /usr/bin/ld\nchecking if the linker (/usr/bin/ld) is GNU ld... yes\nchecking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B\nchecking the name lister (/usr/bin/nm -B) interface... BSD nm\nchecking the maximum length of command line arguments... 98304\nchecking whether the shell understands some XSI constructs... yes\nchecking whether the shell understands \"+=\"... yes\nchecking for /usr/bin/ld option to reload object files... -r\nchecking how to recognize dependent libraries... pass_all\nchecking for ar... ar\nchecking for strip... strip\nchecking for ranlib... ranlib\nchecking command to parse /usr/bin/nm -B output from gcc object... ok\nchecking for ANSI C header files... yes\nchecking for sys/types.h... yes\nchecking for sys/stat.h... yes\nchecking for stdlib.h... yes\nchecking for string.h... yes\nchecking for memory.h... yes\nchecking for strings.h... yes\nchecking for inttypes.h... yes\nchecking for stdint.h... yes\nchecking for unistd.h... yes\nchecking for dlfcn.h... yes\nchecking whether we are using the GNU C++ compiler... (cached) yes\nchecking whether g++ accepts -g... (cached) yes\nchecking dependency style of g++... (cached) gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for objdir... .libs\nchecking if gcc supports -fno-rtti -fno-exceptions... no\nchecking for gcc option to produce PIC... -fPIC -DPIC\nchecking if gcc PIC flag -fPIC -DPIC works... yes\nchecking if gcc static flag -static works... yes\nchecking if gcc supports -c -o file.o... yes\nchecking if gcc supports -c -o file.o... (cached) yes\nchecking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking whether -lc should be explicitly linked in... no\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking whether stripping libraries is possible... yes\nchecking if libtool supports shared libraries... yes\nchecking whether to build shared libraries... yes\nchecking whether to build static libraries... yes\nchecking for ld used by g++... /usr/bin/ld -m elf_x86_64\nchecking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking for g++ option to produce PIC... -fPIC -DPIC\nchecking if g++ PIC flag -fPIC -DPIC works... yes\nchecking if g++ static flag -static works... yes\nchecking if g++ supports -c -o file.o... yes\nchecking if g++ supports -c -o file.o... (cached) yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking for ANSI C header files... (cached) yes\nchecking for stdbool.h that conforms to C99... yes\nchecking for _Bool... yes\nchecking limits.h usability... yes\nchecking limits.h presence... yes\nchecking for limits.h... yes\nchecking malloc.h usability... yes\nchecking malloc.h presence... yes\nchecking for malloc.h... yes\nchecking for an ANSI C-conforming const... yes\nchecking for size_t... yes\nchecking for int8_t... yes\nchecking for int16_t... yes\nchecking for int32_t... yes\nchecking for int64_t... yes\nchecking for off_t... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible malloc... yes\nchecking for stdlib.h... (cached) yes\nchecking for unistd.h... (cached) yes\nchecking for getpagesize... yes\nchecking for working mmap... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible realloc... yes\nchecking for working strtod... yes\nchecking for strstr... yes\nchecking for strtod... (cached) yes\nchecking for strtol... yes\nchecking for strerror... yes\nchecking for memset... yes\nchecking for floor... no\nchecking for library containing floor... -lm\nchecking for pow... yes\nchecking for sqrt... yes\nchecking whether to enable debug mode... yes\ndisabled cernlib dependent code... yes\nconfigure: creating ./config.status\nconfig.status: creating Makefile\nconfig.status: creating lib/Makefile\nconfig.status: creating src/Makefile\nconfig.status: creating ice/Makefile\nconfig.status: creating scripts/Makefile\nconfig.status: creating amasim/Makefile\nconfig.status: creating level2/Makefile\nconfig.status: creating config.h\nconfig.status: executing depfiles commands\nconfig.status: executing libtool commands\n-------------------------------------------------------------------\n Photonics: \"1.73: pyrosoma r4\"\n please refer to the 'INSTALL' file for further instructions.\n Hints:\n\n Building photonics......................'make'\n Compiling tool directory................'make tool'\n Performing post compile test............'make tests'\n All of the above........................'make everything'\n\n Clean objects and binaries..............'make clean'\n Clean tool directory....................'make toolclean'\n Remove traces of previous configure.....'make distclean'\n\n Also consider trying 'scripts/install_in_icetray.sh --help'\n\n\n---> Building photonics_1.73 with target all\nDEBUG: Executing com.apple.build (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all'\nmake all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\nError: Target com.apple.build returned: shell command \"cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all\" returned error 2\nCommand output: make all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\n\nWarning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install\nicecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS\n}}}",
"reporter": "boersma",
"cc": "",
"resolution": "worksforme",
"_ts": "1340381107000000",
"component": "booking",
"summary": "port photonics_1.73 does not build on SL5 64 bit",
"priority": "normal",
"keywords": "photonics I3_PORTS gcc",
"time": "2012-06-22T15:13:34",
"milestone": "",
"owner": "",
"type": "defect"
}
```
|
1.0
|
port photonics_1.73 does not build on SL5 64 bit (Trac #686) - Scientific Linux release 5.8 (Boron)
```text
$ gcc --version
gcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)
Copyright (C) 2006 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
```
```text
$ ./bin/port -vd install photonics_1.73
DEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73
DEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73
DEBUG: Requested variant x86_64 is not provided by port photonics_1.73.
DEBUG: Executing variant linux provides linux
DEBUG: Executing com.apple.main (photonics_1.73)
DEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz
DEBUG: Skipping unarchive (photonics_1.73) since no archive found
DEBUG: Skipping completed com.apple.unarchive (photonics_1.73)
---> Fetching photonics_1.73
DEBUG: Executing com.apple.fetch (photonics_1.73)
---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73
---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M
---> Verifying checksum(s) for photonics_1.73
DEBUG: Executing com.apple.checksum (photonics_1.73)
---> Checksumming photonics-1.73.tar.bz2
DEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2
DEBUG: setting option extract.cmd to /usr/bin/bzip2
---> Extracting photonics_1.73
DEBUG: Executing com.apple.extract (photonics_1.73)
---> Extracting photonics-1.73.tar.bz2
DEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'
DEBUG: Executing com.apple.patch (photonics_1.73)
---> Configuring photonics_1.73
DEBUG: Executing com.apple.configure (photonics_1.73)
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'
checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... yes
checking for a thread-safe mkdir -p... /bin/mkdir -p
checking for gawk... gawk
checking whether make sets $(MAKE)... yes
checking for gawk... (cached) gawk
checking for gcc... gcc
checking for C compiler default output file name... a.out
checking whether the C compiler works... yes
checking whether we are cross compiling... no
checking for suffix of executables...
checking for suffix of object files... o
checking whether we are using the GNU C compiler... yes
checking whether gcc accepts -g... yes
checking for gcc option to accept ISO C89... none needed
checking for style of include used by make... GNU
checking dependency style of gcc... gcc3
checking how to run the C preprocessor... gcc -E
checking for g++... g++
checking whether we are using the GNU C++ compiler... yes
checking whether g++ accepts -g... yes
checking dependency style of g++... gcc3
checking how to run the C++ preprocessor... g++ -E
checking for a BSD-compatible install... /usr/bin/install -c
checking whether ln -s works... yes
checking whether make sets $(MAKE)... (cached) yes
checking build system type... x86_64-unknown-linux-gnu
checking host system type... x86_64-unknown-linux-gnu
checking for a sed that does not truncate output... /bin/sed
checking for grep that handles long lines and -e... /bin/grep
checking for egrep... /bin/grep -E
checking for fgrep... /bin/grep -F
checking for ld used by gcc... /usr/bin/ld
checking if the linker (/usr/bin/ld) is GNU ld... yes
checking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B
checking the name lister (/usr/bin/nm -B) interface... BSD nm
checking the maximum length of command line arguments... 98304
checking whether the shell understands some XSI constructs... yes
checking whether the shell understands "+="... yes
checking for /usr/bin/ld option to reload object files... -r
checking how to recognize dependent libraries... pass_all
checking for ar... ar
checking for strip... strip
checking for ranlib... ranlib
checking command to parse /usr/bin/nm -B output from gcc object... ok
checking for ANSI C header files... yes
checking for sys/types.h... yes
checking for sys/stat.h... yes
checking for stdlib.h... yes
checking for string.h... yes
checking for memory.h... yes
checking for strings.h... yes
checking for inttypes.h... yes
checking for stdint.h... yes
checking for unistd.h... yes
checking for dlfcn.h... yes
checking whether we are using the GNU C++ compiler... (cached) yes
checking whether g++ accepts -g... (cached) yes
checking dependency style of g++... (cached) gcc3
checking how to run the C++ preprocessor... g++ -E
checking for objdir... .libs
checking if gcc supports -fno-rtti -fno-exceptions... no
checking for gcc option to produce PIC... -fPIC -DPIC
checking if gcc PIC flag -fPIC -DPIC works... yes
checking if gcc static flag -static works... yes
checking if gcc supports -c -o file.o... yes
checking if gcc supports -c -o file.o... (cached) yes
checking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking whether -lc should be explicitly linked in... no
checking dynamic linker characteristics... GNU/Linux ld.so
checking how to hardcode library paths into programs... immediate
checking whether stripping libraries is possible... yes
checking if libtool supports shared libraries... yes
checking whether to build shared libraries... yes
checking whether to build static libraries... yes
checking for ld used by g++... /usr/bin/ld -m elf_x86_64
checking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes
checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking for g++ option to produce PIC... -fPIC -DPIC
checking if g++ PIC flag -fPIC -DPIC works... yes
checking if g++ static flag -static works... yes
checking if g++ supports -c -o file.o... yes
checking if g++ supports -c -o file.o... (cached) yes
checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking dynamic linker characteristics... GNU/Linux ld.so
checking how to hardcode library paths into programs... immediate
checking for ANSI C header files... (cached) yes
checking for stdbool.h that conforms to C99... yes
checking for _Bool... yes
checking limits.h usability... yes
checking limits.h presence... yes
checking for limits.h... yes
checking malloc.h usability... yes
checking malloc.h presence... yes
checking for malloc.h... yes
checking for an ANSI C-conforming const... yes
checking for size_t... yes
checking for int8_t... yes
checking for int16_t... yes
checking for int32_t... yes
checking for int64_t... yes
checking for off_t... yes
checking for stdlib.h... (cached) yes
checking for GNU libc compatible malloc... yes
checking for stdlib.h... (cached) yes
checking for unistd.h... (cached) yes
checking for getpagesize... yes
checking for working mmap... yes
checking for stdlib.h... (cached) yes
checking for GNU libc compatible realloc... yes
checking for working strtod... yes
checking for strstr... yes
checking for strtod... (cached) yes
checking for strtol... yes
checking for strerror... yes
checking for memset... yes
checking for floor... no
checking for library containing floor... -lm
checking for pow... yes
checking for sqrt... yes
checking whether to enable debug mode... yes
disabled cernlib dependent code... yes
configure: creating ./config.status
config.status: creating Makefile
config.status: creating lib/Makefile
config.status: creating src/Makefile
config.status: creating ice/Makefile
config.status: creating scripts/Makefile
config.status: creating amasim/Makefile
config.status: creating level2/Makefile
config.status: creating config.h
config.status: executing depfiles commands
config.status: executing libtool commands
-------------------------------------------------------------------
Photonics: "1.73: pyrosoma r4"
please refer to the 'INSTALL' file for further instructions.
Hints:
Building photonics......................'make'
Compiling tool directory................'make tool'
Performing post compile test............'make tests'
All of the above........................'make everything'
Clean objects and binaries..............'make clean'
Clean tool directory....................'make toolclean'
Remove traces of previous configure.....'make distclean'
Also consider trying 'scripts/install_in_icetray.sh --help'
---> Building photonics_1.73 with target all
DEBUG: Executing com.apple.build (photonics_1.73)
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all'
make all-recursive
make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
Making all in lib
make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c
libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o
boundary.c:1: error: bad value (native) for -march= switch
boundary.c:1: error: bad value (native) for -mtune= switch
make[2]: *** [boundary.lo] Error 1
make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
make: *** [all] Error 2
Error: Target com.apple.build returned: shell command "cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all" returned error 2
Command output: make all-recursive
make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
Making all in lib
make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c
libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o
boundary.c:1: error: bad value (native) for -march= switch
boundary.c:1: error: bad value (native) for -mtune= switch
make[2]: *** [boundary.lo] Error 1
make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
make: *** [all] Error 2
Warning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install
icecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS
```
Migrated from https://code.icecube.wisc.edu/ticket/686
```json
{
"status": "closed",
"changetime": "2012-06-22T16:05:07",
"description": "Scientific Linux release 5.8 (Boron)\n\n{{{\n$ gcc --version\ngcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)\nCopyright (C) 2006 Free Software Foundation, Inc.\nThis is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n}}}\n\n{{{\n$ ./bin/port -vd install photonics_1.73\nDEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Requested variant x86_64 is not provided by port photonics_1.73.\nDEBUG: Executing variant linux provides linux\nDEBUG: Executing com.apple.main (photonics_1.73)\nDEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz\nDEBUG: Skipping unarchive (photonics_1.73) since no archive found\nDEBUG: Skipping completed com.apple.unarchive (photonics_1.73)\n---> Fetching photonics_1.73\nDEBUG: Executing com.apple.fetch (photonics_1.73)\n---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\n---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M\n---> Verifying checksum(s) for photonics_1.73\nDEBUG: Executing com.apple.checksum (photonics_1.73)\n---> Checksumming photonics-1.73.tar.bz2\nDEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2\nDEBUG: setting option extract.cmd to /usr/bin/bzip2\n---> Extracting photonics_1.73\nDEBUG: Executing com.apple.extract (photonics_1.73)\n---> Extracting photonics-1.73.tar.bz2\nDEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work\" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'\nDEBUG: Executing com.apple.patch (photonics_1.73)\n---> Configuring photonics_1.73\nDEBUG: Executing com.apple.configure (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether build environment is sane... yes\nchecking for a thread-safe mkdir -p... /bin/mkdir -p\nchecking for gawk... gawk\nchecking whether make sets $(MAKE)... yes\nchecking for gawk... (cached) gawk\nchecking for gcc... gcc\nchecking for C compiler default output file name... a.out\nchecking whether the C compiler works... yes\nchecking whether we are cross compiling... no\nchecking for suffix of executables... \nchecking for suffix of object files... o\nchecking whether we are using the GNU C compiler... yes\nchecking whether gcc accepts -g... yes\nchecking for gcc option to accept ISO C89... none needed\nchecking for style of include used by make... GNU\nchecking dependency style of gcc... gcc3\nchecking how to run the C preprocessor... gcc -E\nchecking for g++... g++\nchecking whether we are using the GNU C++ compiler... yes\nchecking whether g++ accepts -g... yes\nchecking dependency style of g++... gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether ln -s works... yes\nchecking whether make sets $(MAKE)... (cached) yes\nchecking build system type... x86_64-unknown-linux-gnu\nchecking host system type... x86_64-unknown-linux-gnu\nchecking for a sed that does not truncate output... /bin/sed\nchecking for grep that handles long lines and -e... /bin/grep\nchecking for egrep... /bin/grep -E\nchecking for fgrep... /bin/grep -F\nchecking for ld used by gcc... /usr/bin/ld\nchecking if the linker (/usr/bin/ld) is GNU ld... yes\nchecking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B\nchecking the name lister (/usr/bin/nm -B) interface... BSD nm\nchecking the maximum length of command line arguments... 98304\nchecking whether the shell understands some XSI constructs... yes\nchecking whether the shell understands \"+=\"... yes\nchecking for /usr/bin/ld option to reload object files... -r\nchecking how to recognize dependent libraries... pass_all\nchecking for ar... ar\nchecking for strip... strip\nchecking for ranlib... ranlib\nchecking command to parse /usr/bin/nm -B output from gcc object... ok\nchecking for ANSI C header files... yes\nchecking for sys/types.h... yes\nchecking for sys/stat.h... yes\nchecking for stdlib.h... yes\nchecking for string.h... yes\nchecking for memory.h... yes\nchecking for strings.h... yes\nchecking for inttypes.h... yes\nchecking for stdint.h... yes\nchecking for unistd.h... yes\nchecking for dlfcn.h... yes\nchecking whether we are using the GNU C++ compiler... (cached) yes\nchecking whether g++ accepts -g... (cached) yes\nchecking dependency style of g++... (cached) gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for objdir... .libs\nchecking if gcc supports -fno-rtti -fno-exceptions... no\nchecking for gcc option to produce PIC... -fPIC -DPIC\nchecking if gcc PIC flag -fPIC -DPIC works... yes\nchecking if gcc static flag -static works... yes\nchecking if gcc supports -c -o file.o... yes\nchecking if gcc supports -c -o file.o... (cached) yes\nchecking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking whether -lc should be explicitly linked in... no\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking whether stripping libraries is possible... yes\nchecking if libtool supports shared libraries... yes\nchecking whether to build shared libraries... yes\nchecking whether to build static libraries... yes\nchecking for ld used by g++... /usr/bin/ld -m elf_x86_64\nchecking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking for g++ option to produce PIC... -fPIC -DPIC\nchecking if g++ PIC flag -fPIC -DPIC works... yes\nchecking if g++ static flag -static works... yes\nchecking if g++ supports -c -o file.o... yes\nchecking if g++ supports -c -o file.o... (cached) yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking for ANSI C header files... (cached) yes\nchecking for stdbool.h that conforms to C99... yes\nchecking for _Bool... yes\nchecking limits.h usability... yes\nchecking limits.h presence... yes\nchecking for limits.h... yes\nchecking malloc.h usability... yes\nchecking malloc.h presence... yes\nchecking for malloc.h... yes\nchecking for an ANSI C-conforming const... yes\nchecking for size_t... yes\nchecking for int8_t... yes\nchecking for int16_t... yes\nchecking for int32_t... yes\nchecking for int64_t... yes\nchecking for off_t... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible malloc... yes\nchecking for stdlib.h... (cached) yes\nchecking for unistd.h... (cached) yes\nchecking for getpagesize... yes\nchecking for working mmap... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible realloc... yes\nchecking for working strtod... yes\nchecking for strstr... yes\nchecking for strtod... (cached) yes\nchecking for strtol... yes\nchecking for strerror... yes\nchecking for memset... yes\nchecking for floor... no\nchecking for library containing floor... -lm\nchecking for pow... yes\nchecking for sqrt... yes\nchecking whether to enable debug mode... yes\ndisabled cernlib dependent code... yes\nconfigure: creating ./config.status\nconfig.status: creating Makefile\nconfig.status: creating lib/Makefile\nconfig.status: creating src/Makefile\nconfig.status: creating ice/Makefile\nconfig.status: creating scripts/Makefile\nconfig.status: creating amasim/Makefile\nconfig.status: creating level2/Makefile\nconfig.status: creating config.h\nconfig.status: executing depfiles commands\nconfig.status: executing libtool commands\n-------------------------------------------------------------------\n Photonics: \"1.73: pyrosoma r4\"\n please refer to the 'INSTALL' file for further instructions.\n Hints:\n\n Building photonics......................'make'\n Compiling tool directory................'make tool'\n Performing post compile test............'make tests'\n All of the above........................'make everything'\n\n Clean objects and binaries..............'make clean'\n Clean tool directory....................'make toolclean'\n Remove traces of previous configure.....'make distclean'\n\n Also consider trying 'scripts/install_in_icetray.sh --help'\n\n\n---> Building photonics_1.73 with target all\nDEBUG: Executing com.apple.build (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all'\nmake all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\nError: Target com.apple.build returned: shell command \"cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all\" returned error 2\nCommand output: make all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\n\nWarning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install\nicecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS\n}}}",
"reporter": "boersma",
"cc": "",
"resolution": "worksforme",
"_ts": "1340381107000000",
"component": "booking",
"summary": "port photonics_1.73 does not build on SL5 64 bit",
"priority": "normal",
"keywords": "photonics I3_PORTS gcc",
"time": "2012-06-22T15:13:34",
"milestone": "",
"owner": "",
"type": "defect"
}
```
|
defect
|
port photonics does not build on bit trac scientific linux release boron text gcc version gcc gcc red hat copyright c free software foundation inc this is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose text bin port vd install photonics debug found port in file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics debug changing to port directory net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics debug requested variant is not provided by port photonics debug executing variant linux provides linux debug executing com apple main photonics debug no tgz archive net software icecube ports var db dports packages linux photonics tgz debug skipping unarchive photonics since no archive found debug skipping completed com apple unarchive photonics fetching photonics debug executing com apple fetch photonics photonics tar doesn t seem to exist in net software icecube ports var db dports distfiles photonics attempting to fetch photonics tar from debug assembled command cd net software icecube ports var db dports distfiles photonics curl f l o photonics tar tmp total received xferd average speed time time time current dload upload total spent left speed verifying checksum s for photonics debug executing com apple checksum photonics checksumming photonics tar debug correct checksum for photonics tar debug setting option extract cmd to usr bin extracting photonics debug executing com apple extract photonics extracting photonics tar debug setting option extract args to net software icecube ports var db dports distfiles photonics photonics tar debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work usr bin dc net software icecube ports var db dports distfiles photonics photonics tar tar no same owner xf debug executing com apple patch photonics configuring photonics debug executing com apple configure photonics debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic configure prefix net software icecube ports enable optimize checking for a bsd compatible install usr bin install c checking whether build environment is sane yes checking for a thread safe mkdir p bin mkdir p checking for gawk gawk checking whether make sets make yes checking for gawk cached gawk checking for gcc gcc checking for c compiler default output file name a out checking whether the c compiler works yes checking whether we are cross compiling no checking for suffix of executables checking for suffix of object files o checking whether we are using the gnu c compiler yes checking whether gcc accepts g yes checking for gcc option to accept iso none needed checking for style of include used by make gnu checking dependency style of gcc checking how to run the c preprocessor gcc e checking for g g checking whether we are using the gnu c compiler yes checking whether g accepts g yes checking dependency style of g checking how to run the c preprocessor g e checking for a bsd compatible install usr bin install c checking whether ln s works yes checking whether make sets make cached yes checking build system type unknown linux gnu checking host system type unknown linux gnu checking for a sed that does not truncate output bin sed checking for grep that handles long lines and e bin grep checking for egrep bin grep e checking for fgrep bin grep f checking for ld used by gcc usr bin ld checking if the linker usr bin ld is gnu ld yes checking for bsd or ms compatible name lister nm usr bin nm b checking the name lister usr bin nm b interface bsd nm checking the maximum length of command line arguments checking whether the shell understands some xsi constructs yes checking whether the shell understands yes checking for usr bin ld option to reload object files r checking how to recognize dependent libraries pass all checking for ar ar checking for strip strip checking for ranlib ranlib checking command to parse usr bin nm b output from gcc object ok checking for ansi c header files yes checking for sys types h yes checking for sys stat h yes checking for stdlib h yes checking for string h yes checking for memory h yes checking for strings h yes checking for inttypes h yes checking for stdint h yes checking for unistd h yes checking for dlfcn h yes checking whether we are using the gnu c compiler cached yes checking whether g accepts g cached yes checking dependency style of g cached checking how to run the c preprocessor g e checking for objdir libs checking if gcc supports fno rtti fno exceptions no checking for gcc option to produce pic fpic dpic checking if gcc pic flag fpic dpic works yes checking if gcc static flag static works yes checking if gcc supports c o file o yes checking if gcc supports c o file o cached yes checking whether the gcc linker usr bin ld m elf supports shared libraries yes checking whether lc should be explicitly linked in no checking dynamic linker characteristics gnu linux ld so checking how to hardcode library paths into programs immediate checking whether stripping libraries is possible yes checking if libtool supports shared libraries yes checking whether to build shared libraries yes checking whether to build static libraries yes checking for ld used by g usr bin ld m elf checking if the linker usr bin ld m elf is gnu ld yes checking whether the g linker usr bin ld m elf supports shared libraries yes checking for g option to produce pic fpic dpic checking if g pic flag fpic dpic works yes checking if g static flag static works yes checking if g supports c o file o yes checking if g supports c o file o cached yes checking whether the g linker usr bin ld m elf supports shared libraries yes checking dynamic linker characteristics gnu linux ld so checking how to hardcode library paths into programs immediate checking for ansi c header files cached yes checking for stdbool h that conforms to yes checking for bool yes checking limits h usability yes checking limits h presence yes checking for limits h yes checking malloc h usability yes checking malloc h presence yes checking for malloc h yes checking for an ansi c conforming const yes checking for size t yes checking for t yes checking for t yes checking for t yes checking for t yes checking for off t yes checking for stdlib h cached yes checking for gnu libc compatible malloc yes checking for stdlib h cached yes checking for unistd h cached yes checking for getpagesize yes checking for working mmap yes checking for stdlib h cached yes checking for gnu libc compatible realloc yes checking for working strtod yes checking for strstr yes checking for strtod cached yes checking for strtol yes checking for strerror yes checking for memset yes checking for floor no checking for library containing floor lm checking for pow yes checking for sqrt yes checking whether to enable debug mode yes disabled cernlib dependent code yes configure creating config status config status creating makefile config status creating lib makefile config status creating src makefile config status creating ice makefile config status creating scripts makefile config status creating amasim makefile config status creating makefile config status creating config h config status executing depfiles commands config status executing libtool commands photonics pyrosoma please refer to the install file for further instructions hints building photonics make compiling tool directory make tool performing post compile test make tests all of the above make everything clean objects and binaries make clean clean tool directory make toolclean remove traces of previous configure make distclean also consider trying scripts install in icetray sh help building photonics with target all debug executing com apple build photonics debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all make all recursive make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics making all in lib make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c libtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o boundary c error bad value native for march switch boundary c error bad value native for mtune switch make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics make error error target com apple build returned shell command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all returned error command output make all recursive make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics making all in lib make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c libtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o boundary c error bad value native for march switch boundary c error bad value native for mtune switch make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics make error warning the following items did not execute for photonics com apple activate com apple build com apple destroot com apple archive com apple install icecubemgr net software icecube ports migrated from json status closed changetime description scientific linux release boron n n n gcc version ngcc gcc red hat ncopyright c free software foundation inc nthis is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose n n n n bin port vd install photonics ndebug found port in file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics ndebug changing to port directory net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics ndebug requested variant is not provided by port photonics ndebug executing variant linux provides linux ndebug executing com apple main photonics ndebug no tgz archive net software icecube ports var db dports packages linux photonics tgz ndebug skipping unarchive photonics since no archive found ndebug skipping completed com apple unarchive photonics n fetching photonics ndebug executing com apple fetch photonics n photonics tar doesn t seem to exist in net software icecube ports var db dports distfiles photonics n attempting to fetch photonics tar from assembled command cd net software icecube ports var db dports distfiles photonics curl f l o photonics tar tmp total received xferd average speed time time time current n dload upload total spent left speed n verifying checksum s for photonics ndebug executing com apple checksum photonics n checksumming photonics tar ndebug correct checksum for photonics tar ndebug setting option extract cmd to usr bin n extracting photonics ndebug executing com apple extract photonics n extracting photonics tar ndebug setting option extract args to net software icecube ports var db dports distfiles photonics photonics tar ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work usr bin dc net software icecube ports var db dports distfiles photonics photonics tar tar no same owner xf ndebug executing com apple patch photonics n configuring photonics ndebug executing com apple configure photonics ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic configure prefix net software icecube ports enable optimize nchecking for a bsd compatible install usr bin install c nchecking whether build environment is sane yes nchecking for a thread safe mkdir p bin mkdir p nchecking for gawk gawk nchecking whether make sets make yes nchecking for gawk cached gawk nchecking for gcc gcc nchecking for c compiler default output file name a out nchecking whether the c compiler works yes nchecking whether we are cross compiling no nchecking for suffix of executables nchecking for suffix of object files o nchecking whether we are using the gnu c compiler yes nchecking whether gcc accepts g yes nchecking for gcc option to accept iso none needed nchecking for style of include used by make gnu nchecking dependency style of gcc nchecking how to run the c preprocessor gcc e nchecking for g g nchecking whether we are using the gnu c compiler yes nchecking whether g accepts g yes nchecking dependency style of g nchecking how to run the c preprocessor g e nchecking for a bsd compatible install usr bin install c nchecking whether ln s works yes nchecking whether make sets make cached yes nchecking build system type unknown linux gnu nchecking host system type unknown linux gnu nchecking for a sed that does not truncate output bin sed nchecking for grep that handles long lines and e bin grep nchecking for egrep bin grep e nchecking for fgrep bin grep f nchecking for ld used by gcc usr bin ld nchecking if the linker usr bin ld is gnu ld yes nchecking for bsd or ms compatible name lister nm usr bin nm b nchecking the name lister usr bin nm b interface bsd nm nchecking the maximum length of command line arguments nchecking whether the shell understands some xsi constructs yes nchecking whether the shell understands yes nchecking for usr bin ld option to reload object files r nchecking how to recognize dependent libraries pass all nchecking for ar ar nchecking for strip strip nchecking for ranlib ranlib nchecking command to parse usr bin nm b output from gcc object ok nchecking for ansi c header files yes nchecking for sys types h yes nchecking for sys stat h yes nchecking for stdlib h yes nchecking for string h yes nchecking for memory h yes nchecking for strings h yes nchecking for inttypes h yes nchecking for stdint h yes nchecking for unistd h yes nchecking for dlfcn h yes nchecking whether we are using the gnu c compiler cached yes nchecking whether g accepts g cached yes nchecking dependency style of g cached nchecking how to run the c preprocessor g e nchecking for objdir libs nchecking if gcc supports fno rtti fno exceptions no nchecking for gcc option to produce pic fpic dpic nchecking if gcc pic flag fpic dpic works yes nchecking if gcc static flag static works yes nchecking if gcc supports c o file o yes nchecking if gcc supports c o file o cached yes nchecking whether the gcc linker usr bin ld m elf supports shared libraries yes nchecking whether lc should be explicitly linked in no nchecking dynamic linker characteristics gnu linux ld so nchecking how to hardcode library paths into programs immediate nchecking whether stripping libraries is possible yes nchecking if libtool supports shared libraries yes nchecking whether to build shared libraries yes nchecking whether to build static libraries yes nchecking for ld used by g usr bin ld m elf nchecking if the linker usr bin ld m elf is gnu ld yes nchecking whether the g linker usr bin ld m elf supports shared libraries yes nchecking for g option to produce pic fpic dpic nchecking if g pic flag fpic dpic works yes nchecking if g static flag static works yes nchecking if g supports c o file o yes nchecking if g supports c o file o cached yes nchecking whether the g linker usr bin ld m elf supports shared libraries yes nchecking dynamic linker characteristics gnu linux ld so nchecking how to hardcode library paths into programs immediate nchecking for ansi c header files cached yes nchecking for stdbool h that conforms to yes nchecking for bool yes nchecking limits h usability yes nchecking limits h presence yes nchecking for limits h yes nchecking malloc h usability yes nchecking malloc h presence yes nchecking for malloc h yes nchecking for an ansi c conforming const yes nchecking for size t yes nchecking for t yes nchecking for t yes nchecking for t yes nchecking for t yes nchecking for off t yes nchecking for stdlib h cached yes nchecking for gnu libc compatible malloc yes nchecking for stdlib h cached yes nchecking for unistd h cached yes nchecking for getpagesize yes nchecking for working mmap yes nchecking for stdlib h cached yes nchecking for gnu libc compatible realloc yes nchecking for working strtod yes nchecking for strstr yes nchecking for strtod cached yes nchecking for strtol yes nchecking for strerror yes nchecking for memset yes nchecking for floor no nchecking for library containing floor lm nchecking for pow yes nchecking for sqrt yes nchecking whether to enable debug mode yes ndisabled cernlib dependent code yes nconfigure creating config status nconfig status creating makefile nconfig status creating lib makefile nconfig status creating src makefile nconfig status creating ice makefile nconfig status creating scripts makefile nconfig status creating amasim makefile nconfig status creating makefile nconfig status creating config h nconfig status executing depfiles commands nconfig status executing libtool commands n n photonics pyrosoma n please refer to the install file for further instructions n hints n n building photonics make n compiling tool directory make tool n performing post compile test make tests n all of the above make everything n n clean objects and binaries make clean n clean tool directory make toolclean n remove traces of previous configure make distclean n n also consider trying scripts install in icetray sh help n n n building photonics with target all ndebug executing com apple build photonics ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all nmake all recursive nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmaking all in lib nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib n bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c nlibtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o nboundary c error bad value native for march switch nboundary c error bad value native for mtune switch nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmake error nerror target com apple build returned shell command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all returned error ncommand output make all recursive nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmaking all in lib nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib n bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c nlibtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o nboundary c error bad value native for march switch nboundary c error bad value native for mtune switch nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmake error n nwarning the following items did not execute for photonics com apple activate com apple build com apple destroot com apple archive com apple install nicecubemgr net software icecube ports n reporter boersma cc resolution worksforme ts component booking summary port photonics does not build on bit priority normal keywords photonics ports gcc time milestone owner type defect
| 1
|
48,405
| 20,144,015,331
|
IssuesEvent
|
2022-02-09 04:23:10
|
Azure/azure-powershell
|
https://api.github.com/repos/Azure/azure-powershell
|
closed
|
Enable-AzFrontDoorCustomDomainHttps doesnt return error / removes https settings from endpoint
|
Service Attention Network - Front Door customer-reported
|
We have a case where one of our certs isnt compatible with azure frontdoor and the error we get through the azure portal is:
'Failed to update the custom https configuration for the frontend host 'ENDPOINTNAME'. Error: The private key is not RSA or it is unreadable. Only RSA private key is supported for BYOC to secure a custom domain.' (we have a support case open about this)
This error is not returned when running the Enable-AzFrontDoorCustomDomainHttps cmdlet but FAR WORSE after the cmdlet is run it removes the currently configured cert from the endpoint so rather than us getting an updated cert as expected we are left with a broken config and a cert error is presented to our users.
|
1.0
|
Enable-AzFrontDoorCustomDomainHttps doesnt return error / removes https settings from endpoint - We have a case where one of our certs isnt compatible with azure frontdoor and the error we get through the azure portal is:
'Failed to update the custom https configuration for the frontend host 'ENDPOINTNAME'. Error: The private key is not RSA or it is unreadable. Only RSA private key is supported for BYOC to secure a custom domain.' (we have a support case open about this)
This error is not returned when running the Enable-AzFrontDoorCustomDomainHttps cmdlet but FAR WORSE after the cmdlet is run it removes the currently configured cert from the endpoint so rather than us getting an updated cert as expected we are left with a broken config and a cert error is presented to our users.
|
non_defect
|
enable azfrontdoorcustomdomainhttps doesnt return error removes https settings from endpoint we have a case where one of our certs isnt compatible with azure frontdoor and the error we get through the azure portal is failed to update the custom https configuration for the frontend host endpointname error the private key is not rsa or it is unreadable only rsa private key is supported for byoc to secure a custom domain we have a support case open about this this error is not returned when running the enable azfrontdoorcustomdomainhttps cmdlet but far worse after the cmdlet is run it removes the currently configured cert from the endpoint so rather than us getting an updated cert as expected we are left with a broken config and a cert error is presented to our users
| 0
|
433,686
| 12,508,571,074
|
IssuesEvent
|
2020-06-02 15:45:12
|
LLK/scratch-paint
|
https://api.github.com/repos/LLK/scratch-paint
|
closed
|
Paint editor grid contrast is too low
|
High Impact Medium Severity priority 2
|
### Expected Behavior
We lowered the contrast of the checkerboard in the background of the paint editor to reduce eye-strain, but now there are complaints that the contrast is too low.
### Actual Behavior
There are reports on the forums that it is hard to see when painting with white, or that the grid lines appear to be "missing" on some devices
### Steps to Reproduce
View paint editor
<img width="665" alt="Screen Shot 2020-05-24 at 23 04 16" src="https://user-images.githubusercontent.com/2855464/82774454-ebbf3c80-9e12-11ea-98dc-3747a48309b6.png">
### Operating System and Browser
all
|
1.0
|
Paint editor grid contrast is too low - ### Expected Behavior
We lowered the contrast of the checkerboard in the background of the paint editor to reduce eye-strain, but now there are complaints that the contrast is too low.
### Actual Behavior
There are reports on the forums that it is hard to see when painting with white, or that the grid lines appear to be "missing" on some devices
### Steps to Reproduce
View paint editor
<img width="665" alt="Screen Shot 2020-05-24 at 23 04 16" src="https://user-images.githubusercontent.com/2855464/82774454-ebbf3c80-9e12-11ea-98dc-3747a48309b6.png">
### Operating System and Browser
all
|
non_defect
|
paint editor grid contrast is too low expected behavior we lowered the contrast of the checkerboard in the background of the paint editor to reduce eye strain but now there are complaints that the contrast is too low actual behavior there are reports on the forums that it is hard to see when painting with white or that the grid lines appear to be missing on some devices steps to reproduce view paint editor img width alt screen shot at src operating system and browser all
| 0
|
20,987
| 3,441,868,298
|
IssuesEvent
|
2015-12-14 20:13:28
|
wdg/blacktree-secrets
|
https://api.github.com/repos/wdg/blacktree-secrets
|
closed
|
Secret #36768 will eat your mac
|
auto-migrated Priority-Medium Type-Defect
|
```
There is something wrong with secret #36768.
Its datatype is boolean where it would be a string "YES" or "NO".
Using a boolean here will result in windows not showing, as described in
http://www.macosxhints.com/article.php?story=20090228011201290
My suggestion:
Set datatype to "string"
Set title to "Window appearance"
Set default value to "YES"
Set Values to
"Leopard" = YES;
"Tiger" = NO;
Keep dangerous checked, as it won't eat your mac anymore, but it's still not
entirely safe.
I tried to change the secret, but it's way too old
I tried to submit a new secret, but it got marked as a duplicate of #36768
```
Original issue reported on code.google.com by `yorndej...@gmail.com` on 13 Jul 2010 at 12:21
|
1.0
|
Secret #36768 will eat your mac - ```
There is something wrong with secret #36768.
Its datatype is boolean where it would be a string "YES" or "NO".
Using a boolean here will result in windows not showing, as described in
http://www.macosxhints.com/article.php?story=20090228011201290
My suggestion:
Set datatype to "string"
Set title to "Window appearance"
Set default value to "YES"
Set Values to
"Leopard" = YES;
"Tiger" = NO;
Keep dangerous checked, as it won't eat your mac anymore, but it's still not
entirely safe.
I tried to change the secret, but it's way too old
I tried to submit a new secret, but it got marked as a duplicate of #36768
```
Original issue reported on code.google.com by `yorndej...@gmail.com` on 13 Jul 2010 at 12:21
|
defect
|
secret will eat your mac there is something wrong with secret its datatype is boolean where it would be a string yes or no using a boolean here will result in windows not showing as described in my suggestion set datatype to string set title to window appearance set default value to yes set values to leopard yes tiger no keep dangerous checked as it won t eat your mac anymore but it s still not entirely safe i tried to change the secret but it s way too old i tried to submit a new secret but it got marked as a duplicate of original issue reported on code google com by yorndej gmail com on jul at
| 1
|
148,850
| 19,552,576,467
|
IssuesEvent
|
2022-01-03 01:14:45
|
madhans23/linux-4.15
|
https://api.github.com/repos/madhans23/linux-4.15
|
opened
|
WS-2021-0566 (Medium) detected in linux-stagingv5.15
|
security vulnerability
|
## WS-2021-0566 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stagingv5.15</b></p></summary>
<p>
<p>hwmon staging tree</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git>https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
net: hns3: fix use-after-free bug in hclgevf_send_mbx_msg
This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven.
This ID is fixed in Linux Kernel version v5.15.11 by commit 4f4a353f6fe033807cd026a5de81c67469ff19b0, it was introduced in version v5.8 by commit d8355240cf8fb8b9e002b5c8458578435cea85c2. For more details please see the references link.
<p>Publish Date: 2021-12-27
<p>URL: <a href=https://github.com/gregkh/linux/commit/4f4a353f6fe033807cd026a5de81c67469ff19b0>WS-2021-0566</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002741">https://osv.dev/vulnerability/GSD-2021-1002741</a></p>
<p>Release Date: 2021-12-27</p>
<p>Fix Resolution: v5.15.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0566 (Medium) detected in linux-stagingv5.15 - ## WS-2021-0566 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stagingv5.15</b></p></summary>
<p>
<p>hwmon staging tree</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git>https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
net: hns3: fix use-after-free bug in hclgevf_send_mbx_msg
This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven.
This ID is fixed in Linux Kernel version v5.15.11 by commit 4f4a353f6fe033807cd026a5de81c67469ff19b0, it was introduced in version v5.8 by commit d8355240cf8fb8b9e002b5c8458578435cea85c2. For more details please see the references link.
<p>Publish Date: 2021-12-27
<p>URL: <a href=https://github.com/gregkh/linux/commit/4f4a353f6fe033807cd026a5de81c67469ff19b0>WS-2021-0566</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002741">https://osv.dev/vulnerability/GSD-2021-1002741</a></p>
<p>Release Date: 2021-12-27</p>
<p>Fix Resolution: v5.15.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws medium detected in linux ws medium severity vulnerability vulnerable library linux hwmon staging tree library home page a href found in base branch master vulnerable source files drivers net ethernet hisilicon hclgevf mbx c vulnerability details net fix use after free bug in hclgevf send mbx msg this is an automated id intended to aid in discovery of potential security vulnerabilities the actual impact and attack plausibility have not yet been proven this id is fixed in linux kernel version by commit it was introduced in version by commit for more details please see the references link publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
70,552
| 23,232,456,289
|
IssuesEvent
|
2022-08-03 08:50:40
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
opened
|
Can't verify user when option to send keys to verified devices only is selected
|
T-Defect A-E2EE-SAS-Verification crypto-team
|
### Steps to reproduce
1. Go to Settings > Security & Privacy
3. Enable 'Encrypt to verified devices only'
4. Go back to a DM with one of your contact, click on his profile
5. Tap on verify.
### Outcome
#### What did you expect?
That you initiate a verification process.
#### What happened instead?
It's failing because the recipient won't be able to decrypt the verification request
<img width="388" alt="image" src="https://user-images.githubusercontent.com/9841565/182565977-7387b034-7f10-4e07-ad98-97d1236cbb72.png">
So it's quite annoying, I enabled 'Encrypt to verified devices only' and I have unverified users in a room. So you would think that verify them will fix the issue, but you can't
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
1.4.32
### Homeserver
local synapse
### Will you send logs?
No
### Are you willing to provide a PR?
Yes
|
1.0
|
Can't verify user when option to send keys to verified devices only is selected - ### Steps to reproduce
1. Go to Settings > Security & Privacy
3. Enable 'Encrypt to verified devices only'
4. Go back to a DM with one of your contact, click on his profile
5. Tap on verify.
### Outcome
#### What did you expect?
That you initiate a verification process.
#### What happened instead?
It's failing because the recipient won't be able to decrypt the verification request
<img width="388" alt="image" src="https://user-images.githubusercontent.com/9841565/182565977-7387b034-7f10-4e07-ad98-97d1236cbb72.png">
So it's quite annoying, I enabled 'Encrypt to verified devices only' and I have unverified users in a room. So you would think that verify them will fix the issue, but you can't
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
1.4.32
### Homeserver
local synapse
### Will you send logs?
No
### Are you willing to provide a PR?
Yes
|
defect
|
can t verify user when option to send keys to verified devices only is selected steps to reproduce go to settings security privacy enable encrypt to verified devices only go back to a dm with one of your contact click on his profile tap on verify outcome what did you expect that you initiate a verification process what happened instead it s failing because the recipient won t be able to decrypt the verification request img width alt image src so it s quite annoying i enabled encrypt to verified devices only and i have unverified users in a room so you would think that verify them will fix the issue but you can t your phone model no response operating system version no response application version and app store homeserver local synapse will you send logs no are you willing to provide a pr yes
| 1
|
315
| 2,525,200,567
|
IssuesEvent
|
2015-01-20 22:51:43
|
AtlasOfLivingAustralia/biocache-hubs
|
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-hubs
|
closed
|
Occurrence records with invalid layer sampling?
|
bug priority-medium status-started type-defect
|
*migrated from:* https://code.google.com/p/ala/issues/detail?id=659
*date:* Thu May 1 22:17:26 2014
*author:* leebel...@gmail.com
---
Summaries of occurrence queries from Kristen Williams are suggesting (serious?) anomalies in the sampling of spatial layers of some records during processing. For example, if a query is constructed on
Country = n/a AND
IBRA = n/a AND
IMCRA = n/a AND
el884 (bathymetry and elevation) = n/a
it should only return records that occur outside Australia. The following query however
[http://biocache.ala.org.au/occurrences/search?fq=-ibra%3A*&fq=-imcra%3A*&fq=-country%3A*&fq=-el848%3A*&wkt=POLYGON((96.6%20-58.0,159.8%20-58.0,159.8%20-10.5,96.6%20-10.5,96.6%20-58.0))#tab_mapView](http://biocache.ala.org.au/occurrences/search?fq=-ibra%3A*&fq=-imcra%3A*&fq=-country%3A*&fq=-el848%3A*&wkt=POLYGON((96.6%20-58.0,159.8%20-58.0,159.8%20-10.5,96.6%20-10.5,96.6%20-58.0))#tab_mapView)
produces 40,954 records, most of which ARE in 'Australia' and are terrestrial.
A few records are marine (correctly or not as I haven't checked yet) but should still be on the bathy/topo layer (and IMCRA).
There are also a few on offshore islands, which should also be on the bathy/topo layer, if not IMBRA or IMCRA - and they should also be 'Australia' but that may be another issue.
The WKT (bounding box) here just includes Australia and associated islands.
PS: Diagnosis of classification of the records would be made easier if `#658` were solved.
|
1.0
|
Occurrence records with invalid layer sampling? -
*migrated from:* https://code.google.com/p/ala/issues/detail?id=659
*date:* Thu May 1 22:17:26 2014
*author:* leebel...@gmail.com
---
Summaries of occurrence queries from Kristen Williams are suggesting (serious?) anomalies in the sampling of spatial layers of some records during processing. For example, if a query is constructed on
Country = n/a AND
IBRA = n/a AND
IMCRA = n/a AND
el884 (bathymetry and elevation) = n/a
it should only return records that occur outside Australia. The following query however
[http://biocache.ala.org.au/occurrences/search?fq=-ibra%3A*&fq=-imcra%3A*&fq=-country%3A*&fq=-el848%3A*&wkt=POLYGON((96.6%20-58.0,159.8%20-58.0,159.8%20-10.5,96.6%20-10.5,96.6%20-58.0))#tab_mapView](http://biocache.ala.org.au/occurrences/search?fq=-ibra%3A*&fq=-imcra%3A*&fq=-country%3A*&fq=-el848%3A*&wkt=POLYGON((96.6%20-58.0,159.8%20-58.0,159.8%20-10.5,96.6%20-10.5,96.6%20-58.0))#tab_mapView)
produces 40,954 records, most of which ARE in 'Australia' and are terrestrial.
A few records are marine (correctly or not as I haven't checked yet) but should still be on the bathy/topo layer (and IMCRA).
There are also a few on offshore islands, which should also be on the bathy/topo layer, if not IMBRA or IMCRA - and they should also be 'Australia' but that may be another issue.
The WKT (bounding box) here just includes Australia and associated islands.
PS: Diagnosis of classification of the records would be made easier if `#658` were solved.
|
defect
|
occurrence records with invalid layer sampling migrated from date thu may author leebel gmail com summaries of occurrence queries from kristen williams are suggesting serious anomalies in the sampling of spatial layers of some records during processing for example if a query is constructed on country n a and ibra n a and imcra n a and bathymetry and elevation n a it should only return records that occur outside australia the following query however produces records most of which are in australia and are terrestrial a few records are marine correctly or not as i haven t checked yet but should still be on the bathy topo layer and imcra there are also a few on offshore islands which should also be on the bathy topo layer if not imbra or imcra and they should also be australia but that may be another issue the wkt bounding box here just includes australia and associated islands ps diagnosis of classification of the records would be made easier if were solved
| 1
|
29,192
| 14,003,839,126
|
IssuesEvent
|
2020-10-28 16:21:25
|
apache/trafficcontrol
|
https://api.github.com/repos/apache/trafficcontrol
|
opened
|
Reduce the number of DB requests made in ValidateServerCapabilities()
|
Traffic Ops good first issue improvement performance
|
## I'm submitting a ...
- improvement request (usability, performance, tech debt, etc.)
## Traffic Control components affected ...
- Traffic Ops
## Current behavior:
TO currently makes a DB request per server in the `ValidateServerCapabilities()` function: https://github.com/apache/trafficcontrol/blob/b99eda51b03505f48d67aadf481d16470455a7f4/traffic_ops/traffic_ops_golang/deliveryservice/servers/servers.go#L541. When a delivery service is being assigned to a large number of servers, this slows down the total execution time unnecessarily.
## New behavior:
Instead of making a DB request per server, the function should make a single DB request to get data for all the servers at once, then perform the validation. This will make it much faster to assign a delivery service to large numbers of servers at a time.
## Minimal reproduction of the problem with instructions:
Assign a delivery service with required capabilities to a large number of servers with those capabilities at a time.
|
True
|
Reduce the number of DB requests made in ValidateServerCapabilities() - ## I'm submitting a ...
- improvement request (usability, performance, tech debt, etc.)
## Traffic Control components affected ...
- Traffic Ops
## Current behavior:
TO currently makes a DB request per server in the `ValidateServerCapabilities()` function: https://github.com/apache/trafficcontrol/blob/b99eda51b03505f48d67aadf481d16470455a7f4/traffic_ops/traffic_ops_golang/deliveryservice/servers/servers.go#L541. When a delivery service is being assigned to a large number of servers, this slows down the total execution time unnecessarily.
## New behavior:
Instead of making a DB request per server, the function should make a single DB request to get data for all the servers at once, then perform the validation. This will make it much faster to assign a delivery service to large numbers of servers at a time.
## Minimal reproduction of the problem with instructions:
Assign a delivery service with required capabilities to a large number of servers with those capabilities at a time.
|
non_defect
|
reduce the number of db requests made in validateservercapabilities i m submitting a improvement request usability performance tech debt etc traffic control components affected traffic ops current behavior to currently makes a db request per server in the validateservercapabilities function when a delivery service is being assigned to a large number of servers this slows down the total execution time unnecessarily new behavior instead of making a db request per server the function should make a single db request to get data for all the servers at once then perform the validation this will make it much faster to assign a delivery service to large numbers of servers at a time minimal reproduction of the problem with instructions assign a delivery service with required capabilities to a large number of servers with those capabilities at a time
| 0
|
77,850
| 27,193,790,096
|
IssuesEvent
|
2023-02-20 02:12:44
|
hyperledger/iroha
|
https://api.github.com/repos/hyperledger/iroha
|
closed
|
[BUG] `This peer is faulty` after 1.5 hours of iroha downtime.
|
Bug iroha2 Dev defect QA-confirmed
|
### OS and Environment
MacOS, Docker Hub
### GIT commit hash
4ee707a3
### Minimum working example / Steps to reproduce
1. `docker compose up`
2. Just wait 1.5 hours

### Actual result
```bash
iroha-iroha1-1 | 2023-02-17T12:09:46.555187Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader
iroha-iroha0-1 | 2023-02-17T12:09:46.674023Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer
iroha-iroha2-1 | 2023-02-17T12:09:48.588423Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail
iroha-iroha3-1 | 2023-02-17T12:09:49.220245Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer
iroha-iroha1-1 | 2023-02-17T12:09:49.573967Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader
iroha-iroha0-1 | 2023-02-17T12:09:49.675306Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer
iroha-iroha0-1 | 2023-02-17T12:09:50.645770Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:50.764237Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:51.625035Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail
iroha-iroha3-1 | 2023-02-17T12:09:52.171543Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer
iroha-iroha1-1 | 2023-02-17T12:09:52.515576Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader
iroha-iroha0-1 | 2023-02-17T12:09:52.615514Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer
iroha-iroha1-1 | 2023-02-17T12:09:52.687375Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha3-1 | 2023-02-17T12:09:52.694003Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha1-1 | 2023-02-17T12:09:53.232625Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha0-1 | 2023-02-17T12:09:53.242050Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:53.591693Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha0-1 | 2023-02-17T12:09:53.599164Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:53.703294Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha3-1 | 2023-02-17T12:09:53.709998Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:54.578374Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail
iroha-iroha3-1 | 2023-02-17T12:09:55.124516Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer
iroha-iroha1-1 | 2023-02-17T12:09:55.458997Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader
iroha-iroha0-1 | 2023-02-17T12:09:55.575338Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer
iroha-iroha1-1 | 2023-02-17T12:09:55.613468Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha3-1 | 2023-02-17T12:09:55.659568Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
```
### Expected result
No errors during downtime.
### Who can help to reproduce?
@astrokov7
|
1.0
|
[BUG] `This peer is faulty` after 1.5 hours of iroha downtime. - ### OS and Environment
MacOS, Docker Hub
### GIT commit hash
4ee707a3
### Minimum working example / Steps to reproduce
1. `docker compose up`
2. Just wait 1.5 hours

### Actual result
```bash
iroha-iroha1-1 | 2023-02-17T12:09:46.555187Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader
iroha-iroha0-1 | 2023-02-17T12:09:46.674023Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer
iroha-iroha2-1 | 2023-02-17T12:09:48.588423Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail
iroha-iroha3-1 | 2023-02-17T12:09:49.220245Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer
iroha-iroha1-1 | 2023-02-17T12:09:49.573967Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader
iroha-iroha0-1 | 2023-02-17T12:09:49.675306Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer
iroha-iroha0-1 | 2023-02-17T12:09:50.645770Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:50.764237Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:51.625035Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail
iroha-iroha3-1 | 2023-02-17T12:09:52.171543Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer
iroha-iroha1-1 | 2023-02-17T12:09:52.515576Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader
iroha-iroha0-1 | 2023-02-17T12:09:52.615514Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer
iroha-iroha1-1 | 2023-02-17T12:09:52.687375Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha3-1 | 2023-02-17T12:09:52.694003Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha1-1 | 2023-02-17T12:09:53.232625Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha0-1 | 2023-02-17T12:09:53.242050Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:53.591693Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha0-1 | 2023-02-17T12:09:53.599164Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:53.703294Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha3-1 | 2023-02-17T12:09:53.709998Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha2-1 | 2023-02-17T12:09:54.578374Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ProxyTail
iroha-iroha3-1 | 2023-02-17T12:09:55.124516Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ObservingPeer
iroha-iroha1-1 | 2023-02-17T12:09:55.458997Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=Leader
iroha-iroha0-1 | 2023-02-17T12:09:55.575338Z INFO run: iroha_core::sumeragi::main_loop: No new transactions, requesting view change... role=ValidatingPeer
iroha-iroha1-1 | 2023-02-17T12:09:55.613468Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
iroha-iroha3-1 | 2023-02-17T12:09:55.659568Z ERROR iroha_core::sumeragi: This peer is faulty. Incoming messages have to be dropped due to low processing speed. error=sending on a full channel
```
### Expected result
No errors during downtime.
### Who can help to reproduce?
@astrokov7
|
defect
|
this peer is faulty after hours of iroha downtime os and environment macos docker hub git commit hash minimum working example steps to reproduce docker compose up just wait hours actual result bash iroha info run iroha core sumeragi main loop no new transactions requesting view change role leader iroha info run iroha core sumeragi main loop no new transactions requesting view change role validatingpeer iroha info run iroha core sumeragi main loop no new transactions requesting view change role proxytail iroha info run iroha core sumeragi main loop no new transactions requesting view change role observingpeer iroha info run iroha core sumeragi main loop no new transactions requesting view change role leader iroha info run iroha core sumeragi main loop no new transactions requesting view change role validatingpeer iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha info run iroha core sumeragi main loop no new transactions requesting view change role proxytail iroha info run iroha core sumeragi main loop no new transactions requesting view change role observingpeer iroha info run iroha core sumeragi main loop no new transactions requesting view change role leader iroha info run iroha core sumeragi main loop no new transactions requesting view change role validatingpeer iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha info run iroha core sumeragi main loop no new transactions requesting view change role proxytail iroha info run iroha core sumeragi main loop no new transactions requesting view change role observingpeer iroha info run iroha core sumeragi main loop no new transactions requesting view change role leader iroha info run iroha core sumeragi main loop no new transactions requesting view change role validatingpeer iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel iroha error iroha core sumeragi this peer is faulty incoming messages have to be dropped due to low processing speed error sending on a full channel expected result no errors during downtime who can help to reproduce
| 1
|
249,596
| 7,963,887,107
|
IssuesEvent
|
2018-07-13 19:14:15
|
phetsims/equality-explorer
|
https://api.github.com/repos/phetsims/equality-explorer
|
closed
|
disable showAnswers in production versions
|
priority:5-deferred
|
phetsims/joist#406 is a prerequisite to this.
In EqualityExplorerQueryParameters, set `showAnswers` to false if we're running a production version, so that answers can't be revealed.
|
1.0
|
disable showAnswers in production versions - phetsims/joist#406 is a prerequisite to this.
In EqualityExplorerQueryParameters, set `showAnswers` to false if we're running a production version, so that answers can't be revealed.
|
non_defect
|
disable showanswers in production versions phetsims joist is a prerequisite to this in equalityexplorerqueryparameters set showanswers to false if we re running a production version so that answers can t be revealed
| 0
|
79,635
| 28,494,306,324
|
IssuesEvent
|
2023-04-18 13:17:26
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Cannot send message
|
T-Defect X-Regression S-Major O-Occasional
|
### Steps to reproduce
1. Be in an encrypted room
2. Try to send a message
### Outcome
#### What did you expect?
Message sent
#### What happened instead?
Error
```
caught (in promise) TypeError: Cannot read properties of null (reading 'getEvents')
at TimelineWindow.getEvents (timeline-window.ts:375:37)
at TimelinePanel_TimelinePanel.getEvents (TimelinePanel.tsx:1546:64)
at TimelinePanel_TimelinePanel.reloadEvents (TimelinePanel.tsx:1533:28)
at MatrixClient.<anonymous> (TimelinePanel.tsx:882:14)
at MatrixClient.emit (events.js:158:1)
at MatrixClient.emit (typed-event-emitter.ts:59:22)
at Room.forSource (ReEmitter.ts:53:29)
at Room.emit (events.js:158:1)
at Room.emit (typed-event-emitter.ts:59:22)
at Room.addPendingEvent (room.ts:2411:14)
```
Got this a couple of times today.
### Operating system
Ubuntu 22.04.2 LTS
### Application version
Element Nightly version: 2023041201 Olm version: 3.2.12
### How did you install the app?
https://packages.element.io/debian/
### Homeserver
https://element.ems.host
### Will you send logs?
Yes
|
1.0
|
Cannot send message - ### Steps to reproduce
1. Be in an encrypted room
2. Try to send a message
### Outcome
#### What did you expect?
Message sent
#### What happened instead?
Error
```
caught (in promise) TypeError: Cannot read properties of null (reading 'getEvents')
at TimelineWindow.getEvents (timeline-window.ts:375:37)
at TimelinePanel_TimelinePanel.getEvents (TimelinePanel.tsx:1546:64)
at TimelinePanel_TimelinePanel.reloadEvents (TimelinePanel.tsx:1533:28)
at MatrixClient.<anonymous> (TimelinePanel.tsx:882:14)
at MatrixClient.emit (events.js:158:1)
at MatrixClient.emit (typed-event-emitter.ts:59:22)
at Room.forSource (ReEmitter.ts:53:29)
at Room.emit (events.js:158:1)
at Room.emit (typed-event-emitter.ts:59:22)
at Room.addPendingEvent (room.ts:2411:14)
```
Got this a couple of times today.
### Operating system
Ubuntu 22.04.2 LTS
### Application version
Element Nightly version: 2023041201 Olm version: 3.2.12
### How did you install the app?
https://packages.element.io/debian/
### Homeserver
https://element.ems.host
### Will you send logs?
Yes
|
defect
|
cannot send message steps to reproduce be in an encrypted room try to send a message outcome what did you expect message sent what happened instead error caught in promise typeerror cannot read properties of null reading getevents at timelinewindow getevents timeline window ts at timelinepanel timelinepanel getevents timelinepanel tsx at timelinepanel timelinepanel reloadevents timelinepanel tsx at matrixclient timelinepanel tsx at matrixclient emit events js at matrixclient emit typed event emitter ts at room forsource reemitter ts at room emit events js at room emit typed event emitter ts at room addpendingevent room ts got this a couple of times today operating system ubuntu lts application version element nightly version olm version how did you install the app homeserver will you send logs yes
| 1
|
11,964
| 18,722,503,454
|
IssuesEvent
|
2021-11-03 13:20:34
|
alper-turgal/SWE-573
|
https://api.github.com/repos/alper-turgal/SWE-573
|
closed
|
Write the first full version of requirements
|
requirements-task
|
- Review the preliminary requirements
- Delete the ones rejected by the product owner
- Add the confirmed requirements
- Add some requirements about the admin
|
1.0
|
Write the first full version of requirements - - Review the preliminary requirements
- Delete the ones rejected by the product owner
- Add the confirmed requirements
- Add some requirements about the admin
|
non_defect
|
write the first full version of requirements review the preliminary requirements delete the ones rejected by the product owner add the confirmed requirements add some requirements about the admin
| 0
|
14,211
| 17,094,345,025
|
IssuesEvent
|
2021-07-08 22:33:23
|
jenkinsci/configuration-as-code-plugin
|
https://api.github.com/repos/jenkinsci/configuration-as-code-plugin
|
opened
|
Unable to login to Jenkins after Jenkins upgrade to 2.289.2
|
plugin-compatibility
|
HI,
After upgrading the Jenkins version to 2.289.2, the jenkins is not working as it throws the below error
java.lang.IllegalArgumentException: Permission can not be null for sid:{USER ID}
There was no changes to the Permission Matrix
|
True
|
Unable to login to Jenkins after Jenkins upgrade to 2.289.2 - HI,
After upgrading the Jenkins version to 2.289.2, the jenkins is not working as it throws the below error
java.lang.IllegalArgumentException: Permission can not be null for sid:{USER ID}
There was no changes to the Permission Matrix
|
non_defect
|
unable to login to jenkins after jenkins upgrade to hi after upgrading the jenkins version to the jenkins is not working as it throws the below error java lang illegalargumentexception permission can not be null for sid user id there was no changes to the permission matrix
| 0
|
338,327
| 10,227,503,581
|
IssuesEvent
|
2019-08-16 20:59:46
|
Alluxio/alluxio
|
https://api.github.com/repos/Alluxio/alluxio
|
closed
|
Display three the same records in BlockInfo tab of worker node WEBUI
|
area-webui priority-high target-2.0.1 type-bug
|
**Alluxio Version:**
2.1.0-SNAPSHOT
**Describe the bug**
when entering page of **BlockInfo** of worker node , each file has three the same records in WEBUI.
clicked the link of file which three the same records, display the same information in WEBUI.


**To Reproduce**
enter page of BlockInfo of worker node
**Expected behavior**
A clear and concise description of what you expected to happen.
**Urgency**
Describe the impact and urgency of the bug.
**Additional context**
Add any other context about the problem here.
|
1.0
|
Display three the same records in BlockInfo tab of worker node WEBUI - **Alluxio Version:**
2.1.0-SNAPSHOT
**Describe the bug**
when entering page of **BlockInfo** of worker node , each file has three the same records in WEBUI.
clicked the link of file which three the same records, display the same information in WEBUI.


**To Reproduce**
enter page of BlockInfo of worker node
**Expected behavior**
A clear and concise description of what you expected to happen.
**Urgency**
Describe the impact and urgency of the bug.
**Additional context**
Add any other context about the problem here.
|
non_defect
|
display three the same records in blockinfo tab of worker node webui alluxio version snapshot describe the bug when entering page of blockinfo of worker node each file has three the same records in webui clicked the link of file which three the same records display the same information in webui to reproduce enter page of blockinfo of worker node expected behavior a clear and concise description of what you expected to happen urgency describe the impact and urgency of the bug additional context add any other context about the problem here
| 0
|
86,013
| 16,813,310,355
|
IssuesEvent
|
2021-06-17 02:37:58
|
QLou895/Terrified_2.0
|
https://api.github.com/repos/QLou895/Terrified_2.0
|
opened
|
Peripheral
|
Code Dev eyeTrackAPI
|
### Eye Tracking Interaction
- [ ] Implement peripheral stuff through a cross-hair for initial interface & testing
- [ ] Acquire face mesh (especially targeting eyes to affect screen)
- [ ] User "blinks" -> create input: (suggestion to darken room and exaggerate shadows and _something comes alive_)
- [ ] See monster in peripheral -> look at monster & disappears
- [ ] Looking Left -> something appears or looks like it occurred on the Right in peripheral
- [ ]
|
1.0
|
Peripheral - ### Eye Tracking Interaction
- [ ] Implement peripheral stuff through a cross-hair for initial interface & testing
- [ ] Acquire face mesh (especially targeting eyes to affect screen)
- [ ] User "blinks" -> create input: (suggestion to darken room and exaggerate shadows and _something comes alive_)
- [ ] See monster in peripheral -> look at monster & disappears
- [ ] Looking Left -> something appears or looks like it occurred on the Right in peripheral
- [ ]
|
non_defect
|
peripheral eye tracking interaction implement peripheral stuff through a cross hair for initial interface testing acquire face mesh especially targeting eyes to affect screen user blinks create input suggestion to darken room and exaggerate shadows and something comes alive see monster in peripheral look at monster disappears looking left something appears or looks like it occurred on the right in peripheral
| 0
|
51,214
| 13,207,395,671
|
IssuesEvent
|
2020-08-14 22:56:38
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
use JEBEventServiceIsolator.PopEvent(...) in JEBEventService (Trac #51)
|
Incomplete Migration Migrated from Trac defect jeb-server
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/51">https://code.icecube.wisc.edu/projects/icecube/ticket/51</a>, reported by tschmidtand owned by tschmidt</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-09T22:48:41",
"_ts": "1194648521000000",
"description": "Add method PopEvent(...) to JEBEventServiceIsolator and use it in JEBEventService instead of directly retrieving an event from the event service. it's working like it is done, but with using the isolator, one decouples the event service much better. Note: the event serice may not be threadsafe.\n",
"reporter": "tschmidt",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-07T15:17:33",
"component": "jeb-server",
"summary": "use JEBEventServiceIsolator.PopEvent(...) in JEBEventService",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
use JEBEventServiceIsolator.PopEvent(...) in JEBEventService (Trac #51) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/51">https://code.icecube.wisc.edu/projects/icecube/ticket/51</a>, reported by tschmidtand owned by tschmidt</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-09T22:48:41",
"_ts": "1194648521000000",
"description": "Add method PopEvent(...) to JEBEventServiceIsolator and use it in JEBEventService instead of directly retrieving an event from the event service. it's working like it is done, but with using the isolator, one decouples the event service much better. Note: the event serice may not be threadsafe.\n",
"reporter": "tschmidt",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-07T15:17:33",
"component": "jeb-server",
"summary": "use JEBEventServiceIsolator.PopEvent(...) in JEBEventService",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
|
defect
|
use jebeventserviceisolator popevent in jebeventservice trac migrated from json status closed changetime ts description add method popevent to jebeventserviceisolator and use it in jebeventservice instead of directly retrieving an event from the event service it s working like it is done but with using the isolator one decouples the event service much better note the event serice may not be threadsafe n reporter tschmidt cc resolution duplicate time component jeb server summary use jebeventserviceisolator popevent in jebeventservice priority normal keywords milestone owner tschmidt type defect
| 1
|
2,054
| 2,603,975,600
|
IssuesEvent
|
2015-02-24 19:01:24
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳湿尤能治疗
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳湿尤能治疗〓沈陽軍區政治部醫院性病〓TEL:024-31023308��
�成立于1946年,68年專注于性傳播疾病的研究和治療。位于沈�
��市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史�
��久、設備精良、技術權威、專家云集,是預防、保健、醫療
、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊��
�院、全國首批醫療規范定點單位,是第四軍醫大學、東南大�
��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤
部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功��
�
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:17
|
1.0
|
沈阳湿尤能治疗 - ```
沈阳湿尤能治疗〓沈陽軍區政治部醫院性病〓TEL:024-31023308��
�成立于1946年,68年專注于性傳播疾病的研究和治療。位于沈�
��市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史�
��久、設備精良、技術權威、專家云集,是預防、保健、醫療
、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊��
�院、全國首批醫療規范定點單位,是第四軍醫大學、東南大�
��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤
部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功��
�
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:17
|
defect
|
沈阳湿尤能治疗 沈阳湿尤能治疗〓沈陽軍區政治部醫院性病〓tel: �� � , 。位于沈� �� 。是一所與新中國同建立共輝煌的歷史� ��久、設備精良、技術權威、專家云集,是預防、保健、醫療 、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊�� �院、全國首批醫療規范定點單位,是第四軍醫大學、東南大� ��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤 部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功�� � original issue reported on code google com by gmail com on jun at
| 1
|
2,734
| 3,006,423,510
|
IssuesEvent
|
2015-07-27 10:17:10
|
Itseez/opencv
|
https://api.github.com/repos/Itseez/opencv
|
opened
|
Opencv 2.4.8, MinGW, cv::CascadeClassifier();
|
affected: 2.4 auto-transferred bug category: build/install priority: normal
|
Transferred from http://code.opencv.org/issues/3593
```
|| Denis Gottardello on 2014-03-05 17:06
|| Priority: Normal
|| Affected: 2.4.8 (latest release)
|| Category: build/install
|| Tracker: Bug
|| Difficulty:
|| PR:
|| Platform: x86 / Windows
```
Opencv 2.4.8, MinGW, cv::CascadeClassifier();
-----------
```
In order to build OpenCV with MinGW32 is necessary to apply the patch described here:
http://stackoverflow.com/questions/21103042/error-while-building-opencv-monitorfromrect-was-not-declared-in-this-scope
But the serious problem is that the object *"cv::CascadeClassifier()"* is unusable.
The row
*"cv::CascadeClassifier *a= new cv::CascadeClassifier();"*
or
*"cv::CascadeClassifier a;"*
causes an application crash!
On Linux and Mac all works well.
```
History
-------
##### Denis Gottardello on 2014-03-05 17:07
```
The problem happens also with BackgroundSubtractorMOG2 and BackgroundSubtractorMOG.
```
##### Dinar Ahmatnurov on 2014-03-05 17:16
```
Roman. please have a look
- Priority changed from Blocker to Normal
- Assignee set to Roman Donchenko
- Category set to build/install
```
##### Dinar Ahmatnurov on 2014-03-05 17:17
```
- Status changed from New to Open
```
##### Denis Gottardello on 2014-06-17 19:49
```
Is this bug still open?
```
##### Denis Gottardello on 2015-01-01 17:58
```
Up!
```
|
1.0
|
Opencv 2.4.8, MinGW, cv::CascadeClassifier(); - Transferred from http://code.opencv.org/issues/3593
```
|| Denis Gottardello on 2014-03-05 17:06
|| Priority: Normal
|| Affected: 2.4.8 (latest release)
|| Category: build/install
|| Tracker: Bug
|| Difficulty:
|| PR:
|| Platform: x86 / Windows
```
Opencv 2.4.8, MinGW, cv::CascadeClassifier();
-----------
```
In order to build OpenCV with MinGW32 is necessary to apply the patch described here:
http://stackoverflow.com/questions/21103042/error-while-building-opencv-monitorfromrect-was-not-declared-in-this-scope
But the serious problem is that the object *"cv::CascadeClassifier()"* is unusable.
The row
*"cv::CascadeClassifier *a= new cv::CascadeClassifier();"*
or
*"cv::CascadeClassifier a;"*
causes an application crash!
On Linux and Mac all works well.
```
History
-------
##### Denis Gottardello on 2014-03-05 17:07
```
The problem happens also with BackgroundSubtractorMOG2 and BackgroundSubtractorMOG.
```
##### Dinar Ahmatnurov on 2014-03-05 17:16
```
Roman. please have a look
- Priority changed from Blocker to Normal
- Assignee set to Roman Donchenko
- Category set to build/install
```
##### Dinar Ahmatnurov on 2014-03-05 17:17
```
- Status changed from New to Open
```
##### Denis Gottardello on 2014-06-17 19:49
```
Is this bug still open?
```
##### Denis Gottardello on 2015-01-01 17:58
```
Up!
```
|
non_defect
|
opencv mingw cv cascadeclassifier transferred from denis gottardello on priority normal affected latest release category build install tracker bug difficulty pr platform windows opencv mingw cv cascadeclassifier in order to build opencv with is necessary to apply the patch described here but the serious problem is that the object cv cascadeclassifier is unusable the row cv cascadeclassifier a new cv cascadeclassifier or cv cascadeclassifier a causes an application crash on linux and mac all works well history denis gottardello on the problem happens also with and backgroundsubtractormog dinar ahmatnurov on roman please have a look priority changed from blocker to normal assignee set to roman donchenko category set to build install dinar ahmatnurov on status changed from new to open denis gottardello on is this bug still open denis gottardello on up
| 0
|
4,565
| 2,610,117,963
|
IssuesEvent
|
2015-02-26 18:36:38
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
text not copied when invoking Scribefire for Safari
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
When I select text on a page and invoke the ScribeFire extension, the text is
not copied into the blog body.
What browser are you using?
Safari Version 5.0.1 (6533.17.8) on Mac OSX 10.6.4
What version of ScribeFire are you running?
1.2.0.0
```
-----
Original issue reported on code.google.com by `fiver.lo...@gmail.com` on 7 Aug 2010 at 10:45
|
1.0
|
text not copied when invoking Scribefire for Safari - ```
What's the problem?
When I select text on a page and invoke the ScribeFire extension, the text is
not copied into the blog body.
What browser are you using?
Safari Version 5.0.1 (6533.17.8) on Mac OSX 10.6.4
What version of ScribeFire are you running?
1.2.0.0
```
-----
Original issue reported on code.google.com by `fiver.lo...@gmail.com` on 7 Aug 2010 at 10:45
|
defect
|
text not copied when invoking scribefire for safari what s the problem when i select text on a page and invoke the scribefire extension the text is not copied into the blog body what browser are you using safari version on mac osx what version of scribefire are you running original issue reported on code google com by fiver lo gmail com on aug at
| 1
|
10,290
| 26,671,665,199
|
IssuesEvent
|
2023-01-26 10:49:20
|
MicrosoftDocs/architecture-center
|
https://api.github.com/repos/MicrosoftDocs/architecture-center
|
closed
|
Conflicting statement on patching AKS nodes
|
doc-enhancement assigned-to-author triaged architecture-center/svc reference-architecture/subsvc Pri2
|
[As per current document security pacthes doesnt get automaticlly applied to nodes by default , in similar document its mentioned that each evening linux node in AKS get sec pacthes vai distro security update and that behaviour is automatically configured as nodes deployed to AKS cluster.
Refernce doc links: https://learn.microsoft.com/en-us/azure/aks/concepts-security#linux-nodes]
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 43da9bdf-a3d3-a65b-b555-60dcc5ac2c55
* Version Independent ID: 9913c705-aeb7-e154-4fdf-164ff1e7ff99
* Content: [AKS Day-2 - Patch and upgrade guidance - Azure Architecture Center](https://learn.microsoft.com/en-us/azure/architecture/operator-guides/aks/aks-upgrade-practices)
* Content Source: [docs/operator-guides/aks/aks-upgrade-practices.md](https://github.com/microsoftdocs/architecture-center/blob/main/docs/operator-guides/aks/aks-upgrade-practices.md)
* Service: **architecture-center**
* Sub-service: **reference-architecture**
* GitHub Login: @rishabhsaha
* Microsoft Alias: **pnp**
|
2.0
|
Conflicting statement on patching AKS nodes -
[As per current document security pacthes doesnt get automaticlly applied to nodes by default , in similar document its mentioned that each evening linux node in AKS get sec pacthes vai distro security update and that behaviour is automatically configured as nodes deployed to AKS cluster.
Refernce doc links: https://learn.microsoft.com/en-us/azure/aks/concepts-security#linux-nodes]
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 43da9bdf-a3d3-a65b-b555-60dcc5ac2c55
* Version Independent ID: 9913c705-aeb7-e154-4fdf-164ff1e7ff99
* Content: [AKS Day-2 - Patch and upgrade guidance - Azure Architecture Center](https://learn.microsoft.com/en-us/azure/architecture/operator-guides/aks/aks-upgrade-practices)
* Content Source: [docs/operator-guides/aks/aks-upgrade-practices.md](https://github.com/microsoftdocs/architecture-center/blob/main/docs/operator-guides/aks/aks-upgrade-practices.md)
* Service: **architecture-center**
* Sub-service: **reference-architecture**
* GitHub Login: @rishabhsaha
* Microsoft Alias: **pnp**
|
non_defect
|
conflicting statement on patching aks nodes as per current document security pacthes doesnt get automaticlly applied to nodes by default in similar document its mentioned that each evening linux node in aks get sec pacthes vai distro security update and that behaviour is automatically configured as nodes deployed to aks cluster refernce doc links document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service reference architecture github login rishabhsaha microsoft alias pnp
| 0
|
251,074
| 7,999,466,359
|
IssuesEvent
|
2018-07-22 01:47:33
|
Zicerite/Gavania-Project
|
https://api.github.com/repos/Zicerite/Gavania-Project
|
opened
|
Make Gunman's Araw rune
|
Medium Priority Skill Related
|
Rune 0 = Pistol. 12 ammo, reload speed = 1.5s
Rune 1 = Shotgun, 6 ammo, reload speed = 4s
Rune 2 = Burst Shot, 15 ammo, 3 shots/burst, reload speed = 2.5s
|
1.0
|
Make Gunman's Araw rune - Rune 0 = Pistol. 12 ammo, reload speed = 1.5s
Rune 1 = Shotgun, 6 ammo, reload speed = 4s
Rune 2 = Burst Shot, 15 ammo, 3 shots/burst, reload speed = 2.5s
|
non_defect
|
make gunman s araw rune rune pistol ammo reload speed rune shotgun ammo reload speed rune burst shot ammo shots burst reload speed
| 0
|
346,522
| 24,886,950,537
|
IssuesEvent
|
2022-10-28 08:36:43
|
songivan00/ped
|
https://api.github.com/repos/songivan00/ped
|
opened
|
Suggestion for documentation and feature improvement for find
|
severity.Medium type.DocumentationBug
|
The boxed portion of the find feature seems to suggest that only the full name is allowed in this find feature, however, the find feature still works with the complete first name or last name. Maybe using 'full name' might not be the right term.
Also, I feel that an improvement to this feature could be to allow substring of the name to work too!

<!--session: 1666943941242-6358995f-a616-4a7c-a8f1-6a6e2d00776b-->
<!--Version: Web v3.4.4-->
|
1.0
|
Suggestion for documentation and feature improvement for find - The boxed portion of the find feature seems to suggest that only the full name is allowed in this find feature, however, the find feature still works with the complete first name or last name. Maybe using 'full name' might not be the right term.
Also, I feel that an improvement to this feature could be to allow substring of the name to work too!

<!--session: 1666943941242-6358995f-a616-4a7c-a8f1-6a6e2d00776b-->
<!--Version: Web v3.4.4-->
|
non_defect
|
suggestion for documentation and feature improvement for find the boxed portion of the find feature seems to suggest that only the full name is allowed in this find feature however the find feature still works with the complete first name or last name maybe using full name might not be the right term also i feel that an improvement to this feature could be to allow substring of the name to work too
| 0
|
491,646
| 14,168,372,948
|
IssuesEvent
|
2020-11-12 11:38:15
|
gnosis/conditional-tokens-explorer
|
https://api.github.com/repos/gnosis/conditional-tokens-explorer
|
closed
|
[UI/UX] Search option is not fully colored (some missing pixels at the top and from the right border)
|
Low priority bug
|
See comment https://github.com/gnosis/conditional-tokens-explorer/pull/571#issuecomment-723940788
Point 7

|
1.0
|
[UI/UX] Search option is not fully colored (some missing pixels at the top and from the right border) - See comment https://github.com/gnosis/conditional-tokens-explorer/pull/571#issuecomment-723940788
Point 7

|
non_defect
|
search option is not fully colored some missing pixels at the top and from the right border see comment point
| 0
|
269,229
| 28,960,038,850
|
IssuesEvent
|
2023-05-10 01:10:16
|
dreamboy9/mongo
|
https://api.github.com/repos/dreamboy9/mongo
|
reopened
|
CVE-2022-0686 (High) detected in url-parse-1.5.1.tgz
|
Mend: dependency security vulnerability
|
## CVE-2022-0686 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p>
<p>Path to dependency file: /buildscripts/libdeps/graph_visualizer_web_stack/package.json</p>
<p>Path to vulnerable library: /buildscripts/libdeps/graph_visualizer_web_stack/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- webpack-dev-server-3.11.1.tgz
- sockjs-client-1.5.1.tgz
- :x: **url-parse-1.5.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/dreamboy9/mongo/commit/60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b">60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.8.
<p>Publish Date: 2022-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0686>CVE-2022-0686</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686</a></p>
<p>Release Date: 2022-02-20</p>
<p>Fix Resolution (url-parse): 1.5.8</p>
<p>Direct dependency fix Resolution (react-scripts): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-0686 (High) detected in url-parse-1.5.1.tgz - ## CVE-2022-0686 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p>
<p>Path to dependency file: /buildscripts/libdeps/graph_visualizer_web_stack/package.json</p>
<p>Path to vulnerable library: /buildscripts/libdeps/graph_visualizer_web_stack/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- webpack-dev-server-3.11.1.tgz
- sockjs-client-1.5.1.tgz
- :x: **url-parse-1.5.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/dreamboy9/mongo/commit/60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b">60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.8.
<p>Publish Date: 2022-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0686>CVE-2022-0686</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686</a></p>
<p>Release Date: 2022-02-20</p>
<p>Fix Resolution (url-parse): 1.5.8</p>
<p>Direct dependency fix Resolution (react-scripts): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in url parse tgz cve high severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file buildscripts libdeps graph visualizer web stack package json path to vulnerable library buildscripts libdeps graph visualizer web stack node modules url parse package json dependency hierarchy react scripts tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution react scripts step up your open source security game with mend
| 0
|
46,407
| 13,055,908,758
|
IssuesEvent
|
2020-07-30 03:05:22
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
CoincSuite split_recombine.py example doesn't run (Trac #1162)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
Migrated from https://code.icecube.wisc.edu/ticket/1162
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "Here is the error message:\n\n\n{{{\nINFO (Python): Using CoincSuite Recombinations (coincsuite.py:82 in Complete)\nTraceback (most recent call last):\n File \"split_recombine.py\", line 77, in <module>\n Split_Recombine( tray, \"Split_Recombine\", params)\n File \"split_recombine.py\", line 58, in Split_Recombine\n SplitPulses = \"MaskedOfflinePulses\")\n File \"/home/jtatar/StrikeTeam/IceRec/build/lib/I3Tray.py\", line 204, in AddSegment\n return _segment(self, _name, **kwargs)\n File \"/home/jtatar/StrikeTeam/IceRec/build/lib/icecube/CoincSuite/coincsuite.py\", line 169, in Complete\n mininame = lilliput.add_minuit_simplex_minimizer_service(tray)\nNameError: global name 'lilliput' is not defined\n\n}}}\n\n* Default input file does not exist.\n* Please add a better summary of what script does at top of script.",
"reporter": "jtatar",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "CoincSuite split_recombine.py example doesn't run",
"priority": "blocker",
"keywords": "",
"time": "2015-08-18T18:26:13",
"milestone": "",
"owner": "mzoll",
"type": "defect"
}
```
|
1.0
|
CoincSuite split_recombine.py example doesn't run (Trac #1162) - Migrated from https://code.icecube.wisc.edu/ticket/1162
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "Here is the error message:\n\n\n{{{\nINFO (Python): Using CoincSuite Recombinations (coincsuite.py:82 in Complete)\nTraceback (most recent call last):\n File \"split_recombine.py\", line 77, in <module>\n Split_Recombine( tray, \"Split_Recombine\", params)\n File \"split_recombine.py\", line 58, in Split_Recombine\n SplitPulses = \"MaskedOfflinePulses\")\n File \"/home/jtatar/StrikeTeam/IceRec/build/lib/I3Tray.py\", line 204, in AddSegment\n return _segment(self, _name, **kwargs)\n File \"/home/jtatar/StrikeTeam/IceRec/build/lib/icecube/CoincSuite/coincsuite.py\", line 169, in Complete\n mininame = lilliput.add_minuit_simplex_minimizer_service(tray)\nNameError: global name 'lilliput' is not defined\n\n}}}\n\n* Default input file does not exist.\n* Please add a better summary of what script does at top of script.",
"reporter": "jtatar",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "CoincSuite split_recombine.py example doesn't run",
"priority": "blocker",
"keywords": "",
"time": "2015-08-18T18:26:13",
"milestone": "",
"owner": "mzoll",
"type": "defect"
}
```
|
defect
|
coincsuite split recombine py example doesn t run trac migrated from json status closed changetime description here is the error message n n n ninfo python using coincsuite recombinations coincsuite py in complete ntraceback most recent call last n file split recombine py line in n split recombine tray split recombine params n file split recombine py line in split recombine n splitpulses maskedofflinepulses n file home jtatar striketeam icerec build lib py line in addsegment n return segment self name kwargs n file home jtatar striketeam icerec build lib icecube coincsuite coincsuite py line in complete n mininame lilliput add minuit simplex minimizer service tray nnameerror global name lilliput is not defined n n n n default input file does not exist n please add a better summary of what script does at top of script reporter jtatar cc resolution fixed ts component combo reconstruction summary coincsuite split recombine py example doesn t run priority blocker keywords time milestone owner mzoll type defect
| 1
|
252,084
| 27,228,676,758
|
IssuesEvent
|
2023-02-21 11:34:24
|
TIBCOSoftware/apiscout
|
https://api.github.com/repos/TIBCOSoftware/apiscout
|
closed
|
CVE-2019-11250 (Medium) detected in github.com/kubernetes/client-go/transport-v2.0.0 - autoclosed
|
security vulnerability
|
## CVE-2019-11250 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/kubernetes/client-go/transport-v2.0.0</b></p></summary>
<p>Go client for Kubernetes.</p>
<p>
Dependency Hierarchy:
- github.com/kubernetes/client-go-v2.0.0 (Root Library)
- github.com/kubernetes/client-go-v2.0.0
- github.com/kubernetes/client-go-v2.0.0
- :x: **github.com/kubernetes/client-go/transport-v2.0.0** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The Kubernetes client-go library logs request headers at verbosity levels of 7 or higher. This can disclose credentials to unauthorized users via logs or command output. Kubernetes components (such as kube-apiserver) prior to v1.16.0, which make use of basic or bearer token authentication, and run at high verbosity levels, are affected.
<p>Publish Date: 2019-08-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11250>CVE-2019-11250</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GO-2020-0065">https://osv.dev/vulnerability/GO-2020-0065</a></p>
<p>Release Date: 2020-10-16</p>
<p>Fix Resolution: v1.16.0-beta.1</p>
</p>
</details>
<p></p>
|
True
|
CVE-2019-11250 (Medium) detected in github.com/kubernetes/client-go/transport-v2.0.0 - autoclosed - ## CVE-2019-11250 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/kubernetes/client-go/transport-v2.0.0</b></p></summary>
<p>Go client for Kubernetes.</p>
<p>
Dependency Hierarchy:
- github.com/kubernetes/client-go-v2.0.0 (Root Library)
- github.com/kubernetes/client-go-v2.0.0
- github.com/kubernetes/client-go-v2.0.0
- :x: **github.com/kubernetes/client-go/transport-v2.0.0** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The Kubernetes client-go library logs request headers at verbosity levels of 7 or higher. This can disclose credentials to unauthorized users via logs or command output. Kubernetes components (such as kube-apiserver) prior to v1.16.0, which make use of basic or bearer token authentication, and run at high verbosity levels, are affected.
<p>Publish Date: 2019-08-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11250>CVE-2019-11250</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GO-2020-0065">https://osv.dev/vulnerability/GO-2020-0065</a></p>
<p>Release Date: 2020-10-16</p>
<p>Fix Resolution: v1.16.0-beta.1</p>
</p>
</details>
<p></p>
|
non_defect
|
cve medium detected in github com kubernetes client go transport autoclosed cve medium severity vulnerability vulnerable library github com kubernetes client go transport go client for kubernetes dependency hierarchy github com kubernetes client go root library github com kubernetes client go github com kubernetes client go x github com kubernetes client go transport vulnerable library vulnerability details the kubernetes client go library logs request headers at verbosity levels of or higher this can disclose credentials to unauthorized users via logs or command output kubernetes components such as kube apiserver prior to which make use of basic or bearer token authentication and run at high verbosity levels are affected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution beta
| 0
|
32,525
| 15,419,233,412
|
IssuesEvent
|
2021-03-05 09:53:10
|
Piwigo/Piwigo
|
https://api.github.com/repos/Piwigo/Piwigo
|
closed
|
storage usage broken with big galleries
|
Section: Technical Type: Performance
|
If your photo gallery contains 1 million photos, the current (11.3) algorithm to calculate the storage space used by each kind of files (photos/videos/formats/other) will reach memory limit.
Furthermore, there is no need to calculate the sum in the PHP. The database can do it, more efficiently.
On my test database with 960k photos, it takes 800ms to fetch all individuals photos and 4000ms to generate the whole page. With an aggregation query, the SQL takes 1700ms (that's longer) but the page generates in 3000ms (that's faster). Here is the aggregation query I used to test:
```
$query = '
SELECT
SUBSTRING_INDEX(path,".",-1) AS ext,
SUM(filesize) AS filesize
FROM `'.IMAGES_TABLE.'`
GROUP BY ext
;';
$file_extensions = query2array($query, 'ext', 'filesize');
```
Even optimized, it takes too much time in my opinion. Next step would be to cache this.
|
True
|
storage usage broken with big galleries - If your photo gallery contains 1 million photos, the current (11.3) algorithm to calculate the storage space used by each kind of files (photos/videos/formats/other) will reach memory limit.
Furthermore, there is no need to calculate the sum in the PHP. The database can do it, more efficiently.
On my test database with 960k photos, it takes 800ms to fetch all individuals photos and 4000ms to generate the whole page. With an aggregation query, the SQL takes 1700ms (that's longer) but the page generates in 3000ms (that's faster). Here is the aggregation query I used to test:
```
$query = '
SELECT
SUBSTRING_INDEX(path,".",-1) AS ext,
SUM(filesize) AS filesize
FROM `'.IMAGES_TABLE.'`
GROUP BY ext
;';
$file_extensions = query2array($query, 'ext', 'filesize');
```
Even optimized, it takes too much time in my opinion. Next step would be to cache this.
|
non_defect
|
storage usage broken with big galleries if your photo gallery contains million photos the current algorithm to calculate the storage space used by each kind of files photos videos formats other will reach memory limit furthermore there is no need to calculate the sum in the php the database can do it more efficiently on my test database with photos it takes to fetch all individuals photos and to generate the whole page with an aggregation query the sql takes that s longer but the page generates in that s faster here is the aggregation query i used to test query select substring index path as ext sum filesize as filesize from images table group by ext file extensions query ext filesize even optimized it takes too much time in my opinion next step would be to cache this
| 0
|
267,771
| 23,318,869,473
|
IssuesEvent
|
2022-08-08 14:42:19
|
gravitational/teleport
|
https://api.github.com/repos/gravitational/teleport
|
opened
|
`TestEC2IsInstanceMetadataAvailable` flakiness
|
flaky tests
|
## Failure
#### Link(s) to logs
- https://console.cloud.google.com/cloud-build/builds/52d61a26-7ba6-4982-a10d-4ca10841ff2a;step=0?project=ci-account
#### Relevant snippet
```
=== CONT TestEC2IsInstanceMetadataAvailable/response_with_new_id_format
ec2_test.go:156:
Error Trace: ec2_test.go:156
Error: Should be true
Test: TestEC2IsInstanceMetadataAvailable/response_with_new_id_format
--- FAIL: TestEC2IsInstanceMetadataAvailable/response_with_new_id_format (0.35s)
```
|
1.0
|
`TestEC2IsInstanceMetadataAvailable` flakiness - ## Failure
#### Link(s) to logs
- https://console.cloud.google.com/cloud-build/builds/52d61a26-7ba6-4982-a10d-4ca10841ff2a;step=0?project=ci-account
#### Relevant snippet
```
=== CONT TestEC2IsInstanceMetadataAvailable/response_with_new_id_format
ec2_test.go:156:
Error Trace: ec2_test.go:156
Error: Should be true
Test: TestEC2IsInstanceMetadataAvailable/response_with_new_id_format
--- FAIL: TestEC2IsInstanceMetadataAvailable/response_with_new_id_format (0.35s)
```
|
non_defect
|
flakiness failure link s to logs relevant snippet cont response with new id format test go error trace test go error should be true test response with new id format fail response with new id format
| 0
|
186,376
| 14,394,660,511
|
IssuesEvent
|
2020-12-03 01:49:26
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
itsivareddy/terrafrom-Oci: oci/core_virtual_circuit_test.go; 16 LoC
|
fresh small test
|
Found a possible issue in [itsivareddy/terrafrom-Oci](https://www.github.com/itsivareddy/terrafrom-Oci) at [oci/core_virtual_circuit_test.go](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_virtual_circuit_test.go#L531-L546)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_virtual_circuit_test.go#L531-L546)
<details>
<summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary>
```go
for _, virtualCircuitId := range virtualCircuitIds {
if ok := SweeperDefaultResourceId[virtualCircuitId]; !ok {
deleteVirtualCircuitRequest := oci_core.DeleteVirtualCircuitRequest{}
deleteVirtualCircuitRequest.VirtualCircuitId = &virtualCircuitId
deleteVirtualCircuitRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "core")
_, error := virtualNetworkClient.DeleteVirtualCircuit(context.Background(), deleteVirtualCircuitRequest)
if error != nil {
fmt.Printf("Error deleting VirtualCircuit %s %s, It is possible that the resource is already deleted. Please verify manually \n", virtualCircuitId, error)
continue
}
waitTillCondition(testAccProvider, &virtualCircuitId, virtualCircuitSweepWaitCondition, time.Duration(3*time.Minute),
virtualCircuitSweepResponseFetchOperation, "core", true)
}
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to virtualCircuitId is reassigned at line 535
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 075608a9e201ee0e32484da68d5ba5370dfde1be
|
1.0
|
itsivareddy/terrafrom-Oci: oci/core_virtual_circuit_test.go; 16 LoC -
Found a possible issue in [itsivareddy/terrafrom-Oci](https://www.github.com/itsivareddy/terrafrom-Oci) at [oci/core_virtual_circuit_test.go](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_virtual_circuit_test.go#L531-L546)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/itsivareddy/terrafrom-Oci/blob/075608a9e201ee0e32484da68d5ba5370dfde1be/oci/core_virtual_circuit_test.go#L531-L546)
<details>
<summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary>
```go
for _, virtualCircuitId := range virtualCircuitIds {
if ok := SweeperDefaultResourceId[virtualCircuitId]; !ok {
deleteVirtualCircuitRequest := oci_core.DeleteVirtualCircuitRequest{}
deleteVirtualCircuitRequest.VirtualCircuitId = &virtualCircuitId
deleteVirtualCircuitRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "core")
_, error := virtualNetworkClient.DeleteVirtualCircuit(context.Background(), deleteVirtualCircuitRequest)
if error != nil {
fmt.Printf("Error deleting VirtualCircuit %s %s, It is possible that the resource is already deleted. Please verify manually \n", virtualCircuitId, error)
continue
}
waitTillCondition(testAccProvider, &virtualCircuitId, virtualCircuitSweepWaitCondition, time.Duration(3*time.Minute),
virtualCircuitSweepResponseFetchOperation, "core", true)
}
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to virtualCircuitId is reassigned at line 535
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 075608a9e201ee0e32484da68d5ba5370dfde1be
|
non_defect
|
itsivareddy terrafrom oci oci core virtual circuit test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for virtualcircuitid range virtualcircuitids if ok sweeperdefaultresourceid ok deletevirtualcircuitrequest oci core deletevirtualcircuitrequest deletevirtualcircuitrequest virtualcircuitid virtualcircuitid deletevirtualcircuitrequest requestmetadata retrypolicy getretrypolicy true core error virtualnetworkclient deletevirtualcircuit context background deletevirtualcircuitrequest if error nil fmt printf error deleting virtualcircuit s s it is possible that the resource is already deleted please verify manually n virtualcircuitid error continue waittillcondition testaccprovider virtualcircuitid virtualcircuitsweepwaitcondition time duration time minute virtualcircuitsweepresponsefetchoperation core true below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to virtualcircuitid is reassigned at line leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
75,312
| 25,763,429,596
|
IssuesEvent
|
2022-12-08 22:48:52
|
vector-im/element-meta
|
https://api.github.com/repos/vector-im/element-meta
|
opened
|
Standardise wording for threads in the labs setting
|
A-Threads T-Defect O-Occasional S-Tolerable
|
We're calling them "threads", "threaded messaging" and "thread messages" across the three platforms. We should use "Threaded messages" in all cases
|
1.0
|
Standardise wording for threads in the labs setting - We're calling them "threads", "threaded messaging" and "thread messages" across the three platforms. We should use "Threaded messages" in all cases
|
defect
|
standardise wording for threads in the labs setting we re calling them threads threaded messaging and thread messages across the three platforms we should use threaded messages in all cases
| 1
|
201,065
| 22,946,722,982
|
IssuesEvent
|
2022-07-19 01:14:23
|
zencoder/castlabs-java-api
|
https://api.github.com/repos/zencoder/castlabs-java-api
|
closed
|
CVE-2020-14062 (High) detected in jackson-databind-2.9.8.jar - autoclosed
|
security vulnerability
|
## CVE-2020-14062 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /ory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/zencoder/castlabs-java-api/commit/cedc82ec94663352957fffd68b1d6bb74aba81df">cedc82ec94663352957fffd68b1d6bb74aba81df</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to com.sun.org.apache.xalan.internal.lib.sql.JNDIConnectionPool (aka xalan2).
<p>Publish Date: 2020-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14062>CVE-2020-14062</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14062">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14062</a></p>
<p>Release Date: 2020-06-14</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-14062","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to com.sun.org.apache.xalan.internal.lib.sql.JNDIConnectionPool (aka xalan2).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14062","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-14062 (High) detected in jackson-databind-2.9.8.jar - autoclosed - ## CVE-2020-14062 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /ory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/zencoder/castlabs-java-api/commit/cedc82ec94663352957fffd68b1d6bb74aba81df">cedc82ec94663352957fffd68b1d6bb74aba81df</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to com.sun.org.apache.xalan.internal.lib.sql.JNDIConnectionPool (aka xalan2).
<p>Publish Date: 2020-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14062>CVE-2020-14062</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14062">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14062</a></p>
<p>Release Date: 2020-06-14</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-14062","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to com.sun.org.apache.xalan.internal.lib.sql.JNDIConnectionPool (aka xalan2).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14062","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library ory com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com sun org apache xalan internal lib sql jndiconnectionpool aka publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com sun org apache xalan internal lib sql jndiconnectionpool aka vulnerabilityurl
| 0
|
275,408
| 23,913,596,773
|
IssuesEvent
|
2022-09-09 10:30:29
|
wpeventmanager/wp-event-manager
|
https://api.github.com/repos/wpeventmanager/wp-event-manager
|
closed
|
wp-event-manager / update-banner does not work !
|
In Testing
|
Hello,
#738 has been close but bug is always here
I tested that again
https://demo.wp-eventmanager.com/t3a6zsurong9qhc/events/#events
admin : I create an event
then I update banner using dashboard
Banner is not updated in events page
But in the dashboard thumbnail is ok
Did you saw my video ?
If it works on your side could you post a video showing what you are doing ?
Thanks
Regards
|
1.0
|
wp-event-manager / update-banner does not work ! - Hello,
#738 has been close but bug is always here
I tested that again
https://demo.wp-eventmanager.com/t3a6zsurong9qhc/events/#events
admin : I create an event
then I update banner using dashboard
Banner is not updated in events page
But in the dashboard thumbnail is ok
Did you saw my video ?
If it works on your side could you post a video showing what you are doing ?
Thanks
Regards
|
non_defect
|
wp event manager update banner does not work hello has been close but bug is always here i tested that again admin i create an event then i update banner using dashboard banner is not updated in events page but in the dashboard thumbnail is ok did you saw my video if it works on your side could you post a video showing what you are doing thanks regards
| 0
|
63,028
| 6,823,221,385
|
IssuesEvent
|
2017-11-07 22:55:51
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
Need an easy way to find leaking animations in running app (not in test)
|
dev: tests performance prod: framework
|
It's easy to end up leaking animations but not easy to catch them.
https://github.com/flutter/flutter/issues/4918#issuecomment-232770246
We have debugAssertNoTransientCallbacks which we use in flutter_test to catch animation leaks, but if you don't happen to have a flutter_test of your widget there isn't a way to catch leaked animations.
@jason-simmons and I just tried collecting to an app with a known leak via Observatory and calling debugAssertNoTransientCallbacks while the isolate was paused and it didn't seem to print anything. Regardless, we should make this less easy to get wrong or more easy to fix if you do.
|
1.0
|
Need an easy way to find leaking animations in running app (not in test) - It's easy to end up leaking animations but not easy to catch them.
https://github.com/flutter/flutter/issues/4918#issuecomment-232770246
We have debugAssertNoTransientCallbacks which we use in flutter_test to catch animation leaks, but if you don't happen to have a flutter_test of your widget there isn't a way to catch leaked animations.
@jason-simmons and I just tried collecting to an app with a known leak via Observatory and calling debugAssertNoTransientCallbacks while the isolate was paused and it didn't seem to print anything. Regardless, we should make this less easy to get wrong or more easy to fix if you do.
|
non_defect
|
need an easy way to find leaking animations in running app not in test it s easy to end up leaking animations but not easy to catch them we have debugassertnotransientcallbacks which we use in flutter test to catch animation leaks but if you don t happen to have a flutter test of your widget there isn t a way to catch leaked animations jason simmons and i just tried collecting to an app with a known leak via observatory and calling debugassertnotransientcallbacks while the isolate was paused and it didn t seem to print anything regardless we should make this less easy to get wrong or more easy to fix if you do
| 0
|
69,469
| 8,408,950,701
|
IssuesEvent
|
2018-10-12 04:48:19
|
VotingVoices/chorus
|
https://api.github.com/repos/VotingVoices/chorus
|
closed
|
Should we tone down the all-caps headers on the Plan page?
|
design feedback
|
Personally I think the all-caps headers come across as shouting, which might contribute to the feedback we've heard (@seattlehedgehog can confirm?) that people find the Plan page a little bit intimidating.
Maybe we can tone down the headers by not using all-caps?
|
1.0
|
Should we tone down the all-caps headers on the Plan page? - Personally I think the all-caps headers come across as shouting, which might contribute to the feedback we've heard (@seattlehedgehog can confirm?) that people find the Plan page a little bit intimidating.
Maybe we can tone down the headers by not using all-caps?
|
non_defect
|
should we tone down the all caps headers on the plan page personally i think the all caps headers come across as shouting which might contribute to the feedback we ve heard seattlehedgehog can confirm that people find the plan page a little bit intimidating maybe we can tone down the headers by not using all caps
| 0
|
56,722
| 15,326,099,148
|
IssuesEvent
|
2021-02-26 02:49:39
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
closed
|
Forms flagged as deleted are not being archived by the nightly forms migration.
|
Defect Product Support Team VA Forms
|
**Describe the defect**
When a form is flagged in the nightly forms DB data dump here https://prod.cms.va.gov/sites/default/files/migrate_source/va_forms_data.csv the form detail page is supposed to be archived by the migration. The form `21-0790` is flagged as deleted but the form detail page was never archived. It does not even show as being updated on the date it was deleted.
https://prod.cms.va.gov/find-forms/about-form-21-0790
**Facts:**
1. The flag was never updated in the CMS data.
2. The deleted date field value does not have the date from the CSV.
3. It was was not updated by the migration since the deletion date in the CSV.
This needs some investigation. https://dsva.slack.com/archives/CBU0KDSB1/p1611590647138900
**Rule Out**
- [x] Migration not running nightly: Last run time and Jenkins confirms it is.
- [x] Editorial mishap : No record of any mistake in the revision history.
- [x] Revision reversion: No record of it.
- [x] Bad archive logic: Tested it by hand. Worked as expected, and worked as expected with 12 of the 13 nodes that should be archived. Also no record the node was updated to reflect the deleted flag or the deleted date, which happens in the migration before the archive logic happens.
- [x] Toggle of deleted flag in the CSV: This caused no errors and worked as expected.
- [x] What if they put in a deleted flag other than 1 (like a string or some other number)? It still archived as expected.
- [x] What if there was bad data on the row and then fixed it later? It fixed it when the correct format was restored to the current state of the row.
- [x] What if they toggled the deleted on and off a few times? It worked as expected.
- [x] What if the archive logic was flawed? I checked it line by line in x-debug AND that would not explain why the deleted date was never updated.
- [x] Row id (the migration key) changed in the CSV and then changed back. This would have resulted in duplicate nodes. There are no duplicates for this form.
- [x] destination id (Node id) and sourceid (rowid from CSV) mismatch? No that looks fine.
- [x] Duplicate destination ids? No.
- [x] Duplicate sourceid in map table? No
- [x] Duplicate rowid in CSV? No
- [ ] The form deleted date is 2020-10-30 So was there anything on that night or the couple nights after that would have been a failure of the site?
- [ ] Node save happens first and then the migrate_map is updated. What if the node was saved on one instance and the migration map was updated on the new instance? (catastrophic failure of some sort?) This would cause the migration to think it had made the update and no other changes to the data row would happen. The migration map would have an updated date that does not correspond to a node save revision.
|
1.0
|
Forms flagged as deleted are not being archived by the nightly forms migration. - **Describe the defect**
When a form is flagged in the nightly forms DB data dump here https://prod.cms.va.gov/sites/default/files/migrate_source/va_forms_data.csv the form detail page is supposed to be archived by the migration. The form `21-0790` is flagged as deleted but the form detail page was never archived. It does not even show as being updated on the date it was deleted.
https://prod.cms.va.gov/find-forms/about-form-21-0790
**Facts:**
1. The flag was never updated in the CMS data.
2. The deleted date field value does not have the date from the CSV.
3. It was was not updated by the migration since the deletion date in the CSV.
This needs some investigation. https://dsva.slack.com/archives/CBU0KDSB1/p1611590647138900
**Rule Out**
- [x] Migration not running nightly: Last run time and Jenkins confirms it is.
- [x] Editorial mishap : No record of any mistake in the revision history.
- [x] Revision reversion: No record of it.
- [x] Bad archive logic: Tested it by hand. Worked as expected, and worked as expected with 12 of the 13 nodes that should be archived. Also no record the node was updated to reflect the deleted flag or the deleted date, which happens in the migration before the archive logic happens.
- [x] Toggle of deleted flag in the CSV: This caused no errors and worked as expected.
- [x] What if they put in a deleted flag other than 1 (like a string or some other number)? It still archived as expected.
- [x] What if there was bad data on the row and then fixed it later? It fixed it when the correct format was restored to the current state of the row.
- [x] What if they toggled the deleted on and off a few times? It worked as expected.
- [x] What if the archive logic was flawed? I checked it line by line in x-debug AND that would not explain why the deleted date was never updated.
- [x] Row id (the migration key) changed in the CSV and then changed back. This would have resulted in duplicate nodes. There are no duplicates for this form.
- [x] destination id (Node id) and sourceid (rowid from CSV) mismatch? No that looks fine.
- [x] Duplicate destination ids? No.
- [x] Duplicate sourceid in map table? No
- [x] Duplicate rowid in CSV? No
- [ ] The form deleted date is 2020-10-30 So was there anything on that night or the couple nights after that would have been a failure of the site?
- [ ] Node save happens first and then the migrate_map is updated. What if the node was saved on one instance and the migration map was updated on the new instance? (catastrophic failure of some sort?) This would cause the migration to think it had made the update and no other changes to the data row would happen. The migration map would have an updated date that does not correspond to a node save revision.
|
defect
|
forms flagged as deleted are not being archived by the nightly forms migration describe the defect when a form is flagged in the nightly forms db data dump here the form detail page is supposed to be archived by the migration the form is flagged as deleted but the form detail page was never archived it does not even show as being updated on the date it was deleted facts the flag was never updated in the cms data the deleted date field value does not have the date from the csv it was was not updated by the migration since the deletion date in the csv this needs some investigation rule out migration not running nightly last run time and jenkins confirms it is editorial mishap no record of any mistake in the revision history revision reversion no record of it bad archive logic tested it by hand worked as expected and worked as expected with of the nodes that should be archived also no record the node was updated to reflect the deleted flag or the deleted date which happens in the migration before the archive logic happens toggle of deleted flag in the csv this caused no errors and worked as expected what if they put in a deleted flag other than like a string or some other number it still archived as expected what if there was bad data on the row and then fixed it later it fixed it when the correct format was restored to the current state of the row what if they toggled the deleted on and off a few times it worked as expected what if the archive logic was flawed i checked it line by line in x debug and that would not explain why the deleted date was never updated row id the migration key changed in the csv and then changed back this would have resulted in duplicate nodes there are no duplicates for this form destination id node id and sourceid rowid from csv mismatch no that looks fine duplicate destination ids no duplicate sourceid in map table no duplicate rowid in csv no the form deleted date is so was there anything on that night or the couple nights after that would have been a failure of the site node save happens first and then the migrate map is updated what if the node was saved on one instance and the migration map was updated on the new instance catastrophic failure of some sort this would cause the migration to think it had made the update and no other changes to the data row would happen the migration map would have an updated date that does not correspond to a node save revision
| 1
|
34,838
| 12,303,003,713
|
IssuesEvent
|
2020-05-11 17:56:09
|
nextgenhealthcare/connect
|
https://api.github.com/repos/nextgenhealthcare/connect
|
closed
|
Add HTTPS Listener and HTTPS Sender
|
Security
|
Add an HTTPS Listener and HTTPS Sender. This was referenced in MIRTH-74 but the HTTPS part of that request was not added.
Imported Issue. Original Details:
Jira Issue Key: MIRTH-645
Reporter: jlehew
Created: 2007-11-26T09:16:32.000-0800
|
True
|
Add HTTPS Listener and HTTPS Sender - Add an HTTPS Listener and HTTPS Sender. This was referenced in MIRTH-74 but the HTTPS part of that request was not added.
Imported Issue. Original Details:
Jira Issue Key: MIRTH-645
Reporter: jlehew
Created: 2007-11-26T09:16:32.000-0800
|
non_defect
|
add https listener and https sender add an https listener and https sender this was referenced in mirth but the https part of that request was not added imported issue original details jira issue key mirth reporter jlehew created
| 0
|
15,528
| 2,859,440,490
|
IssuesEvent
|
2015-06-03 10:35:38
|
netty/netty
|
https://api.github.com/repos/netty/netty
|
closed
|
DNS resolver loses ADDITIONAL records
|
defect
|
#### Netty Version
4.1.0.Beta5 (using latest code/API per https://github.com/netty/netty/commit/63a02fc04e6ce45aee38b0437ae0279880e5fe40 @trustin )
#### Context
Additional records section is important for cutting down the number of necessary DNS questions - mainly to avoid follow-up queries for resolving hostnames (MX, CNAME or NS) to A / AAAA records. Unfortunately we've discovered that these records are **never** included when querying by Netty and we've tried to do MX queries for ~1 million different domains.
#### Step to Reproduce
First, try this on your shell prompt -
```
[root@localhost]# dig MX hotmail.com
; <<>> DiG 9.8.2rc1-RedHat-9.8.2-0.30.rc1.el6 <<>> MX hotmail.com
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 38795
;; flags: qr rd ra; QUERY: 1, ANSWER: 4, AUTHORITY: 4, ADDITIONAL: 16
;; QUESTION SECTION:
;hotmail.com. IN MX
;; ANSWER SECTION:
hotmail.com. 3600 IN MX 5 mx4.hotmail.com.
hotmail.com. 3600 IN MX 5 mx1.hotmail.com.
hotmail.com. 3600 IN MX 5 mx2.hotmail.com.
hotmail.com. 3600 IN MX 5 mx3.hotmail.com.
;; AUTHORITY SECTION:
hotmail.com. 42729 IN NS ns1.msft.net.
hotmail.com. 42729 IN NS ns2.msft.net.
hotmail.com. 42729 IN NS ns3.msft.net.
hotmail.com. 42729 IN NS ns4.msft.net.
;; ADDITIONAL SECTION:
mx1.hotmail.com. 3600 IN A 65.54.188.126
mx1.hotmail.com. 3600 IN A 65.55.33.119
mx1.hotmail.com. 3600 IN A 65.55.33.135
mx1.hotmail.com. 3600 IN A 65.55.37.72
mx1.hotmail.com. 3600 IN A 65.55.37.88
mx1.hotmail.com. 3600 IN A 65.55.37.104
mx1.hotmail.com. 3600 IN A 65.55.37.120
mx1.hotmail.com. 3600 IN A 65.55.92.136
mx1.hotmail.com. 3600 IN A 65.55.92.152
mx1.hotmail.com. 3600 IN A 65.55.92.168
mx1.hotmail.com. 3600 IN A 65.55.92.184
mx1.hotmail.com. 3600 IN A 207.46.8.167
mx1.hotmail.com. 3600 IN A 207.46.8.199
mx1.hotmail.com. 3600 IN A 65.54.188.72
mx1.hotmail.com. 3600 IN A 65.54.188.94
mx1.hotmail.com. 3600 IN A 65.54.188.110
;; Query time: 14 msec
;; SERVER: 72.13.91.2#53(72.13.91.2)
;; WHEN: Sun May 17 22:08:02 2015
;; MSG SIZE rcvd: 445
```
Then do the same query via Netty and observe no A records in the additional section of the response
```
resolver.query(new DefaultDnsQuestion("hotmail.com", DnsRecordType.MX)).addListener(this);
```
#### Possible Causes?
First of all, my knowledge of DNS protocol intricacies is limited, so I am just reporting a few things that seem different from the above ```dig``` test. I see that resolver is including some sort of ```OPT``` record in every DNS query ( https://github.com/netty/netty/blob/netty-4.1.0.Beta5/resolver-dns/src/main/java/io/netty/resolver/dns/DnsQueryContext.java#L74 ) and the DNS responses sometimes include an ```OPT``` record in the additional section. Seems to be this might be interfering with normal retrieval of additional records?
In any case perhaps it would be a good idea to allow a finer control over the additional section of the question vs. have the resolver always do exactly the same unchangeable thing...
#### Environment
[root@localhost]# java -version
java version "1.7.0_51"
Java(TM) SE Runtime Environment (build 1.7.0_51-b13)
Java HotSpot(TM) 64-Bit Server VM (build 24.51-b03, mixed mode)
[root@localhost]# uname -a
Linux localhost 2.6.32-431.el6.x86_64 #1 SMP Fri Nov 22 03:15:09 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
|
1.0
|
DNS resolver loses ADDITIONAL records - #### Netty Version
4.1.0.Beta5 (using latest code/API per https://github.com/netty/netty/commit/63a02fc04e6ce45aee38b0437ae0279880e5fe40 @trustin )
#### Context
Additional records section is important for cutting down the number of necessary DNS questions - mainly to avoid follow-up queries for resolving hostnames (MX, CNAME or NS) to A / AAAA records. Unfortunately we've discovered that these records are **never** included when querying by Netty and we've tried to do MX queries for ~1 million different domains.
#### Step to Reproduce
First, try this on your shell prompt -
```
[root@localhost]# dig MX hotmail.com
; <<>> DiG 9.8.2rc1-RedHat-9.8.2-0.30.rc1.el6 <<>> MX hotmail.com
;; global options: +cmd
;; Got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 38795
;; flags: qr rd ra; QUERY: 1, ANSWER: 4, AUTHORITY: 4, ADDITIONAL: 16
;; QUESTION SECTION:
;hotmail.com. IN MX
;; ANSWER SECTION:
hotmail.com. 3600 IN MX 5 mx4.hotmail.com.
hotmail.com. 3600 IN MX 5 mx1.hotmail.com.
hotmail.com. 3600 IN MX 5 mx2.hotmail.com.
hotmail.com. 3600 IN MX 5 mx3.hotmail.com.
;; AUTHORITY SECTION:
hotmail.com. 42729 IN NS ns1.msft.net.
hotmail.com. 42729 IN NS ns2.msft.net.
hotmail.com. 42729 IN NS ns3.msft.net.
hotmail.com. 42729 IN NS ns4.msft.net.
;; ADDITIONAL SECTION:
mx1.hotmail.com. 3600 IN A 65.54.188.126
mx1.hotmail.com. 3600 IN A 65.55.33.119
mx1.hotmail.com. 3600 IN A 65.55.33.135
mx1.hotmail.com. 3600 IN A 65.55.37.72
mx1.hotmail.com. 3600 IN A 65.55.37.88
mx1.hotmail.com. 3600 IN A 65.55.37.104
mx1.hotmail.com. 3600 IN A 65.55.37.120
mx1.hotmail.com. 3600 IN A 65.55.92.136
mx1.hotmail.com. 3600 IN A 65.55.92.152
mx1.hotmail.com. 3600 IN A 65.55.92.168
mx1.hotmail.com. 3600 IN A 65.55.92.184
mx1.hotmail.com. 3600 IN A 207.46.8.167
mx1.hotmail.com. 3600 IN A 207.46.8.199
mx1.hotmail.com. 3600 IN A 65.54.188.72
mx1.hotmail.com. 3600 IN A 65.54.188.94
mx1.hotmail.com. 3600 IN A 65.54.188.110
;; Query time: 14 msec
;; SERVER: 72.13.91.2#53(72.13.91.2)
;; WHEN: Sun May 17 22:08:02 2015
;; MSG SIZE rcvd: 445
```
Then do the same query via Netty and observe no A records in the additional section of the response
```
resolver.query(new DefaultDnsQuestion("hotmail.com", DnsRecordType.MX)).addListener(this);
```
#### Possible Causes?
First of all, my knowledge of DNS protocol intricacies is limited, so I am just reporting a few things that seem different from the above ```dig``` test. I see that resolver is including some sort of ```OPT``` record in every DNS query ( https://github.com/netty/netty/blob/netty-4.1.0.Beta5/resolver-dns/src/main/java/io/netty/resolver/dns/DnsQueryContext.java#L74 ) and the DNS responses sometimes include an ```OPT``` record in the additional section. Seems to be this might be interfering with normal retrieval of additional records?
In any case perhaps it would be a good idea to allow a finer control over the additional section of the question vs. have the resolver always do exactly the same unchangeable thing...
#### Environment
[root@localhost]# java -version
java version "1.7.0_51"
Java(TM) SE Runtime Environment (build 1.7.0_51-b13)
Java HotSpot(TM) 64-Bit Server VM (build 24.51-b03, mixed mode)
[root@localhost]# uname -a
Linux localhost 2.6.32-431.el6.x86_64 #1 SMP Fri Nov 22 03:15:09 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
|
defect
|
dns resolver loses additional records netty version using latest code api per trustin context additional records section is important for cutting down the number of necessary dns questions mainly to avoid follow up queries for resolving hostnames mx cname or ns to a aaaa records unfortunately we ve discovered that these records are never included when querying by netty and we ve tried to do mx queries for million different domains step to reproduce first try this on your shell prompt dig mx hotmail com dig redhat mx hotmail com global options cmd got answer header opcode query status noerror id flags qr rd ra query answer authority additional question section hotmail com in mx answer section hotmail com in mx hotmail com hotmail com in mx hotmail com hotmail com in mx hotmail com hotmail com in mx hotmail com authority section hotmail com in ns msft net hotmail com in ns msft net hotmail com in ns msft net hotmail com in ns msft net additional section hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a hotmail com in a query time msec server when sun may msg size rcvd then do the same query via netty and observe no a records in the additional section of the response resolver query new defaultdnsquestion hotmail com dnsrecordtype mx addlistener this possible causes first of all my knowledge of dns protocol intricacies is limited so i am just reporting a few things that seem different from the above dig test i see that resolver is including some sort of opt record in every dns query and the dns responses sometimes include an opt record in the additional section seems to be this might be interfering with normal retrieval of additional records in any case perhaps it would be a good idea to allow a finer control over the additional section of the question vs have the resolver always do exactly the same unchangeable thing environment java version java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode uname a linux localhost smp fri nov utc gnu linux
| 1
|
35,613
| 7,794,610,089
|
IssuesEvent
|
2018-06-08 03:45:08
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
RICO changed on May 29 Plots (Trac #48)
|
Migrated from Trac clubb_src defect dschanen@uwm.edu
|
The RICO case (but no other cases), changed substantially on the nightly plots for 29 May 2009. Did the initial sounding change?
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/48
```json
{
"status": "closed",
"changetime": "2009-09-29T20:02:55",
"description": "The RICO case (but no other cases), changed substantially on the nightly plots for 29 May 2009. Did the initial sounding change?",
"reporter": "dschanen@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1254254575000000",
"component": "clubb_src",
"summary": "RICO changed on May 29 Plots",
"priority": "major",
"keywords": "RICO, plotgen, night plots",
"time": "2009-05-29T17:53:24",
"milestone": "",
"owner": "dschanen@uwm.edu",
"type": "defect"
}
```
|
1.0
|
RICO changed on May 29 Plots (Trac #48) - The RICO case (but no other cases), changed substantially on the nightly plots for 29 May 2009. Did the initial sounding change?
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/48
```json
{
"status": "closed",
"changetime": "2009-09-29T20:02:55",
"description": "The RICO case (but no other cases), changed substantially on the nightly plots for 29 May 2009. Did the initial sounding change?",
"reporter": "dschanen@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1254254575000000",
"component": "clubb_src",
"summary": "RICO changed on May 29 Plots",
"priority": "major",
"keywords": "RICO, plotgen, night plots",
"time": "2009-05-29T17:53:24",
"milestone": "",
"owner": "dschanen@uwm.edu",
"type": "defect"
}
```
|
defect
|
rico changed on may plots trac the rico case but no other cases changed substantially on the nightly plots for may did the initial sounding change attachments migrated from json status closed changetime description the rico case but no other cases changed substantially on the nightly plots for may did the initial sounding change reporter dschanen uwm edu cc resolution verified by v larson ts component clubb src summary rico changed on may plots priority major keywords rico plotgen night plots time milestone owner dschanen uwm edu type defect
| 1
|
233,891
| 17,911,286,386
|
IssuesEvent
|
2021-09-09 05:50:20
|
OneTwentyFour/unicorndocs
|
https://api.github.com/repos/OneTwentyFour/unicorndocs
|
closed
|
docs reorganization
|
documentation
|
We're working on reorganizing the docs as we began doing in PR #2 ... individual branches will be created and merged eventually to main as we continue working on the documentation. This issue will track our progress as we go.
|
1.0
|
docs reorganization - We're working on reorganizing the docs as we began doing in PR #2 ... individual branches will be created and merged eventually to main as we continue working on the documentation. This issue will track our progress as we go.
|
non_defect
|
docs reorganization we re working on reorganizing the docs as we began doing in pr individual branches will be created and merged eventually to main as we continue working on the documentation this issue will track our progress as we go
| 0
|
447,316
| 31,682,573,137
|
IssuesEvent
|
2023-09-08 01:59:55
|
DeMarcoLab/fibsem
|
https://api.github.com/repos/DeMarcoLab/fibsem
|
closed
|
[DOC] User Guide
|
documentation
|
Create a user guide including:
- [x] System Configuration
- [x] Connecting / Setup
- [x] Imaging + Settings
- [x] Movement (Modes, Hotkeys)
- [x] Milling (Patterns, Settings, Hotkeys)
- [x] Manipulator (Positions, Modes)
- [x] Detection (Supervision, Features, Segmentation, Data)
|
1.0
|
[DOC] User Guide - Create a user guide including:
- [x] System Configuration
- [x] Connecting / Setup
- [x] Imaging + Settings
- [x] Movement (Modes, Hotkeys)
- [x] Milling (Patterns, Settings, Hotkeys)
- [x] Manipulator (Positions, Modes)
- [x] Detection (Supervision, Features, Segmentation, Data)
|
non_defect
|
user guide create a user guide including system configuration connecting setup imaging settings movement modes hotkeys milling patterns settings hotkeys manipulator positions modes detection supervision features segmentation data
| 0
|
139,373
| 12,855,311,096
|
IssuesEvent
|
2020-07-09 04:49:36
|
deathlyrage/pot-demo-bugs
|
https://api.github.com/repos/deathlyrage/pot-demo-bugs
|
closed
|
Skin will not load
|
documentation more info
|
**Location:** (X=-54359.347656,Y=380192.40625,Z=-2523.67749)
**Message:** My skin will not load unless I scroll in up to the dinosaurs scales
.png)
**Version:** 8222 (demo-public-test)
**Reporter:** Vigil (658-017-677)
|
1.0
|
Skin will not load -
**Location:** (X=-54359.347656,Y=380192.40625,Z=-2523.67749)
**Message:** My skin will not load unless I scroll in up to the dinosaurs scales
.png)
**Version:** 8222 (demo-public-test)
**Reporter:** Vigil (658-017-677)
|
non_defect
|
skin will not load location x y z message my skin will not load unless i scroll in up to the dinosaurs scales version demo public test reporter vigil
| 0
|
50,725
| 13,187,698,487
|
IssuesEvent
|
2020-08-13 04:16:33
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
Example scripts in weighting do not run (Trac #1238)
|
Migrated from Trac combo core defect
|
neither example scripts work out of the box
nugen_weights.py complains about NewNuFlux
corsika_weights.py complains about no input files and overwrites itslef with an hdf5 file
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1238">https://code.icecube.wisc.edu/ticket/1238</a>, reported by kjmeagher and owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:13:59",
"description": "neither example scripts work out of the box\nnugen_weights.py complains about NewNuFlux\ncorsika_weights.py complains about no input files and overwrites itslef with an hdf5 file",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1458335639558230",
"component": "combo core",
"summary": "Example scripts in weighting do not run",
"priority": "blocker",
"keywords": "",
"time": "2015-08-20T08:28:57",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Example scripts in weighting do not run (Trac #1238) - neither example scripts work out of the box
nugen_weights.py complains about NewNuFlux
corsika_weights.py complains about no input files and overwrites itslef with an hdf5 file
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1238">https://code.icecube.wisc.edu/ticket/1238</a>, reported by kjmeagher and owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:13:59",
"description": "neither example scripts work out of the box\nnugen_weights.py complains about NewNuFlux\ncorsika_weights.py complains about no input files and overwrites itslef with an hdf5 file",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1458335639558230",
"component": "combo core",
"summary": "Example scripts in weighting do not run",
"priority": "blocker",
"keywords": "",
"time": "2015-08-20T08:28:57",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
defect
|
example scripts in weighting do not run trac neither example scripts work out of the box nugen weights py complains about newnuflux corsika weights py complains about no input files and overwrites itslef with an file migrated from json status closed changetime description neither example scripts work out of the box nnugen weights py complains about newnuflux ncorsika weights py complains about no input files and overwrites itslef with an file reporter kjmeagher cc resolution fixed ts component combo core summary example scripts in weighting do not run priority blocker keywords time milestone owner jvansanten type defect
| 1
|
74,232
| 25,016,634,527
|
IssuesEvent
|
2022-11-03 19:26:16
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
closed
|
Investigation into Content release failures due to runner permission issues
|
Defect Needs refining Unplanned work Platform CMS Team
|
## Background
There were a [number of Content release failures on 10-27-2022](https://github.com/department-of-veterans-affairs/content-build/actions/workflows/content-release.yml?query=created%3A2022-10-27..2022-10-27). The failures were around lack of disk-write permissions. Example:
```
https://github.com/department-of-veterans-affairs/content-build
Removing previously created refs, to avoid conflicts
/usr/bin/git rev-parse --symbolic-full-name --verify --quiet HEAD
refs/heads/main
/usr/bin/git checkout --detach
Error: fatal: Unable to create '/home/runner/runner2/_work/content-build/content-build/.git/index.lock': Permission denied
Warning: Unable to prepare the existing repository. The repository will be recreated instead.
Deleting the contents of '/home/runner/runner2/_work/content-build/content-build'
Error: stderr maxBuffer length exceeded
```
These types of failures have happened previously and generally resolve on their own, but on 10-27 the failures lasted longer than previously (nearly 60 minutes) and were more visible to the Sitewide CMS team. We decided to spend some time investigating the issue.
A discussion of the issue took place in #platform-cop-devops (closed channel): https://dsva.slack.com/archives/C040CF55H0A/p1666928783525639
## Findings
Any [workflow that uses a self-hosted runner on the Content Build repo](https://github.com/department-of-veterans-affairs/content-build/actions) uses runners with the name [dsva-vagov-content-build-gha-runner](https://console.amazonaws-us-gov.com/ec2/v2/home?region=us-gov-west-1#Instances:instanceState=running;tag:Name=dsva-vagov-content-build-gha-runner;v=3;$case=tags:true%5C,client:false;$regex=tags:false%5C,client:false;sort=desc:launchTime). When these workflows run, they check out and cache code to defined locations on the runner that they're running on. These locations, their permissions, and any changes to those permissions are persistent for the life of the runner.
Normally, a user `runner` is designated as the user which runs the workflow job.
A select number of these workflows run parts of their workflow within containers. These containers are stood up using the root user. Examples:
* [a11y workflow](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/a11y.yml#L160)
* [a11y heading order](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/a11y-heading-order.yml#L126)
* [continuous integration](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/continuous-integration.yml#L356)
When these jobs using containers run as root, they change ownership of the directories for code checkouts & yarn/npm installation to root. This means that, when a subsequent workflow which is not containerized attempts to run, it attempts to install code to a root-owned directory. This fails because the runner user does not have permission to write to those directories.
## Recommendations
If we could isolate installation location for everything - node & node_modules and also repository checkouts - then we could allow containerized 'root' jobs to coexist on the same runners as non-containerized jobs. However, [action/setup-node](https://github.com/actions/setup-node) does not provide any mechanism for setting install location.
Therefore, the recommendation is:
* Ensure either that all jobs treat permissions in the same way, or else ensure that there are instances that can be used only for containerized and only for non-containerized.
## Follow-up tickets
* [doing the above work to create the new ASG](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11402)
* [doing the work to assign the appropriate ASG to the appropriate jobs](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11403)
* [culling any outstanding instances (there are currently something like 96 running instances which is very expensive)](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11405)
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [x] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
|
1.0
|
Investigation into Content release failures due to runner permission issues - ## Background
There were a [number of Content release failures on 10-27-2022](https://github.com/department-of-veterans-affairs/content-build/actions/workflows/content-release.yml?query=created%3A2022-10-27..2022-10-27). The failures were around lack of disk-write permissions. Example:
```
https://github.com/department-of-veterans-affairs/content-build
Removing previously created refs, to avoid conflicts
/usr/bin/git rev-parse --symbolic-full-name --verify --quiet HEAD
refs/heads/main
/usr/bin/git checkout --detach
Error: fatal: Unable to create '/home/runner/runner2/_work/content-build/content-build/.git/index.lock': Permission denied
Warning: Unable to prepare the existing repository. The repository will be recreated instead.
Deleting the contents of '/home/runner/runner2/_work/content-build/content-build'
Error: stderr maxBuffer length exceeded
```
These types of failures have happened previously and generally resolve on their own, but on 10-27 the failures lasted longer than previously (nearly 60 minutes) and were more visible to the Sitewide CMS team. We decided to spend some time investigating the issue.
A discussion of the issue took place in #platform-cop-devops (closed channel): https://dsva.slack.com/archives/C040CF55H0A/p1666928783525639
## Findings
Any [workflow that uses a self-hosted runner on the Content Build repo](https://github.com/department-of-veterans-affairs/content-build/actions) uses runners with the name [dsva-vagov-content-build-gha-runner](https://console.amazonaws-us-gov.com/ec2/v2/home?region=us-gov-west-1#Instances:instanceState=running;tag:Name=dsva-vagov-content-build-gha-runner;v=3;$case=tags:true%5C,client:false;$regex=tags:false%5C,client:false;sort=desc:launchTime). When these workflows run, they check out and cache code to defined locations on the runner that they're running on. These locations, their permissions, and any changes to those permissions are persistent for the life of the runner.
Normally, a user `runner` is designated as the user which runs the workflow job.
A select number of these workflows run parts of their workflow within containers. These containers are stood up using the root user. Examples:
* [a11y workflow](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/a11y.yml#L160)
* [a11y heading order](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/a11y-heading-order.yml#L126)
* [continuous integration](https://github.com/department-of-veterans-affairs/content-build/blob/main/.github/workflows/continuous-integration.yml#L356)
When these jobs using containers run as root, they change ownership of the directories for code checkouts & yarn/npm installation to root. This means that, when a subsequent workflow which is not containerized attempts to run, it attempts to install code to a root-owned directory. This fails because the runner user does not have permission to write to those directories.
## Recommendations
If we could isolate installation location for everything - node & node_modules and also repository checkouts - then we could allow containerized 'root' jobs to coexist on the same runners as non-containerized jobs. However, [action/setup-node](https://github.com/actions/setup-node) does not provide any mechanism for setting install location.
Therefore, the recommendation is:
* Ensure either that all jobs treat permissions in the same way, or else ensure that there are instances that can be used only for containerized and only for non-containerized.
## Follow-up tickets
* [doing the above work to create the new ASG](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11402)
* [doing the work to assign the appropriate ASG to the appropriate jobs](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11403)
* [culling any outstanding instances (there are currently something like 96 running instances which is very expensive)](https://github.com/department-of-veterans-affairs/va.gov-cms/issues/11405)
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [x] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
|
defect
|
investigation into content release failures due to runner permission issues background there were a the failures were around lack of disk write permissions example removing previously created refs to avoid conflicts usr bin git rev parse symbolic full name verify quiet head refs heads main usr bin git checkout detach error fatal unable to create home runner work content build content build git index lock permission denied warning unable to prepare the existing repository the repository will be recreated instead deleting the contents of home runner work content build content build error stderr maxbuffer length exceeded these types of failures have happened previously and generally resolve on their own but on the failures lasted longer than previously nearly minutes and were more visible to the sitewide cms team we decided to spend some time investigating the issue a discussion of the issue took place in platform cop devops closed channel findings any uses runners with the name when these workflows run they check out and cache code to defined locations on the runner that they re running on these locations their permissions and any changes to those permissions are persistent for the life of the runner normally a user runner is designated as the user which runs the workflow job a select number of these workflows run parts of their workflow within containers these containers are stood up using the root user examples when these jobs using containers run as root they change ownership of the directories for code checkouts yarn npm installation to root this means that when a subsequent workflow which is not containerized attempts to run it attempts to install code to a root owned directory this fails because the runner user does not have permission to write to those directories recommendations if we could isolate installation location for everything node node modules and also repository checkouts then we could allow containerized root jobs to coexist on the same runners as non containerized jobs however does not provide any mechanism for setting install location therefore the recommendation is ensure either that all jobs treat permissions in the same way or else ensure that there are instances that can be used only for containerized and only for non containerized follow up tickets cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
| 1
|
76,265
| 26,337,173,809
|
IssuesEvent
|
2023-01-10 15:07:52
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Creating polls does not scale well for small windows
|
T-Defect S-Minor A11y O-Occasional A-Polls
|
### Steps to reproduce

### Outcome
#### What did you expect?
A responsive modal
#### What happened instead?
Overlapping buttons and horizontal overscroll
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
1.10.10
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Creating polls does not scale well for small windows - ### Steps to reproduce

### Outcome
#### What did you expect?
A responsive modal
#### What happened instead?
Overlapping buttons and horizontal overscroll
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
1.10.10
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
creating polls does not scale well for small windows steps to reproduce outcome what did you expect a responsive modal what happened instead overlapping buttons and horizontal overscroll operating system no response browser information no response url for webapp no response application version homeserver no response will you send logs no
| 1
|
69,876
| 9,344,647,945
|
IssuesEvent
|
2019-03-30 00:00:59
|
rancher/k3s
|
https://api.github.com/repos/rancher/k3s
|
closed
|
Document image preloading
|
help wanted kind/documentation
|
You can now put docker image export tars at `/var/lib/rancher/k3s/agent/images` and they will be read and loaded into containerd on start. We need to document this :)
|
1.0
|
Document image preloading - You can now put docker image export tars at `/var/lib/rancher/k3s/agent/images` and they will be read and loaded into containerd on start. We need to document this :)
|
non_defect
|
document image preloading you can now put docker image export tars at var lib rancher agent images and they will be read and loaded into containerd on start we need to document this
| 0
|
14,016
| 2,789,853,312
|
IssuesEvent
|
2015-05-08 21:55:54
|
google/google-visualization-api-issues
|
https://api.github.com/repos/google/google-visualization-api-issues
|
opened
|
In line chart bug
|
Priority-Medium Type-Defect
|
Original [issue 302](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=302) created by orwant on 2010-06-03T11:59:41.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1.when i called the url of line chart in iphone. its width is so long. it
might not decreases. i try all the ways. at same, legends are displayed
non-order. plz rectify those two things
<b>2.</b>
<b>3.</b>
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
<b>Are you using the test environment (version 1.1)?</b>
<b>(If you are not sure, answer NO)</b>
<b>What operating system and browser are you using?</b>
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
|
1.0
|
In line chart bug - Original [issue 302](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=302) created by orwant on 2010-06-03T11:59:41.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1.when i called the url of line chart in iphone. its width is so long. it
might not decreases. i try all the ways. at same, legends are displayed
non-order. plz rectify those two things
<b>2.</b>
<b>3.</b>
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
<b>Are you using the test environment (version 1.1)?</b>
<b>(If you are not sure, answer NO)</b>
<b>What operating system and browser are you using?</b>
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
|
defect
|
in line chart bug original created by orwant on what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code when i called the url of line chart in iphone its width is so long it might not decreases i try all the ways at same legends are displayed non order plz rectify those two things what component is this issue related to piechart linechart datatable query etc are you using the test environment version if you are not sure answer no what operating system and browser are you using for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved
| 1
|
41,531
| 10,510,933,309
|
IssuesEvent
|
2019-09-27 14:26:09
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
DSL.field(Name) fields should support schema mapping
|
C: Functionality E: All Editions P: Medium R: Fixed T: Defect
|
When using `DSL.field(Name)` with qualified field references, it should be possible to apply schema and table mappings. This is currently not being done.
~We should enable this feature, including a new Setting to turn it off to allow or users to have backwards compatible behaviour.~
----
See also:
https://stackoverflow.com/q/57222533/521799
|
1.0
|
DSL.field(Name) fields should support schema mapping - When using `DSL.field(Name)` with qualified field references, it should be possible to apply schema and table mappings. This is currently not being done.
~We should enable this feature, including a new Setting to turn it off to allow or users to have backwards compatible behaviour.~
----
See also:
https://stackoverflow.com/q/57222533/521799
|
defect
|
dsl field name fields should support schema mapping when using dsl field name with qualified field references it should be possible to apply schema and table mappings this is currently not being done we should enable this feature including a new setting to turn it off to allow or users to have backwards compatible behaviour see also
| 1
|
405,994
| 27,544,647,403
|
IssuesEvent
|
2023-03-07 10:53:35
|
rupali-codes/LinksHub
|
https://api.github.com/repos/rupali-codes/LinksHub
|
closed
|
[DOCS]: Improve README.md
|
documentation good first issue chore
|
### Description
We can improve our readme by putting everything under a section like putting the last sentence under support section!
### Screenshots
_No response_
### Additional Information
We can also add a discord icon in the community joining section and linked it to our discord server as it looks really cool :)
|
1.0
|
[DOCS]: Improve README.md - ### Description
We can improve our readme by putting everything under a section like putting the last sentence under support section!
### Screenshots
_No response_
### Additional Information
We can also add a discord icon in the community joining section and linked it to our discord server as it looks really cool :)
|
non_defect
|
improve readme md description we can improve our readme by putting everything under a section like putting the last sentence under support section screenshots no response additional information we can also add a discord icon in the community joining section and linked it to our discord server as it looks really cool
| 0
|
35,359
| 7,714,927,045
|
IssuesEvent
|
2018-05-23 05:13:44
|
bridgedotnet/Bridge.Newtonsoft.Json
|
https://api.github.com/repos/bridgedotnet/Bridge.Newtonsoft.Json
|
closed
|
DateTimeOffset deserialization fails from different time zone to deck server
|
defect
|
When I serialize and deserialize a new DateTimeOffset object on deck.net, I get the following error: `Argument_OffsetPrecision, Parameter name: offset`
I am working in the UK so my timezone is set to BST which is UTC +1 hour.
You can view the deck here:
https://deck.net/ba083a00c886571deec195adc6497b53
And the screenshot of the issue i'm experiencing if your timezone does not match mine and does not recreate issue:

Originally reported here:
https://github.com/bridgedotnet/Bridge.Newtonsoft.Json/issues/89
|
1.0
|
DateTimeOffset deserialization fails from different time zone to deck server - When I serialize and deserialize a new DateTimeOffset object on deck.net, I get the following error: `Argument_OffsetPrecision, Parameter name: offset`
I am working in the UK so my timezone is set to BST which is UTC +1 hour.
You can view the deck here:
https://deck.net/ba083a00c886571deec195adc6497b53
And the screenshot of the issue i'm experiencing if your timezone does not match mine and does not recreate issue:

Originally reported here:
https://github.com/bridgedotnet/Bridge.Newtonsoft.Json/issues/89
|
defect
|
datetimeoffset deserialization fails from different time zone to deck server when i serialize and deserialize a new datetimeoffset object on deck net i get the following error argument offsetprecision parameter name offset i am working in the uk so my timezone is set to bst which is utc hour you can view the deck here and the screenshot of the issue i m experiencing if your timezone does not match mine and does not recreate issue originally reported here
| 1
|
107,460
| 11,544,739,845
|
IssuesEvent
|
2020-02-18 12:05:29
|
trimox/angular-mdc-web
|
https://api.github.com/repos/trimox/angular-mdc-web
|
closed
|
demos: Use scrollTarget for top-app-bar elevation changes on scroll
|
documentation
|
Noticed the demo site is not changing elevation on scroll event.
|
1.0
|
demos: Use scrollTarget for top-app-bar elevation changes on scroll - Noticed the demo site is not changing elevation on scroll event.
|
non_defect
|
demos use scrolltarget for top app bar elevation changes on scroll noticed the demo site is not changing elevation on scroll event
| 0
|
48,796
| 13,184,743,048
|
IssuesEvent
|
2020-08-12 20:00:46
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
Too many back-to-back short runs in Pdaq cause PnF processing to stall (Trac #222)
|
Incomplete Migration Migrated from Trac defect jeb + pnf
|
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/222
, reported by blaufuss and owned by tschmidt_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-05-25T13:41:49",
"description": "Too many short pdaq runs (that actually produce a few events each) cause PnF processing to slow to a crawl. \n\nRun transitions are hard at PnF, with clients needing to request new GCD, etc, server needing to close open files, etc...\n\nNeed to manage this transition better. Options:\n\n1. Improve GCDispatch, so it caches several recent runs, so a client flopping between runs isn't a problem\n\n2. Make PFServer block data from new run until it finishes current run. May slow transitons a bit, but might be more robust.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1337953309000000",
"component": "jeb + pnf",
"summary": "Too many back-to-back short runs in Pdaq cause PnF processing to stall",
"priority": "normal",
"keywords": "",
"time": "2010-12-01T16:43:51",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Too many back-to-back short runs in Pdaq cause PnF processing to stall (Trac #222) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/222
, reported by blaufuss and owned by tschmidt_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-05-25T13:41:49",
"description": "Too many short pdaq runs (that actually produce a few events each) cause PnF processing to slow to a crawl. \n\nRun transitions are hard at PnF, with clients needing to request new GCD, etc, server needing to close open files, etc...\n\nNeed to manage this transition better. Options:\n\n1. Improve GCDispatch, so it caches several recent runs, so a client flopping between runs isn't a problem\n\n2. Make PFServer block data from new run until it finishes current run. May slow transitons a bit, but might be more robust.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1337953309000000",
"component": "jeb + pnf",
"summary": "Too many back-to-back short runs in Pdaq cause PnF processing to stall",
"priority": "normal",
"keywords": "",
"time": "2010-12-01T16:43:51",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
|
defect
|
too many back to back short runs in pdaq cause pnf processing to stall trac migrated from reported by blaufuss and owned by tschmidt json status closed changetime description too many short pdaq runs that actually produce a few events each cause pnf processing to slow to a crawl n nrun transitions are hard at pnf with clients needing to request new gcd etc server needing to close open files etc n nneed to manage this transition better options n improve gcdispatch so it caches several recent runs so a client flopping between runs isn t a problem n make pfserver block data from new run until it finishes current run may slow transitons a bit but might be more robust reporter blaufuss cc resolution fixed ts component jeb pnf summary too many back to back short runs in pdaq cause pnf processing to stall priority normal keywords time milestone owner tschmidt type defect
| 1
|
32,929
| 6,970,593,374
|
IssuesEvent
|
2017-12-11 10:51:46
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
p:calendar DateTimePicker error in javascript for mode="inline"
|
defect
|
I´m reopening an bug from your old issuetracker...
https://code.google.com/p/primefaces/issues/detail?id=6790&thanks=6790&ts=1397477544
for more information also read
http://stackoverflow.com/questions/23036356/bug-in-primefaces-calendar-timepicker
issues still exists in 5.1.19 and 5.2.6
|
1.0
|
p:calendar DateTimePicker error in javascript for mode="inline" - I´m reopening an bug from your old issuetracker...
https://code.google.com/p/primefaces/issues/detail?id=6790&thanks=6790&ts=1397477544
for more information also read
http://stackoverflow.com/questions/23036356/bug-in-primefaces-calendar-timepicker
issues still exists in 5.1.19 and 5.2.6
|
defect
|
p calendar datetimepicker error in javascript for mode inline i´m reopening an bug from your old issuetracker for more information also read issues still exists in and
| 1
|
447,709
| 12,892,074,755
|
IssuesEvent
|
2020-07-13 18:55:32
|
coyiutoc/Honda-Capstone-Salmon
|
https://api.github.com/repos/coyiutoc/Honda-Capstone-Salmon
|
opened
|
Order tags in search column by # of supporting evidence
|
high priority
|
Tags in search column need # of supporting evidence included + should be ordered from most to greatest.
e.g. [ Salmon (12) ] [ Tuna (10) ] [ Snapper (4) ]
|
1.0
|
Order tags in search column by # of supporting evidence - Tags in search column need # of supporting evidence included + should be ordered from most to greatest.
e.g. [ Salmon (12) ] [ Tuna (10) ] [ Snapper (4) ]
|
non_defect
|
order tags in search column by of supporting evidence tags in search column need of supporting evidence included should be ordered from most to greatest e g
| 0
|
67,679
| 21,051,814,942
|
IssuesEvent
|
2022-03-31 21:15:18
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
closed
|
[🐛 Bug]: InvocationTargetException while using dynamic Selenium grid 4
|
I-defect needs-triaging
|
### What happened?
Getting InvocationTargetException while using dynamic Selenium grid 4.
We have downloaded docker compose file (docker-compose-v3-dynamic-grid.yml) from https://github.com/SeleniumHQ/docker-selenium/blob/trunk/docker-compose-v3-dynamic-grid.yml and created config.toml file using the code from https://github.com/SeleniumHQ/docker-selenium#dynamic-grid without changing anything. The config.toml file is place in NodeDocker folder.
### How can we reproduce the issue?
```shell
selenium-compose.yml:
version: "3"
services:
node-docker:
image: selenium/node-docker:latest
volumes:
- ./assets:/opt/selenium/assets
- ./config/config.toml:/opt/bin/config.toml
depends_on:
- selenium-hub
environment:
- SE_EVENT_BUS_HOST=selenium-hub
- SE_EVENT_BUS_PUBLISH_PORT=4442
- SE_EVENT_BUS_SUBSCRIBE_PORT=4443
selenium-hub:
image: selenium/hub:4.1.3-20220327
container_name: selenium-hub
ports:
- "4442:4442"
- "4443:4443"
- "4444:4444"
config.toml
[docker]
# Configs have a mapping between the Docker image to use and the capabilities that need to be matched to
# start a container with the given image.
configs = [
"selenium/standalone-chrome:latest", "{\"browserName\": \"chrome\"}"
"selenium/standalone-firefox:latest", "{\"browserName\": \"firefox\"}"
]
# URL for connecting to the docker daemon
# Most simple approach, leave it as http://127.0.0.1:2375, and mount /var/run/docker.sock.
# 127.0.0.1 is used because interally the container uses socat when /var/run/docker.sock is mounted
# If var/run/docker.sock is not mounted:
# Windows: make sure Docker Desktop exposes the daemon via tcp, and use http://host.docker.internal:2375.
# macOS: install socat and run the following command, socat -4 TCP-LISTEN:2375,fork UNIX-CONNECT:/var/run/docker.sock,
# then use http://host.docker.internal:2375.
# Linux: varies from machine to machine, please mount /var/run/docker.sock. If this does not work, please create an issue.
url = "http://host.docker.internal:2375"
# Docker image used for video recording
video-image = "selenium/video:latest"
# Uncomment the following section if you are running the node on a separate VM
# Fill out the placeholders with appropriate values
#[server]
#host = <ip-from-node-machine>
#port = <port-from-node-machine>
Running below docker compose command:
docker-compose -f selenium-compose.yml up
```
### Relevant log output
```shell
selenium-hub | 2022-03-31 20:24:30,036 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-hub.conf" during parsing
selenium-hub | 2022-03-31 20:24:30,039 INFO supervisord started with pid 8
selenium-hub | 2022-03-31 20:24:31,041 INFO spawned: 'selenium-grid-hub' with pid 10
selenium-hub | 20:24:31.411 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
selenium-hub | 2022-03-31 20:24:31,413 INFO success: selenium-grid-hub entered RUNNING state, process has stayed up for > than 0 seconds (startsecs)
selenium-hub | 20:24:31.416 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
selenium-hub | 20:24:31.517 INFO [BoundZmqEventBus.<init>] - XPUB binding to [binding to tcp://*:4442, advertising as tcp://172.22.0.2:4442], XSUB binding to [binding to tcp://*:4443, advertising as tcp://172.22.0.2:4443]
selenium-hub | 20:24:31.567 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://172.22.0.2:4442 and tcp://172.22.0.2:4443
selenium-hub | 20:24:31.587 INFO [UnboundZmqEventBus.<init>] - Sockets created
selenium-hub | 20:24:32.589 INFO [UnboundZmqEventBus.<init>] - Event bus ready
selenium-hub | 20:24:33.198 INFO [Hub.execute] - Started Selenium Hub 4.1.3 (revision 7b1ebf28ef): http://172.22.0.2:4444
selenium-hub | Trapped SIGTERM/SIGINT/x so shutting down supervisord...
selenium-hub | 2022-03-31 20:31:13,705 WARN received SIGTERM indicating exit request
selenium-hub | 2022-03-31 20:31:13,706 INFO waiting for selenium-grid-hub to die
selenium-hub | 2022-03-31 20:31:14,707 INFO stopped: selenium-grid-hub (terminated by SIGTERM)
selenium-hub | Shutdown complete
selenium-hub | 2022-03-31 20:31:35,542 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-hub.conf" during parsing
selenium-hub | 2022-03-31 20:31:35,545 INFO supervisord started with pid 9
selenium-hub | 2022-03-31 20:31:36,547 INFO spawned: 'selenium-grid-hub' with pid 11
selenium-hub | 20:31:36.964 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
selenium-hub | 2022-03-31 20:31:36,968 INFO success: selenium-grid-hub entered RUNNING state, process has stayed up for > than 0 seconds (startsecs)
selenium-hub | 20:31:36.971 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
selenium-hub | 20:31:37.102 INFO [BoundZmqEventBus.<init>] - XPUB binding to [binding to tcp://*:4442, advertising as tcp://172.22.0.2:4442], XSUB binding to [binding to tcp://*:4443, advertising as tcp://172.22.0.2:4443]
selenium-hub | 20:31:37.158 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://172.22.0.2:4442 and tcp://172.22.0.2:4443
selenium-hub | 20:31:37.181 INFO [UnboundZmqEventBus.<init>] - Sockets created
selenium-hub | 20:31:38.183 INFO [UnboundZmqEventBus.<init>] - Event bus ready
selenium-hub | 20:31:38.936 INFO [Hub.execute] - Started Selenium Hub 4.1.3 (revision 7b1ebf28ef): http://172.22.0.2:4444
node-docker_1 | 2022-03-31 20:33:12,162 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-docker.conf" during parsing
node-docker_1 | 2022-03-31 20:33:12,165 INFO supervisord started with pid 8
node-docker_1 | 2022-03-31 20:33:13,167 INFO spawned: 'socat' with pid 10
node-docker_1 | 2022-03-31 20:33:13,169 INFO spawned: 'selenium-grid-docker' with pid 11
node-docker_1 | 2022-03-31 20:33:13,176 INFO success: socat entered RUNNING state, process has stayed up for > than 0 seconds (startsecs)
node-docker_1 | 2022-03-31 20:33:13,177 INFO success: selenium-grid-docker entered RUNNING state, process has stayed up for > than 0 seconds (startsecs)
node-docker_1 | 2022-03-31 20:33:13,177 INFO exited: socat (exit status 0; expected)
node-docker_1 | Starting Selenium Grid Node Docker...
node-docker_1 | 20:33:13.656 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
node-docker_1 | 20:33:13.663 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
node-docker_1 | 20:33:13.831 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://selenium-hub:4442 and tcp://selenium-hub:4443
node-docker_1 | 20:33:13.889 INFO [UnboundZmqEventBus.<init>] - Sockets created
node-docker_1 | 20:33:14.892 INFO [UnboundZmqEventBus.<init>] - Event bus ready
node-docker_1 | 20:33:15.016 INFO [NodeServer.createHandlers] - Reporting self as: http://172.22.0.3:5555
node-docker_1 | 20:33:15.032 INFO [NodeOptions.getSessionFactories] - Detected 12 available processors
node-docker_1 | java.lang.reflect.InvocationTargetException
node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
node-docker_1 | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
node-docker_1 | at java.base/java.lang.reflect.Method.invoke(Method.java:566)
node-docker_1 | at org.openqa.selenium.grid.Bootstrap.runMain(Bootstrap.java:77)
node-docker_1 | at org.openqa.selenium.grid.Bootstrap.main(Bootstrap.java:70)
node-docker_1 | Caused by: org.openqa.selenium.grid.config.ConfigException: java.lang.reflect.InvocationTargetException
node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.getClass(MemoizedConfig.java:115)
node-docker_1 | at org.openqa.selenium.grid.node.config.NodeOptions.getNode(NodeOptions.java:149)
node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.createHandlers(NodeServer.java:127)
node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.asServer(NodeServer.java:183)
node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.execute(NodeServer.java:231)
node-docker_1 | at org.openqa.selenium.grid.TemplateGridCommand.lambda$configure$4(TemplateGridCommand.java:129)
node-docker_1 | at org.openqa.selenium.grid.Main.launch(Main.java:83)
node-docker_1 | at org.openqa.selenium.grid.Main.go(Main.java:57)
node-docker_1 | at org.openqa.selenium.grid.Main.main(Main.java:42)
node-docker_1 | ... 6 more
node-docker_1 | Caused by: java.lang.reflect.InvocationTargetException
node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
node-docker_1 | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
node-docker_1 | at java.base/java.lang.reflect.Method.invoke(Method.java:566)
node-docker_1 | at org.openqa.selenium.grid.config.ClassCreation.callCreateMethod(ClassCreation.java:50)
node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.lambda$getClass$4(MemoizedConfig.java:100)
node-docker_1 | at java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1737)
node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.getClass(MemoizedConfig.java:95)
node-docker_1 | ... 14 more
node-docker_1 | Caused by: org.openqa.selenium.docker.DockerException: Unable to reach the Docker daemon at http://127.0.0.1:2375
node-docker_1 | at org.openqa.selenium.grid.node.docker.DockerOptions.getDockerSessionFactories(DockerOptions.java:125)
node-docker_1 | at org.openqa.selenium.grid.node.local.LocalNodeFactory.create(LocalNodeFactory.java:80)
node-docker_1 | ... 22 more
node-docker_1 | Exception in thread "Thread-0" java.lang.NullPointerException
node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.lambda$new$0(NodeServer.java:79)
node-docker_1 | at java.base/java.lang.Thread.run(Thread.java:829)
node-docker_1 | 2022-03-31 20:33:15,858 INFO exited: selenium-grid-docker (exit status 1; not expected)
```
### Operating System
Windows 10
### Selenium version
Java 4.1.3
### What are the browser(s) and version(s) where you see this issue?
dynamic grid using docker
### What are the browser driver(s) and version(s) where you see this issue?
selenium/node-docker:latest, selenium/standalone-chrome:latest, selenium/standalone-firefox:latest
### Are you using Selenium Grid?
4
|
1.0
|
[🐛 Bug]: InvocationTargetException while using dynamic Selenium grid 4 - ### What happened?
Getting InvocationTargetException while using dynamic Selenium grid 4.
We have downloaded docker compose file (docker-compose-v3-dynamic-grid.yml) from https://github.com/SeleniumHQ/docker-selenium/blob/trunk/docker-compose-v3-dynamic-grid.yml and created config.toml file using the code from https://github.com/SeleniumHQ/docker-selenium#dynamic-grid without changing anything. The config.toml file is place in NodeDocker folder.
### How can we reproduce the issue?
```shell
selenium-compose.yml:
version: "3"
services:
node-docker:
image: selenium/node-docker:latest
volumes:
- ./assets:/opt/selenium/assets
- ./config/config.toml:/opt/bin/config.toml
depends_on:
- selenium-hub
environment:
- SE_EVENT_BUS_HOST=selenium-hub
- SE_EVENT_BUS_PUBLISH_PORT=4442
- SE_EVENT_BUS_SUBSCRIBE_PORT=4443
selenium-hub:
image: selenium/hub:4.1.3-20220327
container_name: selenium-hub
ports:
- "4442:4442"
- "4443:4443"
- "4444:4444"
config.toml
[docker]
# Configs have a mapping between the Docker image to use and the capabilities that need to be matched to
# start a container with the given image.
configs = [
"selenium/standalone-chrome:latest", "{\"browserName\": \"chrome\"}"
"selenium/standalone-firefox:latest", "{\"browserName\": \"firefox\"}"
]
# URL for connecting to the docker daemon
# Most simple approach, leave it as http://127.0.0.1:2375, and mount /var/run/docker.sock.
# 127.0.0.1 is used because interally the container uses socat when /var/run/docker.sock is mounted
# If var/run/docker.sock is not mounted:
# Windows: make sure Docker Desktop exposes the daemon via tcp, and use http://host.docker.internal:2375.
# macOS: install socat and run the following command, socat -4 TCP-LISTEN:2375,fork UNIX-CONNECT:/var/run/docker.sock,
# then use http://host.docker.internal:2375.
# Linux: varies from machine to machine, please mount /var/run/docker.sock. If this does not work, please create an issue.
url = "http://host.docker.internal:2375"
# Docker image used for video recording
video-image = "selenium/video:latest"
# Uncomment the following section if you are running the node on a separate VM
# Fill out the placeholders with appropriate values
#[server]
#host = <ip-from-node-machine>
#port = <port-from-node-machine>
Running below docker compose command:
docker-compose -f selenium-compose.yml up
```
### Relevant log output
```shell
selenium-hub | 2022-03-31 20:24:30,036 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-hub.conf" during parsing
selenium-hub | 2022-03-31 20:24:30,039 INFO supervisord started with pid 8
selenium-hub | 2022-03-31 20:24:31,041 INFO spawned: 'selenium-grid-hub' with pid 10
selenium-hub | 20:24:31.411 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
selenium-hub | 2022-03-31 20:24:31,413 INFO success: selenium-grid-hub entered RUNNING state, process has stayed up for > than 0 seconds (startsecs)
selenium-hub | 20:24:31.416 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
selenium-hub | 20:24:31.517 INFO [BoundZmqEventBus.<init>] - XPUB binding to [binding to tcp://*:4442, advertising as tcp://172.22.0.2:4442], XSUB binding to [binding to tcp://*:4443, advertising as tcp://172.22.0.2:4443]
selenium-hub | 20:24:31.567 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://172.22.0.2:4442 and tcp://172.22.0.2:4443
selenium-hub | 20:24:31.587 INFO [UnboundZmqEventBus.<init>] - Sockets created
selenium-hub | 20:24:32.589 INFO [UnboundZmqEventBus.<init>] - Event bus ready
selenium-hub | 20:24:33.198 INFO [Hub.execute] - Started Selenium Hub 4.1.3 (revision 7b1ebf28ef): http://172.22.0.2:4444
selenium-hub | Trapped SIGTERM/SIGINT/x so shutting down supervisord...
selenium-hub | 2022-03-31 20:31:13,705 WARN received SIGTERM indicating exit request
selenium-hub | 2022-03-31 20:31:13,706 INFO waiting for selenium-grid-hub to die
selenium-hub | 2022-03-31 20:31:14,707 INFO stopped: selenium-grid-hub (terminated by SIGTERM)
selenium-hub | Shutdown complete
selenium-hub | 2022-03-31 20:31:35,542 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-hub.conf" during parsing
selenium-hub | 2022-03-31 20:31:35,545 INFO supervisord started with pid 9
selenium-hub | 2022-03-31 20:31:36,547 INFO spawned: 'selenium-grid-hub' with pid 11
selenium-hub | 20:31:36.964 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
selenium-hub | 2022-03-31 20:31:36,968 INFO success: selenium-grid-hub entered RUNNING state, process has stayed up for > than 0 seconds (startsecs)
selenium-hub | 20:31:36.971 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
selenium-hub | 20:31:37.102 INFO [BoundZmqEventBus.<init>] - XPUB binding to [binding to tcp://*:4442, advertising as tcp://172.22.0.2:4442], XSUB binding to [binding to tcp://*:4443, advertising as tcp://172.22.0.2:4443]
selenium-hub | 20:31:37.158 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://172.22.0.2:4442 and tcp://172.22.0.2:4443
selenium-hub | 20:31:37.181 INFO [UnboundZmqEventBus.<init>] - Sockets created
selenium-hub | 20:31:38.183 INFO [UnboundZmqEventBus.<init>] - Event bus ready
selenium-hub | 20:31:38.936 INFO [Hub.execute] - Started Selenium Hub 4.1.3 (revision 7b1ebf28ef): http://172.22.0.2:4444
node-docker_1 | 2022-03-31 20:33:12,162 INFO Included extra file "/etc/supervisor/conf.d/selenium-grid-docker.conf" during parsing
node-docker_1 | 2022-03-31 20:33:12,165 INFO supervisord started with pid 8
node-docker_1 | 2022-03-31 20:33:13,167 INFO spawned: 'socat' with pid 10
node-docker_1 | 2022-03-31 20:33:13,169 INFO spawned: 'selenium-grid-docker' with pid 11
node-docker_1 | 2022-03-31 20:33:13,176 INFO success: socat entered RUNNING state, process has stayed up for > than 0 seconds (startsecs)
node-docker_1 | 2022-03-31 20:33:13,177 INFO success: selenium-grid-docker entered RUNNING state, process has stayed up for > than 0 seconds (startsecs)
node-docker_1 | 2022-03-31 20:33:13,177 INFO exited: socat (exit status 0; expected)
node-docker_1 | Starting Selenium Grid Node Docker...
node-docker_1 | 20:33:13.656 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
node-docker_1 | 20:33:13.663 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
node-docker_1 | 20:33:13.831 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://selenium-hub:4442 and tcp://selenium-hub:4443
node-docker_1 | 20:33:13.889 INFO [UnboundZmqEventBus.<init>] - Sockets created
node-docker_1 | 20:33:14.892 INFO [UnboundZmqEventBus.<init>] - Event bus ready
node-docker_1 | 20:33:15.016 INFO [NodeServer.createHandlers] - Reporting self as: http://172.22.0.3:5555
node-docker_1 | 20:33:15.032 INFO [NodeOptions.getSessionFactories] - Detected 12 available processors
node-docker_1 | java.lang.reflect.InvocationTargetException
node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
node-docker_1 | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
node-docker_1 | at java.base/java.lang.reflect.Method.invoke(Method.java:566)
node-docker_1 | at org.openqa.selenium.grid.Bootstrap.runMain(Bootstrap.java:77)
node-docker_1 | at org.openqa.selenium.grid.Bootstrap.main(Bootstrap.java:70)
node-docker_1 | Caused by: org.openqa.selenium.grid.config.ConfigException: java.lang.reflect.InvocationTargetException
node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.getClass(MemoizedConfig.java:115)
node-docker_1 | at org.openqa.selenium.grid.node.config.NodeOptions.getNode(NodeOptions.java:149)
node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.createHandlers(NodeServer.java:127)
node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.asServer(NodeServer.java:183)
node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.execute(NodeServer.java:231)
node-docker_1 | at org.openqa.selenium.grid.TemplateGridCommand.lambda$configure$4(TemplateGridCommand.java:129)
node-docker_1 | at org.openqa.selenium.grid.Main.launch(Main.java:83)
node-docker_1 | at org.openqa.selenium.grid.Main.go(Main.java:57)
node-docker_1 | at org.openqa.selenium.grid.Main.main(Main.java:42)
node-docker_1 | ... 6 more
node-docker_1 | Caused by: java.lang.reflect.InvocationTargetException
node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
node-docker_1 | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
node-docker_1 | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
node-docker_1 | at java.base/java.lang.reflect.Method.invoke(Method.java:566)
node-docker_1 | at org.openqa.selenium.grid.config.ClassCreation.callCreateMethod(ClassCreation.java:50)
node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.lambda$getClass$4(MemoizedConfig.java:100)
node-docker_1 | at java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1737)
node-docker_1 | at org.openqa.selenium.grid.config.MemoizedConfig.getClass(MemoizedConfig.java:95)
node-docker_1 | ... 14 more
node-docker_1 | Caused by: org.openqa.selenium.docker.DockerException: Unable to reach the Docker daemon at http://127.0.0.1:2375
node-docker_1 | at org.openqa.selenium.grid.node.docker.DockerOptions.getDockerSessionFactories(DockerOptions.java:125)
node-docker_1 | at org.openqa.selenium.grid.node.local.LocalNodeFactory.create(LocalNodeFactory.java:80)
node-docker_1 | ... 22 more
node-docker_1 | Exception in thread "Thread-0" java.lang.NullPointerException
node-docker_1 | at org.openqa.selenium.grid.node.httpd.NodeServer.lambda$new$0(NodeServer.java:79)
node-docker_1 | at java.base/java.lang.Thread.run(Thread.java:829)
node-docker_1 | 2022-03-31 20:33:15,858 INFO exited: selenium-grid-docker (exit status 1; not expected)
```
### Operating System
Windows 10
### Selenium version
Java 4.1.3
### What are the browser(s) and version(s) where you see this issue?
dynamic grid using docker
### What are the browser driver(s) and version(s) where you see this issue?
selenium/node-docker:latest, selenium/standalone-chrome:latest, selenium/standalone-firefox:latest
### Are you using Selenium Grid?
4
|
defect
|
invocationtargetexception while using dynamic selenium grid what happened getting invocationtargetexception while using dynamic selenium grid we have downloaded docker compose file docker compose dynamic grid yml from and created config toml file using the code from without changing anything the config toml file is place in nodedocker folder how can we reproduce the issue shell selenium compose yml version services node docker image selenium node docker latest volumes assets opt selenium assets config config toml opt bin config toml depends on selenium hub environment se event bus host selenium hub se event bus publish port se event bus subscribe port selenium hub image selenium hub container name selenium hub ports config toml configs have a mapping between the docker image to use and the capabilities that need to be matched to start a container with the given image configs selenium standalone chrome latest browsername chrome selenium standalone firefox latest browsername firefox url for connecting to the docker daemon most simple approach leave it as and mount var run docker sock is used because interally the container uses socat when var run docker sock is mounted if var run docker sock is not mounted windows make sure docker desktop exposes the daemon via tcp and use macos install socat and run the following command socat tcp listen fork unix connect var run docker sock then use linux varies from machine to machine please mount var run docker sock if this does not work please create an issue url docker image used for video recording video image selenium video latest uncomment the following section if you are running the node on a separate vm fill out the placeholders with appropriate values host port running below docker compose command docker compose f selenium compose yml up relevant log output shell selenium hub info included extra file etc supervisor conf d selenium grid hub conf during parsing selenium hub info supervisord started with pid selenium hub info spawned selenium grid hub with pid selenium hub info using the system default encoding selenium hub info success selenium grid hub entered running state process has stayed up for than seconds startsecs selenium hub info using opentelemetry for tracing selenium hub info xpub binding to xsub binding to selenium hub info connecting to tcp and tcp selenium hub info sockets created selenium hub info event bus ready selenium hub info started selenium hub revision selenium hub trapped sigterm sigint x so shutting down supervisord selenium hub warn received sigterm indicating exit request selenium hub info waiting for selenium grid hub to die selenium hub info stopped selenium grid hub terminated by sigterm selenium hub shutdown complete selenium hub info included extra file etc supervisor conf d selenium grid hub conf during parsing selenium hub info supervisord started with pid selenium hub info spawned selenium grid hub with pid selenium hub info using the system default encoding selenium hub info success selenium grid hub entered running state process has stayed up for than seconds startsecs selenium hub info using opentelemetry for tracing selenium hub info xpub binding to xsub binding to selenium hub info connecting to tcp and tcp selenium hub info sockets created selenium hub info event bus ready selenium hub info started selenium hub revision node docker info included extra file etc supervisor conf d selenium grid docker conf during parsing node docker info supervisord started with pid node docker info spawned socat with pid node docker info spawned selenium grid docker with pid node docker info success socat entered running state process has stayed up for than seconds startsecs node docker info success selenium grid docker entered running state process has stayed up for than seconds startsecs node docker info exited socat exit status expected node docker starting selenium grid node docker node docker info using the system default encoding node docker info using opentelemetry for tracing node docker info connecting to tcp selenium hub and tcp selenium hub node docker info sockets created node docker info event bus ready node docker info reporting self as node docker info detected available processors node docker java lang reflect invocationtargetexception node docker at java base jdk internal reflect nativemethodaccessorimpl native method node docker at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java node docker at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java node docker at java base java lang reflect method invoke method java node docker at org openqa selenium grid bootstrap runmain bootstrap java node docker at org openqa selenium grid bootstrap main bootstrap java node docker caused by org openqa selenium grid config configexception java lang reflect invocationtargetexception node docker at org openqa selenium grid config memoizedconfig getclass memoizedconfig java node docker at org openqa selenium grid node config nodeoptions getnode nodeoptions java node docker at org openqa selenium grid node httpd nodeserver createhandlers nodeserver java node docker at org openqa selenium grid node httpd nodeserver asserver nodeserver java node docker at org openqa selenium grid node httpd nodeserver execute nodeserver java node docker at org openqa selenium grid templategridcommand lambda configure templategridcommand java node docker at org openqa selenium grid main launch main java node docker at org openqa selenium grid main go main java node docker at org openqa selenium grid main main main java node docker more node docker caused by java lang reflect invocationtargetexception node docker at java base jdk internal reflect nativemethodaccessorimpl native method node docker at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java node docker at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java node docker at java base java lang reflect method invoke method java node docker at org openqa selenium grid config classcreation callcreatemethod classcreation java node docker at org openqa selenium grid config memoizedconfig lambda getclass memoizedconfig java node docker at java base java util concurrent concurrenthashmap computeifabsent concurrenthashmap java node docker at org openqa selenium grid config memoizedconfig getclass memoizedconfig java node docker more node docker caused by org openqa selenium docker dockerexception unable to reach the docker daemon at node docker at org openqa selenium grid node docker dockeroptions getdockersessionfactories dockeroptions java node docker at org openqa selenium grid node local localnodefactory create localnodefactory java node docker more node docker exception in thread thread java lang nullpointerexception node docker at org openqa selenium grid node httpd nodeserver lambda new nodeserver java node docker at java base java lang thread run thread java node docker info exited selenium grid docker exit status not expected operating system windows selenium version java what are the browser s and version s where you see this issue dynamic grid using docker what are the browser driver s and version s where you see this issue selenium node docker latest selenium standalone chrome latest selenium standalone firefox latest are you using selenium grid
| 1
|
106,414
| 13,284,191,262
|
IssuesEvent
|
2020-08-24 05:40:35
|
Qiskit/qiskit.org
|
https://api.github.com/repos/Qiskit/qiskit.org
|
closed
|
Compare Marina and Human design work
|
Design
|
Design kickoff session with human to compare and align on work to be done
|
1.0
|
Compare Marina and Human design work - Design kickoff session with human to compare and align on work to be done
|
non_defect
|
compare marina and human design work design kickoff session with human to compare and align on work to be done
| 0
|
6,587
| 2,610,257,312
|
IssuesEvent
|
2015-02-26 19:22:10
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳激光去痤疮有副作用吗
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳激光去痤疮有副作用吗【深圳韩方科颜全国热线400-869-181
8,24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构��
�韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳�
��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不
反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创��
�内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客�
��上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:40
|
1.0
|
深圳激光去痤疮有副作用吗 - ```
深圳激光去痤疮有副作用吗【深圳韩方科颜全国热线400-869-181
8,24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构��
�韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳�
��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不
反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创��
�内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客�
��上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:40
|
defect
|
深圳激光去痤疮有副作用吗 深圳激光去痤疮有副作用吗【 , 】深圳韩方科颜专业祛痘连锁机构,机构�� �韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳� ��,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不 反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创�� �内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客� ��上的痘痘。 original issue reported on code google com by szft com on may at
| 1
|
133,191
| 28,761,071,044
|
IssuesEvent
|
2023-05-01 00:44:39
|
muradshahsuvarov/employee-commute-route-optimizer
|
https://api.github.com/repos/muradshahsuvarov/employee-commute-route-optimizer
|
closed
|
Implement multiple location route finder algorithm
|
code
|
Let N be the number of commute locations. Using the route finder, find the shortest route considering traffic, and etc.
|
1.0
|
Implement multiple location route finder algorithm - Let N be the number of commute locations. Using the route finder, find the shortest route considering traffic, and etc.
|
non_defect
|
implement multiple location route finder algorithm let n be the number of commute locations using the route finder find the shortest route considering traffic and etc
| 0
|
50,348
| 13,187,454,304
|
IssuesEvent
|
2020-08-13 03:28:00
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters (Trac #565)
|
Migrated from Trac defect tools/ports
|
There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.
A patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/565
, reported by kislat and owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2009-06-30T14:16:42",
"description": "There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.\n\nA patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.",
"reporter": "kislat",
"cc": "",
"resolution": "fixed",
"_ts": "1246371402000000",
"component": "tools/ports",
"summary": "rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters",
"priority": "normal",
"keywords": "",
"time": "2009-06-30T12:29:54",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters (Trac #565) - There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.
A patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/565
, reported by kislat and owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2009-06-30T14:16:42",
"description": "There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.\n\nA patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.",
"reporter": "kislat",
"cc": "",
"resolution": "fixed",
"_ts": "1246371402000000",
"component": "tools/ports",
"summary": "rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters",
"priority": "normal",
"keywords": "",
"time": "2009-06-30T12:29:54",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
|
defect
|
rootcint won t create dictionaries if rootsys is longer than characters trac there is a bug in rootcint root and later versions that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable after resolving symlinks is longer than characters a patch for the file root source file core utils src rootcint cxx for root is attached migrated from reported by kislat and owned by cgils json status closed changetime description there is a bug in rootcint root and later versions that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable after resolving symlinks is longer than characters n na patch for the file root source file core utils src rootcint cxx for root is attached reporter kislat cc resolution fixed ts component tools ports summary rootcint won t create dictionaries if rootsys is longer than characters priority normal keywords time milestone owner cgils type defect
| 1
|
114,045
| 9,672,687,759
|
IssuesEvent
|
2019-05-22 04:46:13
|
ansible/ansible
|
https://api.github.com/repos/ansible/ansible
|
closed
|
mysql_user should flush privileges when changed
|
affects_2.3 bug database module mysql support:community test waiting_on_contributor
|
From @lordnynex on 2015-10-26T22:48:07Z
##### Issue Type:
Bug Report
##### Component Name:
mysql_user module
##### Ansible Version:
Any version using mysql_user
##### Ansible Configuration:
N/A
##### Environment:
N/A
##### Summary:
mysql_user should flush privileges when changed. Currently requires extra task to reload mysql or manually flush privileges.
##### Steps To Reproduce:
``` yaml
- mysql_user: name=root password=root priv=*.*:ALL state=present check_implicit_admin=yes
```
##### Expected Results:
Changed credentials should be usable in subsequent tasks without a mysql reload. Ex
``` yaml
- shell: mysqladmin -u root password root
```
##### Actual Results:
Tasks/commands attempting to use the modified credentials will fail execution.
Copied from original issue: ansible/ansible-modules-core#2357
|
1.0
|
mysql_user should flush privileges when changed - From @lordnynex on 2015-10-26T22:48:07Z
##### Issue Type:
Bug Report
##### Component Name:
mysql_user module
##### Ansible Version:
Any version using mysql_user
##### Ansible Configuration:
N/A
##### Environment:
N/A
##### Summary:
mysql_user should flush privileges when changed. Currently requires extra task to reload mysql or manually flush privileges.
##### Steps To Reproduce:
``` yaml
- mysql_user: name=root password=root priv=*.*:ALL state=present check_implicit_admin=yes
```
##### Expected Results:
Changed credentials should be usable in subsequent tasks without a mysql reload. Ex
``` yaml
- shell: mysqladmin -u root password root
```
##### Actual Results:
Tasks/commands attempting to use the modified credentials will fail execution.
Copied from original issue: ansible/ansible-modules-core#2357
|
non_defect
|
mysql user should flush privileges when changed from lordnynex on issue type bug report component name mysql user module ansible version any version using mysql user ansible configuration n a environment n a summary mysql user should flush privileges when changed currently requires extra task to reload mysql or manually flush privileges steps to reproduce yaml mysql user name root password root priv all state present check implicit admin yes expected results changed credentials should be usable in subsequent tasks without a mysql reload ex yaml shell mysqladmin u root password root actual results tasks commands attempting to use the modified credentials will fail execution copied from original issue ansible ansible modules core
| 0
|
65,974
| 19,842,995,580
|
IssuesEvent
|
2022-01-21 00:45:24
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
opened
|
Mentioning a user with emojis in the display name causes Element Android to 'mention' a few times
|
T-Defect
|
### Steps to reproduce
1. Mention a person with emojis in their username
2. Send it
### Outcome
#### What did you expect?
Mentions once
#### What happened instead?
It 'mentions' a few times, though in other Matrix clients it appears as one


### Your phone model
Mito W1
### Operating system version
Android 9
### Application version and app store
1.3.15 [40103152] (G-b5583)
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Mentioning a user with emojis in the display name causes Element Android to 'mention' a few times - ### Steps to reproduce
1. Mention a person with emojis in their username
2. Send it
### Outcome
#### What did you expect?
Mentions once
#### What happened instead?
It 'mentions' a few times, though in other Matrix clients it appears as one


### Your phone model
Mito W1
### Operating system version
Android 9
### Application version and app store
1.3.15 [40103152] (G-b5583)
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
mentioning a user with emojis in the display name causes element android to mention a few times steps to reproduce mention a person with emojis in their username send it outcome what did you expect mentions once what happened instead it mentions a few times though in other matrix clients it appears as one your phone model mito operating system version android application version and app store g homeserver no response will you send logs no
| 1
|
79,396
| 28,147,604,437
|
IssuesEvent
|
2023-04-02 17:09:39
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
BUG: <Please write a comprehensive title after the 'BUG: ' prefix>
|
defect
|
### Describe your issue.
Issue:
'keep_feasible=True' in the 'scipy.optimize.NonlinearConstraint' requires 'x0' in 'scipy.optimize.minimize' to meet the constraints. This is so unreasonable for users.
Besides, sometimes, it also can't converge to the right results, while ''keep_feasible=False' can.
However, if I set ''keep_feasible=False', it sometimes converges to the point of minimum objective but violating constraints.
Analysis:
The reason of the failure is quite obvious, because once we set 'keep_feasible=True' , the optimization process doesn't accept the "infeasible" trials to compute the finte difference to let the process go further.
Improvement suggestion:
Why can't we make the process go as "keep_feasible=False", so that it can use the "infeasible" trials to converge to the final feasible result, but only accept the "feasible" trials as the valid steps towards the final result.
We should also set a number limit to the **consecutive** "infeasible" trials so that it can come back to the last valid step and make it as the final result or continue trials in different directions or step-sizes.
### Reproducing Code Example
```python
cons=NonlinearConstraint(fun_con, 0+1e-6, np.inf, keep_feasible=True) #require x0 to fulfil cons
res = opt.minimize(obj, x0, #method = 'SLSQP', #'COBYLA'(fail), #'SLSQP'(default)
constraints=cons,
bounds=bounds,
tol=1e-4)
```
### Error message
```shell
Exception has occurred: ValueError
`x0` is infeasible with respect to some inequality constraint with `keep_feasible` set to True.
```
### SciPy/NumPy/Python version and system information
```shell
{
"Compilers": {
"c": {
"name": "gcc",
"linker": "ld.bfd",
"version": "10.3.0",
"commands": "cc"
},
"cython": {
"name": "cython",
"linker": "cython",
"version": "0.29.33",
"commands": "cython"
},
"c++": {
"name": "gcc",
"linker": "ld.bfd",
"version": "10.3.0",
"commands": "c++"
},
"fortran": {
"name": "gcc",
"linker": "ld.bfd",
"version": "10.3.0",
"commands": "gfortran"
},
"pythran": {
"version": "0.12.1",
"include directory": "C:\\Users\\runneradmin\\AppData\\Local\\Temp\\pip-build-env-u63ta2f1\\overlay\\Lib\\site-packages\\pythran"
}
},
"Machine Information": {
"host": {
"cpu": "x86_64",
"family": "x86_64",
"endian": "little",
"system": "windows"
},
"build": {
"cpu": "x86_64",
"family": "x86_64",
"endian": "little",
"system": "windows"
},
"cross-compiled": false
},
"Build Dependencies": {
"blas": {
"name": "openblas",
"found": true,
"version": "0.3.18",
"detection method": "pkgconfig",
"include directory": "c:/opt/openblas/if_32/64/include",
"lib directory": "c:/opt/openblas/if_32/64/lib",
"openblas configuration": "USE_64BITINT= DYNAMIC_ARCH=1 DYNAMIC_OLDER= NO_CBLAS= NO_LAPACK= NO_LAPACKE= NO_AFFINITY=1 USE_OPENMP= PRESCOTT MAX_THREADS=4",
"pc file directory": "c:/opt/openblas/if_32/64/lib/pkgconfig"
},
"lapack": {
"name": "openblas",
"found": true,
"version": "0.3.18",
"detection method": "pkgconfig",
"include directory": "c:/opt/openblas/if_32/64/include",
"lib directory": "c:/opt/openblas/if_32/64/lib",
"openblas configuration": "USE_64BITINT= DYNAMIC_ARCH=1 DYNAMIC_OLDER= NO_CBLAS= NO_LAPACK= NO_LAPACKE= NO_AFFINITY=1 USE_OPENMP= PRESCOTT MAX_THREADS=4",
"pc file directory": "c:/opt/openblas/if_32/64/lib/pkgconfig"
}
},
"Python Information": {
"path": "C:\\Users\\runneradmin\\AppData\\Local\\Temp\\cibw-run-a1px0t3e\\cp39-win_amd64\\build\\venv\\Scripts\\python.exe",
"version": "3.9"
}
}
```
|
1.0
|
BUG: <Please write a comprehensive title after the 'BUG: ' prefix> - ### Describe your issue.
Issue:
'keep_feasible=True' in the 'scipy.optimize.NonlinearConstraint' requires 'x0' in 'scipy.optimize.minimize' to meet the constraints. This is so unreasonable for users.
Besides, sometimes, it also can't converge to the right results, while ''keep_feasible=False' can.
However, if I set ''keep_feasible=False', it sometimes converges to the point of minimum objective but violating constraints.
Analysis:
The reason of the failure is quite obvious, because once we set 'keep_feasible=True' , the optimization process doesn't accept the "infeasible" trials to compute the finte difference to let the process go further.
Improvement suggestion:
Why can't we make the process go as "keep_feasible=False", so that it can use the "infeasible" trials to converge to the final feasible result, but only accept the "feasible" trials as the valid steps towards the final result.
We should also set a number limit to the **consecutive** "infeasible" trials so that it can come back to the last valid step and make it as the final result or continue trials in different directions or step-sizes.
### Reproducing Code Example
```python
cons=NonlinearConstraint(fun_con, 0+1e-6, np.inf, keep_feasible=True) #require x0 to fulfil cons
res = opt.minimize(obj, x0, #method = 'SLSQP', #'COBYLA'(fail), #'SLSQP'(default)
constraints=cons,
bounds=bounds,
tol=1e-4)
```
### Error message
```shell
Exception has occurred: ValueError
`x0` is infeasible with respect to some inequality constraint with `keep_feasible` set to True.
```
### SciPy/NumPy/Python version and system information
```shell
{
"Compilers": {
"c": {
"name": "gcc",
"linker": "ld.bfd",
"version": "10.3.0",
"commands": "cc"
},
"cython": {
"name": "cython",
"linker": "cython",
"version": "0.29.33",
"commands": "cython"
},
"c++": {
"name": "gcc",
"linker": "ld.bfd",
"version": "10.3.0",
"commands": "c++"
},
"fortran": {
"name": "gcc",
"linker": "ld.bfd",
"version": "10.3.0",
"commands": "gfortran"
},
"pythran": {
"version": "0.12.1",
"include directory": "C:\\Users\\runneradmin\\AppData\\Local\\Temp\\pip-build-env-u63ta2f1\\overlay\\Lib\\site-packages\\pythran"
}
},
"Machine Information": {
"host": {
"cpu": "x86_64",
"family": "x86_64",
"endian": "little",
"system": "windows"
},
"build": {
"cpu": "x86_64",
"family": "x86_64",
"endian": "little",
"system": "windows"
},
"cross-compiled": false
},
"Build Dependencies": {
"blas": {
"name": "openblas",
"found": true,
"version": "0.3.18",
"detection method": "pkgconfig",
"include directory": "c:/opt/openblas/if_32/64/include",
"lib directory": "c:/opt/openblas/if_32/64/lib",
"openblas configuration": "USE_64BITINT= DYNAMIC_ARCH=1 DYNAMIC_OLDER= NO_CBLAS= NO_LAPACK= NO_LAPACKE= NO_AFFINITY=1 USE_OPENMP= PRESCOTT MAX_THREADS=4",
"pc file directory": "c:/opt/openblas/if_32/64/lib/pkgconfig"
},
"lapack": {
"name": "openblas",
"found": true,
"version": "0.3.18",
"detection method": "pkgconfig",
"include directory": "c:/opt/openblas/if_32/64/include",
"lib directory": "c:/opt/openblas/if_32/64/lib",
"openblas configuration": "USE_64BITINT= DYNAMIC_ARCH=1 DYNAMIC_OLDER= NO_CBLAS= NO_LAPACK= NO_LAPACKE= NO_AFFINITY=1 USE_OPENMP= PRESCOTT MAX_THREADS=4",
"pc file directory": "c:/opt/openblas/if_32/64/lib/pkgconfig"
}
},
"Python Information": {
"path": "C:\\Users\\runneradmin\\AppData\\Local\\Temp\\cibw-run-a1px0t3e\\cp39-win_amd64\\build\\venv\\Scripts\\python.exe",
"version": "3.9"
}
}
```
|
defect
|
bug describe your issue issue keep feasible true in the scipy optimize nonlinearconstraint requires in scipy optimize minimize to meet the constraints this is so unreasonable for users besides sometimes it also can t converge to the right results while keep feasible false can however if i set keep feasible false it sometimes converges to the point of minimum objective but violating constraints analysis the reason of the failure is quite obvious because once we set keep feasible true the optimization process doesn t accept the infeasible trials to compute the finte difference to let the process go further improvement suggestion why can t we make the process go as keep feasible false so that it can use the infeasible trials to converge to the final feasible result but only accept the feasible trials as the valid steps towards the final result we should also set a number limit to the consecutive infeasible trials so that it can come back to the last valid step and make it as the final result or continue trials in different directions or step sizes reproducing code example python cons nonlinearconstraint fun con np inf keep feasible true require to fulfil cons res opt minimize obj method slsqp cobyla fail slsqp default constraints cons bounds bounds tol error message shell exception has occurred valueerror is infeasible with respect to some inequality constraint with keep feasible set to true scipy numpy python version and system information shell compilers c name gcc linker ld bfd version commands cc cython name cython linker cython version commands cython c name gcc linker ld bfd version commands c fortran name gcc linker ld bfd version commands gfortran pythran version include directory c users runneradmin appdata local temp pip build env overlay lib site packages pythran machine information host cpu family endian little system windows build cpu family endian little system windows cross compiled false build dependencies blas name openblas found true version detection method pkgconfig include directory c opt openblas if include lib directory c opt openblas if lib openblas configuration use dynamic arch dynamic older no cblas no lapack no lapacke no affinity use openmp prescott max threads pc file directory c opt openblas if lib pkgconfig lapack name openblas found true version detection method pkgconfig include directory c opt openblas if include lib directory c opt openblas if lib openblas configuration use dynamic arch dynamic older no cblas no lapack no lapacke no affinity use openmp prescott max threads pc file directory c opt openblas if lib pkgconfig python information path c users runneradmin appdata local temp cibw run win build venv scripts python exe version
| 1
|
432,953
| 12,500,434,559
|
IssuesEvent
|
2020-06-01 22:16:25
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
[Android] User wallets not permitted up to a threshold of BAT earner
|
OS/Android priority/P1
|
Issue: Currently a lot of users are being pushed in to the user wallet funnel w/o having a lot of earned BAT creating a load for the wallet partner.
Fix: Prevent any UI to trigger 'Verify Wallet' until XX BAT has been earned as a virtual grant before user is allowed to verify their user wallet.
Please block user wallet release on to Android until fix for this issue is made. We will lead with Android and make fix on Desktop after that.
cc @aekeus @LaurenWags @kjozwiak @jamesmudgett
|
1.0
|
[Android] User wallets not permitted up to a threshold of BAT earner - Issue: Currently a lot of users are being pushed in to the user wallet funnel w/o having a lot of earned BAT creating a load for the wallet partner.
Fix: Prevent any UI to trigger 'Verify Wallet' until XX BAT has been earned as a virtual grant before user is allowed to verify their user wallet.
Please block user wallet release on to Android until fix for this issue is made. We will lead with Android and make fix on Desktop after that.
cc @aekeus @LaurenWags @kjozwiak @jamesmudgett
|
non_defect
|
user wallets not permitted up to a threshold of bat earner issue currently a lot of users are being pushed in to the user wallet funnel w o having a lot of earned bat creating a load for the wallet partner fix prevent any ui to trigger verify wallet until xx bat has been earned as a virtual grant before user is allowed to verify their user wallet please block user wallet release on to android until fix for this issue is made we will lead with android and make fix on desktop after that cc aekeus laurenwags kjozwiak jamesmudgett
| 0
|
72,926
| 13,937,102,554
|
IssuesEvent
|
2020-10-22 13:46:04
|
microsoft/Azure-Kinect-Sensor-SDK
|
https://api.github.com/repos/microsoft/Azure-Kinect-Sensor-SDK
|
opened
|
Unity | Can't display color or depth image in the Body Tracking example
|
Bug Code Sample Triage Needed
|
**Description**
I am trying to create a script to display Depth and Color images on a `RawImage` object.
I added to the `DeviceConfiguration` : `ColorResolution.R720p, ImageFormat.ColorBGRA32`
Then assigned it to the `currentFrameData` by modifying `BackgroundData` and `SkeletalTrackingProvider`.
But I'm having trouble on retrieving the bytes in a good format. I can't find any help online, and I hope someone here knows how to deal with it. Also I don't understand the "// Repack data and store image data" code.
I tried multiple workaround but nothing seems to work.
Any help would be appreciated
|
1.0
|
Unity | Can't display color or depth image in the Body Tracking example - **Description**
I am trying to create a script to display Depth and Color images on a `RawImage` object.
I added to the `DeviceConfiguration` : `ColorResolution.R720p, ImageFormat.ColorBGRA32`
Then assigned it to the `currentFrameData` by modifying `BackgroundData` and `SkeletalTrackingProvider`.
But I'm having trouble on retrieving the bytes in a good format. I can't find any help online, and I hope someone here knows how to deal with it. Also I don't understand the "// Repack data and store image data" code.
I tried multiple workaround but nothing seems to work.
Any help would be appreciated
|
non_defect
|
unity can t display color or depth image in the body tracking example description i am trying to create a script to display depth and color images on a rawimage object i added to the deviceconfiguration colorresolution imageformat then assigned it to the currentframedata by modifying backgrounddata and skeletaltrackingprovider but i m having trouble on retrieving the bytes in a good format i can t find any help online and i hope someone here knows how to deal with it also i don t understand the repack data and store image data code i tried multiple workaround but nothing seems to work any help would be appreciated
| 0
|
209,982
| 23,731,003,802
|
IssuesEvent
|
2022-08-31 01:41:41
|
zettatips/wedding-website
|
https://api.github.com/repos/zettatips/wedding-website
|
closed
|
CVE-2021-35065 (High) detected in glob-parent-3.1.0.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- gulp-4.0.2.tgz (Root Library)
- vinyl-fs-3.0.3.tgz
- glob-stream-6.1.0.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/zettatips/wedding-website/commit/a31757c82d2be8eafc3d6f819f065ae51339ae2f">a31757c82d2be8eafc3d6f819f065ae51339ae2f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-35065 (High) detected in glob-parent-3.1.0.tgz - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- gulp-4.0.2.tgz (Root Library)
- vinyl-fs-3.0.3.tgz
- glob-stream-6.1.0.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/zettatips/wedding-website/commit/a31757c82d2be8eafc3d6f819f065ae51339ae2f">a31757c82d2be8eafc3d6f819f065ae51339ae2f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in glob parent tgz autoclosed cve high severity vulnerability vulnerable library glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy gulp tgz root library vinyl fs tgz glob stream tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with mend
| 0
|
65,045
| 19,060,142,204
|
IssuesEvent
|
2021-11-26 06:10:46
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
ztest crashes with default zfs_abd_scatter_min_size
|
Type: Defect
|
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | 11
Kernel Version | not relevant
Architecture | x86_64
OpenZFS Version | ded851b2
### Describe the problem you're observing
ztest crashes an awful lot.
Most of the crashes, IME, look something like:
```
(gdb) bt
#0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50
#1 0x00007ffff7737537 in __GI_abort () at abort.c:79
#2 0x00007ffff7ae6923 in vpanic (fmt=0x7ffff7d74540 "Got SIGSEGV at address: 0x%lx\n", adx=adx@entry=0x7fffffffce58) at kernel.c:612
#3 0x00007ffff7ae69bb in panic (fmt=fmt@entry=0x7ffff7d74540 "Got SIGSEGV at address: 0x%lx\n") at kernel.c:621
#4 0x00007ffff7afaeb6 in arc_buf_sigsegv (sig=<optimized out>, si=<optimized out>, unused=<optimized out>) at ../../module/zfs/arc.c:1515
#5 <signal handler called>
#6 __memmove_avx_unaligned_erms () at ../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:437
#7 0x00007ffff7aef9be in abd_copy_to_buf_off_cb (buf=<optimized out>, size=size@entry=4096, private=private@entry=0x7fffffffd5a8) at ../../module/zfs/abd.c:828
#8 0x00007ffff7af1044 in abd_iterate_func (private=0x7fffffffd5a8, func=0x7ffff7aef9a0 <abd_copy_to_buf_off_cb>, size=4096, off=<optimized out>, abd=0x7fffd8021c90) at ../../module/zfs/abd.c:805
#9 abd_iterate_func (abd=0x7fffd8021c90, off=<optimized out>, size=<optimized out>, func=0x7ffff7aef9a0 <abd_copy_to_buf_off_cb>, private=0x7fffffffd5a8) at ../../module/zfs/abd.c:780
#10 0x00007ffff7af1278 in abd_copy_to_buf_off (buf=<optimized out>, abd=<optimized out>, off=off@entry=0, size=<optimized out>) at ../../module/zfs/abd.c:842
#11 0x00007ffff7b0239e in abd_copy_to_buf (size=<optimized out>, abd=<optimized out>, buf=<optimized out>) at ../../include/sys/abd.h:159
#12 arc_buf_fill (buf=0x555555b35a90, spa=spa@entry=0x5555556569a0, zb=zb@entry=0x7fffffffd6a0, flags=flags@entry=0) at ../../module/zfs/arc.c:2067
#13 0x00007ffff7b0327d in arc_untransform (buf=<optimized out>, spa=0x5555556569a0, zb=zb@entry=0x7fffffffd6a0, in_place=in_place@entry=B_FALSE) at ../../module/zfs/arc.c:2171
#14 0x00007ffff7b366f6 in dmu_objset_own_impl (ds=ds@entry=0x5555556af3b0, type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, decrypt=decrypt@entry=B_TRUE, osp=osp@entry=0x7fffffffd8c8, tag=<optimized out>)
at ../../module/zfs/dmu_objset.c:774
#15 0x00007ffff7b3abed in dmu_objset_own_impl (tag=0x555555573cf0 <__func__.9>, osp=0x7fffffffd8c8, decrypt=B_TRUE, readonly=B_TRUE, type=DMU_OST_ANY, ds=0x5555556af3b0) at ../../module/zfs/dmu_objset.c:757
#16 dmu_objset_own (name=name@entry=0x5555565cbfa0 "ztest/ds_4", type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, decrypt=decrypt@entry=B_TRUE, tag=tag@entry=0x555555573cf0 <__func__.9>, osp=osp@entry=0x7fffffffd8c8)
at ../../module/zfs/dmu_objset.c:808
#17 0x0000555555563fcb in ztest_dmu_objset_own (name=name@entry=0x5555565cbfa0 "ztest/ds_4", type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, tag=tag@entry=0x555555573cf0 <__func__.9>, osp=osp@entry=0x7fffffffd8c8,
decrypt=B_TRUE) at ztest.c:1602
#18 0x0000555555567d9c in ztest_replay_zil_cb (name=name@entry=0x5555565cbfa0 "ztest/ds_4", arg=arg@entry=0x0) at ztest.c:7255
#19 0x00007ffff7b35d0d in dmu_objset_find_impl (spa=spa@entry=0x5555556569a0, name=name@entry=0x5555565cbfa0 "ztest/ds_4", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2)
at ../../module/zfs/dmu_objset.c:2951
#20 0x00007ffff7b35e40 in dmu_objset_find_impl (spa=0x5555556569a0, name=name@entry=0x55555557a960 <ztest_opts> "ztest", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2)
at ../../module/zfs/dmu_objset.c:2894
#21 0x00007ffff7b3b303 in dmu_objset_find (name=name@entry=0x55555557a960 <ztest_opts> "ztest", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2) at ../../module/zfs/dmu_objset.c:2967
#22 0x000055555555cdc4 in ztest_run (zs=0x7ffff7ffb738) at ztest.c:7563
#23 main (argc=<optimized out>, argv=<optimized out>) at ztest.c:8062
```
After a round of bisecting, I ended up at 87c25d56, which I would not have guessed, but here we are. And lo, if you [extend ztest to set
zfs_abd_scatter_min_size to 4097](https://github.com/rincebrain/zfs/commit/74828db4139ba4a690a99c348c4ada86830660ee) on x86_64, it goes from crashing practically always to crashing never so far.
If we ask valgrind, first it complains a lot about [uninitialized values in the crypto code being read a bunch](https://github.com/openzfs/zfs/blob/6217656da33c0920cb9f213742fd51dd215bc455/module/zfs/dsl_crypt.c#L1197-L1200), but if you zero those, it becomes limited to eventually spitting out:
```
==129414== Thread 109:
==129414== Conditional jump or move depends on uninitialised value(s)
==129414== at 0x483EEEE: bcmp (vg_replace_strmem.c:1111)
==129414== by 0x48ABA33: abd_cmp_buf_off_cb (abd.c:852)
==129414== by 0x48AD043: abd_iterate_func (abd.c:805)
==129414== by 0x48AD043: abd_iterate_func (abd.c:780)
==129414== by 0x48AD304: abd_cmp_buf_off (abd.c:866)
==129414== by 0x48AD377: abd_cmp_buf (abd.h:165)
==129414== by 0x48AD377: abd_return_buf (abd.c:673)
==129414== by 0x48C0F42: arc_read_done (arc.c:5692)
==129414== by 0x4A606A3: zio_done (zio.c:4835)
==129414== by 0x4A54968: __zio_execute (zio.c:2209)
==129414== by 0x4A54968: zio_execute (zio.c:2122)
==129414== by 0x48A3961: taskq_thread (taskq.c:237)
==129414== by 0x4EFFEA6: start_thread (pthread_create.c:477)
==129414== by 0x5018DEE: clone (clone.S:95)
==129414== Uninitialised value was created by a heap allocation
==129414== at 0x483AEB8: memalign (vg_replace_malloc.c:906)
==129414== by 0x483AFCE: posix_memalign (vg_replace_malloc.c:1070)
==129414== by 0x48AF0BF: umem_alloc_aligned (umem.h:105)
==129414== by 0x48AF0BF: abd_alloc_chunks (abd_os.c:579)
==129414== by 0x48ABF8C: abd_alloc (abd.c:192)
==129414== by 0x48BC838: arc_hdr_alloc_abd (arc.c:3191)
==129414== by 0x48C24DD: arc_read (arc.c:6188)
==129414== by 0x4A4C5BF: zil_read_log_block (zil.c:241)
==129414== by 0x4A4C5BF: zil_parse (zil.c:398)
==129414== by 0x4A4D659: zil_check_log_chain (zil.c:975)
==129414== by 0x48F225A: dmu_objset_find_dp_impl (dmu_objset.c:2725)
==129414== by 0x48F2625: dmu_objset_find_dp_cb (dmu_objset.c:2758)
==129414== by 0x48A3961: taskq_thread (taskq.c:237)
==129414== by 0x4EFFEA6: start_thread (pthread_create.c:477)
==129414==
==129414==
==129414== Process terminating with default action of signal 6 (SIGABRT): dumping core
==129414== at 0x4F56CE1: raise (raise.c:51)
==129414== by 0x4F40536: abort (abort.c:79)
==129414== by 0x48A2922: vpanic (kernel.c:612)
==129414== by 0x48A29BA: panic (kernel.c:621)
==129414== by 0x48B6EB5: arc_buf_sigsegv (arc.c:1515)
==129414== by 0x4F0B13F: ??? (in /usr/lib/x86_64-linux-gnu/libpthread-2.31.so)
==129414== by 0x483F7F2: memmove (vg_replace_strmem.c:1270)
==129414== by 0x48AB9BD: abd_copy_to_buf_off_cb (abd.c:828)
==129414== by 0x48AD043: abd_iterate_func (abd.c:805)
==129414== by 0x48AD043: abd_iterate_func (abd.c:780)
==129414== by 0x48AD277: abd_copy_to_buf_off (abd.c:842)
==129414== by 0x48BE39D: abd_copy_to_buf (abd.h:159)
==129414== by 0x48BE39D: arc_buf_fill (arc.c:2067)
==129414== by 0x48BF27C: arc_untransform (arc.c:2171)
```
Obviously we could just...make ztest do that for now, but that seems problematic, and it's not presently clear to me whether the logical flaw is in the umem implementations of things or elsewhere? (Will continue looking, of course, but.)
### Describe how to reproduce the problem
Above.
### Include any warning/errors/backtraces from the system logs
Above.
|
1.0
|
ztest crashes with default zfs_abd_scatter_min_size - ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | 11
Kernel Version | not relevant
Architecture | x86_64
OpenZFS Version | ded851b2
### Describe the problem you're observing
ztest crashes an awful lot.
Most of the crashes, IME, look something like:
```
(gdb) bt
#0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50
#1 0x00007ffff7737537 in __GI_abort () at abort.c:79
#2 0x00007ffff7ae6923 in vpanic (fmt=0x7ffff7d74540 "Got SIGSEGV at address: 0x%lx\n", adx=adx@entry=0x7fffffffce58) at kernel.c:612
#3 0x00007ffff7ae69bb in panic (fmt=fmt@entry=0x7ffff7d74540 "Got SIGSEGV at address: 0x%lx\n") at kernel.c:621
#4 0x00007ffff7afaeb6 in arc_buf_sigsegv (sig=<optimized out>, si=<optimized out>, unused=<optimized out>) at ../../module/zfs/arc.c:1515
#5 <signal handler called>
#6 __memmove_avx_unaligned_erms () at ../sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S:437
#7 0x00007ffff7aef9be in abd_copy_to_buf_off_cb (buf=<optimized out>, size=size@entry=4096, private=private@entry=0x7fffffffd5a8) at ../../module/zfs/abd.c:828
#8 0x00007ffff7af1044 in abd_iterate_func (private=0x7fffffffd5a8, func=0x7ffff7aef9a0 <abd_copy_to_buf_off_cb>, size=4096, off=<optimized out>, abd=0x7fffd8021c90) at ../../module/zfs/abd.c:805
#9 abd_iterate_func (abd=0x7fffd8021c90, off=<optimized out>, size=<optimized out>, func=0x7ffff7aef9a0 <abd_copy_to_buf_off_cb>, private=0x7fffffffd5a8) at ../../module/zfs/abd.c:780
#10 0x00007ffff7af1278 in abd_copy_to_buf_off (buf=<optimized out>, abd=<optimized out>, off=off@entry=0, size=<optimized out>) at ../../module/zfs/abd.c:842
#11 0x00007ffff7b0239e in abd_copy_to_buf (size=<optimized out>, abd=<optimized out>, buf=<optimized out>) at ../../include/sys/abd.h:159
#12 arc_buf_fill (buf=0x555555b35a90, spa=spa@entry=0x5555556569a0, zb=zb@entry=0x7fffffffd6a0, flags=flags@entry=0) at ../../module/zfs/arc.c:2067
#13 0x00007ffff7b0327d in arc_untransform (buf=<optimized out>, spa=0x5555556569a0, zb=zb@entry=0x7fffffffd6a0, in_place=in_place@entry=B_FALSE) at ../../module/zfs/arc.c:2171
#14 0x00007ffff7b366f6 in dmu_objset_own_impl (ds=ds@entry=0x5555556af3b0, type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, decrypt=decrypt@entry=B_TRUE, osp=osp@entry=0x7fffffffd8c8, tag=<optimized out>)
at ../../module/zfs/dmu_objset.c:774
#15 0x00007ffff7b3abed in dmu_objset_own_impl (tag=0x555555573cf0 <__func__.9>, osp=0x7fffffffd8c8, decrypt=B_TRUE, readonly=B_TRUE, type=DMU_OST_ANY, ds=0x5555556af3b0) at ../../module/zfs/dmu_objset.c:757
#16 dmu_objset_own (name=name@entry=0x5555565cbfa0 "ztest/ds_4", type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, decrypt=decrypt@entry=B_TRUE, tag=tag@entry=0x555555573cf0 <__func__.9>, osp=osp@entry=0x7fffffffd8c8)
at ../../module/zfs/dmu_objset.c:808
#17 0x0000555555563fcb in ztest_dmu_objset_own (name=name@entry=0x5555565cbfa0 "ztest/ds_4", type=type@entry=DMU_OST_ANY, readonly=readonly@entry=B_TRUE, tag=tag@entry=0x555555573cf0 <__func__.9>, osp=osp@entry=0x7fffffffd8c8,
decrypt=B_TRUE) at ztest.c:1602
#18 0x0000555555567d9c in ztest_replay_zil_cb (name=name@entry=0x5555565cbfa0 "ztest/ds_4", arg=arg@entry=0x0) at ztest.c:7255
#19 0x00007ffff7b35d0d in dmu_objset_find_impl (spa=spa@entry=0x5555556569a0, name=name@entry=0x5555565cbfa0 "ztest/ds_4", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2)
at ../../module/zfs/dmu_objset.c:2951
#20 0x00007ffff7b35e40 in dmu_objset_find_impl (spa=0x5555556569a0, name=name@entry=0x55555557a960 <ztest_opts> "ztest", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2)
at ../../module/zfs/dmu_objset.c:2894
#21 0x00007ffff7b3b303 in dmu_objset_find (name=name@entry=0x55555557a960 <ztest_opts> "ztest", func=func@entry=0x555555567d70 <ztest_replay_zil_cb>, arg=arg@entry=0x0, flags=flags@entry=2) at ../../module/zfs/dmu_objset.c:2967
#22 0x000055555555cdc4 in ztest_run (zs=0x7ffff7ffb738) at ztest.c:7563
#23 main (argc=<optimized out>, argv=<optimized out>) at ztest.c:8062
```
After a round of bisecting, I ended up at 87c25d56, which I would not have guessed, but here we are. And lo, if you [extend ztest to set
zfs_abd_scatter_min_size to 4097](https://github.com/rincebrain/zfs/commit/74828db4139ba4a690a99c348c4ada86830660ee) on x86_64, it goes from crashing practically always to crashing never so far.
If we ask valgrind, first it complains a lot about [uninitialized values in the crypto code being read a bunch](https://github.com/openzfs/zfs/blob/6217656da33c0920cb9f213742fd51dd215bc455/module/zfs/dsl_crypt.c#L1197-L1200), but if you zero those, it becomes limited to eventually spitting out:
```
==129414== Thread 109:
==129414== Conditional jump or move depends on uninitialised value(s)
==129414== at 0x483EEEE: bcmp (vg_replace_strmem.c:1111)
==129414== by 0x48ABA33: abd_cmp_buf_off_cb (abd.c:852)
==129414== by 0x48AD043: abd_iterate_func (abd.c:805)
==129414== by 0x48AD043: abd_iterate_func (abd.c:780)
==129414== by 0x48AD304: abd_cmp_buf_off (abd.c:866)
==129414== by 0x48AD377: abd_cmp_buf (abd.h:165)
==129414== by 0x48AD377: abd_return_buf (abd.c:673)
==129414== by 0x48C0F42: arc_read_done (arc.c:5692)
==129414== by 0x4A606A3: zio_done (zio.c:4835)
==129414== by 0x4A54968: __zio_execute (zio.c:2209)
==129414== by 0x4A54968: zio_execute (zio.c:2122)
==129414== by 0x48A3961: taskq_thread (taskq.c:237)
==129414== by 0x4EFFEA6: start_thread (pthread_create.c:477)
==129414== by 0x5018DEE: clone (clone.S:95)
==129414== Uninitialised value was created by a heap allocation
==129414== at 0x483AEB8: memalign (vg_replace_malloc.c:906)
==129414== by 0x483AFCE: posix_memalign (vg_replace_malloc.c:1070)
==129414== by 0x48AF0BF: umem_alloc_aligned (umem.h:105)
==129414== by 0x48AF0BF: abd_alloc_chunks (abd_os.c:579)
==129414== by 0x48ABF8C: abd_alloc (abd.c:192)
==129414== by 0x48BC838: arc_hdr_alloc_abd (arc.c:3191)
==129414== by 0x48C24DD: arc_read (arc.c:6188)
==129414== by 0x4A4C5BF: zil_read_log_block (zil.c:241)
==129414== by 0x4A4C5BF: zil_parse (zil.c:398)
==129414== by 0x4A4D659: zil_check_log_chain (zil.c:975)
==129414== by 0x48F225A: dmu_objset_find_dp_impl (dmu_objset.c:2725)
==129414== by 0x48F2625: dmu_objset_find_dp_cb (dmu_objset.c:2758)
==129414== by 0x48A3961: taskq_thread (taskq.c:237)
==129414== by 0x4EFFEA6: start_thread (pthread_create.c:477)
==129414==
==129414==
==129414== Process terminating with default action of signal 6 (SIGABRT): dumping core
==129414== at 0x4F56CE1: raise (raise.c:51)
==129414== by 0x4F40536: abort (abort.c:79)
==129414== by 0x48A2922: vpanic (kernel.c:612)
==129414== by 0x48A29BA: panic (kernel.c:621)
==129414== by 0x48B6EB5: arc_buf_sigsegv (arc.c:1515)
==129414== by 0x4F0B13F: ??? (in /usr/lib/x86_64-linux-gnu/libpthread-2.31.so)
==129414== by 0x483F7F2: memmove (vg_replace_strmem.c:1270)
==129414== by 0x48AB9BD: abd_copy_to_buf_off_cb (abd.c:828)
==129414== by 0x48AD043: abd_iterate_func (abd.c:805)
==129414== by 0x48AD043: abd_iterate_func (abd.c:780)
==129414== by 0x48AD277: abd_copy_to_buf_off (abd.c:842)
==129414== by 0x48BE39D: abd_copy_to_buf (abd.h:159)
==129414== by 0x48BE39D: arc_buf_fill (arc.c:2067)
==129414== by 0x48BF27C: arc_untransform (arc.c:2171)
```
Obviously we could just...make ztest do that for now, but that seems problematic, and it's not presently clear to me whether the logical flaw is in the umem implementations of things or elsewhere? (Will continue looking, of course, but.)
### Describe how to reproduce the problem
Above.
### Include any warning/errors/backtraces from the system logs
Above.
|
defect
|
ztest crashes with default zfs abd scatter min size system information type version name distribution name debian distribution version kernel version not relevant architecture openzfs version describe the problem you re observing ztest crashes an awful lot most of the crashes ime look something like gdb bt gi raise sig sig entry at sysdeps unix sysv linux raise c in gi abort at abort c in vpanic fmt got sigsegv at address lx n adx adx entry at kernel c in panic fmt fmt entry got sigsegv at address lx n at kernel c in arc buf sigsegv sig si unused at module zfs arc c memmove avx unaligned erms at sysdeps multiarch memmove vec unaligned erms s in abd copy to buf off cb buf size size entry private private entry at module zfs abd c in abd iterate func private func size off abd at module zfs abd c abd iterate func abd off size func private at module zfs abd c in abd copy to buf off buf abd off off entry size at module zfs abd c in abd copy to buf size abd buf at include sys abd h arc buf fill buf spa spa entry zb zb entry flags flags entry at module zfs arc c in arc untransform buf spa zb zb entry in place in place entry b false at module zfs arc c in dmu objset own impl ds ds entry type type entry dmu ost any readonly readonly entry b true decrypt decrypt entry b true osp osp entry tag at module zfs dmu objset c in dmu objset own impl tag osp decrypt b true readonly b true type dmu ost any ds at module zfs dmu objset c dmu objset own name name entry ztest ds type type entry dmu ost any readonly readonly entry b true decrypt decrypt entry b true tag tag entry osp osp entry at module zfs dmu objset c in ztest dmu objset own name name entry ztest ds type type entry dmu ost any readonly readonly entry b true tag tag entry osp osp entry decrypt b true at ztest c in ztest replay zil cb name name entry ztest ds arg arg entry at ztest c in dmu objset find impl spa spa entry name name entry ztest ds func func entry arg arg entry flags flags entry at module zfs dmu objset c in dmu objset find impl spa name name entry ztest func func entry arg arg entry flags flags entry at module zfs dmu objset c in dmu objset find name name entry ztest func func entry arg arg entry flags flags entry at module zfs dmu objset c in ztest run zs at ztest c main argc argv at ztest c after a round of bisecting i ended up at which i would not have guessed but here we are and lo if you extend ztest to set zfs abd scatter min size to on it goes from crashing practically always to crashing never so far if we ask valgrind first it complains a lot about but if you zero those it becomes limited to eventually spitting out thread conditional jump or move depends on uninitialised value s at bcmp vg replace strmem c by abd cmp buf off cb abd c by abd iterate func abd c by abd iterate func abd c by abd cmp buf off abd c by abd cmp buf abd h by abd return buf abd c by arc read done arc c by zio done zio c by zio execute zio c by zio execute zio c by taskq thread taskq c by start thread pthread create c by clone clone s uninitialised value was created by a heap allocation at memalign vg replace malloc c by posix memalign vg replace malloc c by umem alloc aligned umem h by abd alloc chunks abd os c by abd alloc abd c by arc hdr alloc abd arc c by arc read arc c by zil read log block zil c by zil parse zil c by zil check log chain zil c by dmu objset find dp impl dmu objset c by dmu objset find dp cb dmu objset c by taskq thread taskq c by start thread pthread create c process terminating with default action of signal sigabrt dumping core at raise raise c by abort abort c by vpanic kernel c by panic kernel c by arc buf sigsegv arc c by in usr lib linux gnu libpthread so by memmove vg replace strmem c by abd copy to buf off cb abd c by abd iterate func abd c by abd iterate func abd c by abd copy to buf off abd c by abd copy to buf abd h by arc buf fill arc c by arc untransform arc c obviously we could just make ztest do that for now but that seems problematic and it s not presently clear to me whether the logical flaw is in the umem implementations of things or elsewhere will continue looking of course but describe how to reproduce the problem above include any warning errors backtraces from the system logs above
| 1
|
49,108
| 13,185,233,982
|
IssuesEvent
|
2020-08-12 20:59:33
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
test ticket for templates (Trac #755)
|
Incomplete Migration Migrated from Trac booking defect
|
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/755
, reported by nega and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-09-05T22:04:25",
"description": "this is the ticket body\n\n{{{\nsome\n formatted\n text\n}}}",
"reporter": "nega",
"cc": "negapluck@gmail.com",
"resolution": "fixed",
"_ts": "1409954665676165",
"component": "booking",
"summary": "test ticket for templates",
"priority": "normal",
"keywords": "",
"time": "2014-09-05T21:57:55",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
test ticket for templates (Trac #755) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/755
, reported by nega and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-09-05T22:04:25",
"description": "this is the ticket body\n\n{{{\nsome\n formatted\n text\n}}}",
"reporter": "nega",
"cc": "negapluck@gmail.com",
"resolution": "fixed",
"_ts": "1409954665676165",
"component": "booking",
"summary": "test ticket for templates",
"priority": "normal",
"keywords": "",
"time": "2014-09-05T21:57:55",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
test ticket for templates trac migrated from reported by nega and owned by nega json status closed changetime description this is the ticket body n n nsome n formatted n text n reporter nega cc negapluck gmail com resolution fixed ts component booking summary test ticket for templates priority normal keywords time milestone owner nega type defect
| 1
|
47,157
| 13,056,043,323
|
IssuesEvent
|
2020-07-30 03:28:57
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
svn version upgrade needed (Trac #73)
|
Migrated from Trac defect infrastructure
|
DART nodes need svn version 1.4.x or better for latest
and greatest dartboard voodoo.
Migrated from https://code.icecube.wisc.edu/ticket/73
```json
{
"status": "closed",
"changetime": "2007-06-26T16:16:23",
"description": "DART nodes need svn version 1.4.x or better for latest\nand greatest dartboard voodoo.\n\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1182874583000000",
"component": "infrastructure",
"summary": "svn version upgrade needed",
"priority": "normal",
"keywords": "",
"time": "2007-06-25T21:20:12",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
|
1.0
|
svn version upgrade needed (Trac #73) - DART nodes need svn version 1.4.x or better for latest
and greatest dartboard voodoo.
Migrated from https://code.icecube.wisc.edu/ticket/73
```json
{
"status": "closed",
"changetime": "2007-06-26T16:16:23",
"description": "DART nodes need svn version 1.4.x or better for latest\nand greatest dartboard voodoo.\n\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1182874583000000",
"component": "infrastructure",
"summary": "svn version upgrade needed",
"priority": "normal",
"keywords": "",
"time": "2007-06-25T21:20:12",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
|
defect
|
svn version upgrade needed trac dart nodes need svn version x or better for latest and greatest dartboard voodoo migrated from json status closed changetime description dart nodes need svn version x or better for latest nand greatest dartboard voodoo n n reporter blaufuss cc resolution wont or cant fix ts component infrastructure summary svn version upgrade needed priority normal keywords time milestone owner cgils type defect
| 1
|
15,675
| 9,593,549,970
|
IssuesEvent
|
2019-05-09 11:52:13
|
molgenis/molgenis
|
https://api.github.com/repos/molgenis/molgenis
|
closed
|
Calls with token plus session change the session's authentication
|
8.0.0-SNAPSHOT mod:security mod:web team-platform
|
### How to Reproduce
Create a token `blah` for user A.
Log in on molgenis as user B. (You get a session cookie for B.)
Make any call in the browser with `?molgenis-token=blah` (So with a session cookie for A and a token for B)
Now go back to making calls with the session cookie.
### Expected behavior
Calls with the session cookie still give results for user A.
### Observed behavior
Calls with the session cookie now give results for user B.
The session got changed into one for B.
|
True
|
Calls with token plus session change the session's authentication - ### How to Reproduce
Create a token `blah` for user A.
Log in on molgenis as user B. (You get a session cookie for B.)
Make any call in the browser with `?molgenis-token=blah` (So with a session cookie for A and a token for B)
Now go back to making calls with the session cookie.
### Expected behavior
Calls with the session cookie still give results for user A.
### Observed behavior
Calls with the session cookie now give results for user B.
The session got changed into one for B.
|
non_defect
|
calls with token plus session change the session s authentication how to reproduce create a token blah for user a log in on molgenis as user b you get a session cookie for b make any call in the browser with molgenis token blah so with a session cookie for a and a token for b now go back to making calls with the session cookie expected behavior calls with the session cookie still give results for user a observed behavior calls with the session cookie now give results for user b the session got changed into one for b
| 0
|
128,223
| 5,051,062,837
|
IssuesEvent
|
2016-12-20 20:41:12
|
scottbaggett/lexus-int
|
https://api.github.com/repos/scottbaggett/lexus-int
|
closed
|
(Article) Input final copy/images
|
Content Missing High Priority
|
I downloaded and put the images into the proper folders.
- I already updated the images for crafted for lexus & all-lit-up. Copy needs to be checked though.
- Copydeck is inside dropbox in copydeck folder.
|
1.0
|
(Article) Input final copy/images - I downloaded and put the images into the proper folders.
- I already updated the images for crafted for lexus & all-lit-up. Copy needs to be checked though.
- Copydeck is inside dropbox in copydeck folder.
|
non_defect
|
article input final copy images i downloaded and put the images into the proper folders i already updated the images for crafted for lexus all lit up copy needs to be checked though copydeck is inside dropbox in copydeck folder
| 0
|
38,629
| 8,950,552,041
|
IssuesEvent
|
2019-01-25 11:07:15
|
svigerske/ipopt-donotuse
|
https://api.github.com/repos/svigerske/ipopt-donotuse
|
closed
|
Release of Ipopt 3.0
|
Initial Release Ipopt defect highest
|
Issue created by migration from Trac.
Original creator: claird
Original creation time: 2005-08-26 22:08:10
Assignee: ipopt-team
Version: 3.0
This milestone marks the initial release of Ipopt 3.0.0.
|
1.0
|
Release of Ipopt 3.0 - Issue created by migration from Trac.
Original creator: claird
Original creation time: 2005-08-26 22:08:10
Assignee: ipopt-team
Version: 3.0
This milestone marks the initial release of Ipopt 3.0.0.
|
defect
|
release of ipopt issue created by migration from trac original creator claird original creation time assignee ipopt team version this milestone marks the initial release of ipopt
| 1
|
54,305
| 13,543,362,155
|
IssuesEvent
|
2020-09-16 18:50:22
|
NREL/EnergyPlus
|
https://api.github.com/repos/NREL/EnergyPlus
|
closed
|
Site-source conversion factor for steam is wrong
|
Defect NotIDDChange
|
Issue overview
--------------
The default site-source conversion factor for steam in the ABUPS report might be wrong. With all defaults, the value is 0.300 for steam. This seems more like the conversion efficiency of steam. Other sources put the site-source factor at 1.20-1.35.

### Details
Some additional details for this issue (if relevant):
- Version of EnergyPlus v9.4 in develop
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
1.0
|
Site-source conversion factor for steam is wrong - Issue overview
--------------
The default site-source conversion factor for steam in the ABUPS report might be wrong. With all defaults, the value is 0.300 for steam. This seems more like the conversion efficiency of steam. Other sources put the site-source factor at 1.20-1.35.

### Details
Some additional details for this issue (if relevant):
- Version of EnergyPlus v9.4 in develop
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
defect
|
site source conversion factor for steam is wrong issue overview the default site source conversion factor for steam in the abups report might be wrong with all defaults the value is for steam this seems more like the conversion efficiency of steam other sources put the site source factor at details some additional details for this issue if relevant version of energyplus in develop checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
| 1
|
3,639
| 4,462,331,542
|
IssuesEvent
|
2016-08-24 09:31:46
|
opensheetmusicdisplay/opensheetmusicdisplay
|
https://api.github.com/repos/opensheetmusicdisplay/opensheetmusicdisplay
|
closed
|
Get rid of `dist` in repo
|
infrastructure
|
Its annoying. Take my PR #16 for instance. you can't properly browse changed files, everything is covered by this random dist stuff.
So I thought about having a npm prepublish hook that builds stuff before publish so it gets included inside the tgz for npm. The only problem here is that npm respects `.gitignore` for packaging, so I have to figure out how to do this.
|
1.0
|
Get rid of `dist` in repo - Its annoying. Take my PR #16 for instance. you can't properly browse changed files, everything is covered by this random dist stuff.
So I thought about having a npm prepublish hook that builds stuff before publish so it gets included inside the tgz for npm. The only problem here is that npm respects `.gitignore` for packaging, so I have to figure out how to do this.
|
non_defect
|
get rid of dist in repo its annoying take my pr for instance you can t properly browse changed files everything is covered by this random dist stuff so i thought about having a npm prepublish hook that builds stuff before publish so it gets included inside the tgz for npm the only problem here is that npm respects gitignore for packaging so i have to figure out how to do this
| 0
|
26,097
| 4,586,712,116
|
IssuesEvent
|
2016-09-20 00:11:50
|
Openki/Openki
|
https://api.github.com/repos/Openki/Openki
|
closed
|
Answers to comments are only removed server-side
|
defect
|
if you click the closing-x to a sub-comment / to an answer to a comment, it will not be removed on the client. Server seems to delete the comment correctly
|
1.0
|
Answers to comments are only removed server-side - if you click the closing-x to a sub-comment / to an answer to a comment, it will not be removed on the client. Server seems to delete the comment correctly
|
defect
|
answers to comments are only removed server side if you click the closing x to a sub comment to an answer to a comment it will not be removed on the client server seems to delete the comment correctly
| 1
|
54,711
| 13,894,597,682
|
IssuesEvent
|
2020-10-19 14:52:19
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
"Show rooms with unread messages first" checkbox is buggy
|
bug defect type:new-room-list
|
Element version: 1.7.9 (app.element.io)
Pressing the "Show rooms with unread messages first" checkbox produces weird but deterministic behaviour. Clicking it does not toggle the checkbox unless pressing more times or reopening the dialog. When pressing multiple times it always shows the wrong state.
Video:

|
1.0
|
"Show rooms with unread messages first" checkbox is buggy - Element version: 1.7.9 (app.element.io)
Pressing the "Show rooms with unread messages first" checkbox produces weird but deterministic behaviour. Clicking it does not toggle the checkbox unless pressing more times or reopening the dialog. When pressing multiple times it always shows the wrong state.
Video:

|
defect
|
show rooms with unread messages first checkbox is buggy element version app element io pressing the show rooms with unread messages first checkbox produces weird but deterministic behaviour clicking it does not toggle the checkbox unless pressing more times or reopening the dialog when pressing multiple times it always shows the wrong state video
| 1
|
60,847
| 17,023,538,095
|
IssuesEvent
|
2021-07-03 02:32:12
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Data browser breaks on many-versioned Relation
|
Component: website Priority: major Resolution: invalid Type: defect
|
**[Submitted to the original trac issue database at 4.02pm, Tuesday, 12th January 2010]**
Relation #16377 (Downs Link path) has 81 versions (as of date of filing this bug). It also has loads and loads of members. Trying to view the edit history using http://www.openstreetmap.org/browse/relation/16377/history always results in a 500.
Out of memory, maybe?
|
1.0
|
Data browser breaks on many-versioned Relation - **[Submitted to the original trac issue database at 4.02pm, Tuesday, 12th January 2010]**
Relation #16377 (Downs Link path) has 81 versions (as of date of filing this bug). It also has loads and loads of members. Trying to view the edit history using http://www.openstreetmap.org/browse/relation/16377/history always results in a 500.
Out of memory, maybe?
|
defect
|
data browser breaks on many versioned relation relation downs link path has versions as of date of filing this bug it also has loads and loads of members trying to view the edit history using always results in a out of memory maybe
| 1
|
389,592
| 11,504,281,155
|
IssuesEvent
|
2020-02-12 22:55:02
|
garden-io/garden
|
https://api.github.com/repos/garden-io/garden
|
closed
|
Unexpected error when templating with ternary operator and ||
|
bug priority:low
|
## Bug
Noticed two issues when using a ternary operator and a `||` in template strings.
### Current Behavior
Firstly, the following template string:
```yaml
TEST_OUTPUT: "${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}"
```
will fail with:
```console
Invalid template string ${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}: Circular reference detected when
resolving key (runtime -> runtime.tasks.echo-hello)
```
Secondly, the following template string:
```yaml
TEST_OUTPUT: "${runtime.tasks.no-task.outputs.log || 'default'}"
```
fails with:
```console
Unable to resolve one or more runtime template values for service 'backend': ${runtime.tasks.no-task.outputs.log}
```
**if** `no-task` doesn't exist.
Finally the error isn't very helpful if the `dependencies` field is not set. It's the same as above.
### Expected behavior
Should work.
### Reproducible example
```yaml
kind: Module
name: frontend
type: container
tasks:
- name: echo-hello
args: ["echo", "hello"]
...
---
kind: Module
name: backend
type: container
services:
- name: backend
...
env:
TEST_OUTPUT: "${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}" # FIXME
TEST_OUTPUT: "${runtime.tasks.echo-hello.outputs.log || 'default'}" # And FIXME
dependencies: ["echo-hello"]
```
### Workaround
N/A
|
1.0
|
Unexpected error when templating with ternary operator and || - ## Bug
Noticed two issues when using a ternary operator and a `||` in template strings.
### Current Behavior
Firstly, the following template string:
```yaml
TEST_OUTPUT: "${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}"
```
will fail with:
```console
Invalid template string ${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}: Circular reference detected when
resolving key (runtime -> runtime.tasks.echo-hello)
```
Secondly, the following template string:
```yaml
TEST_OUTPUT: "${runtime.tasks.no-task.outputs.log || 'default'}"
```
fails with:
```console
Unable to resolve one or more runtime template values for service 'backend': ${runtime.tasks.no-task.outputs.log}
```
**if** `no-task` doesn't exist.
Finally the error isn't very helpful if the `dependencies` field is not set. It's the same as above.
### Expected behavior
Should work.
### Reproducible example
```yaml
kind: Module
name: frontend
type: container
tasks:
- name: echo-hello
args: ["echo", "hello"]
...
---
kind: Module
name: backend
type: container
services:
- name: backend
...
env:
TEST_OUTPUT: "${runtime.tasks.echo-hello ? runtime.tasks.echo-hello.outputs.log : 'default'}" # FIXME
TEST_OUTPUT: "${runtime.tasks.echo-hello.outputs.log || 'default'}" # And FIXME
dependencies: ["echo-hello"]
```
### Workaround
N/A
|
non_defect
|
unexpected error when templating with ternary operator and bug noticed two issues when using a ternary operator and a in template strings current behavior firstly the following template string yaml test output runtime tasks echo hello runtime tasks echo hello outputs log default will fail with console invalid template string runtime tasks echo hello runtime tasks echo hello outputs log default circular reference detected when resolving key runtime runtime tasks echo hello secondly the following template string yaml test output runtime tasks no task outputs log default fails with console unable to resolve one or more runtime template values for service backend runtime tasks no task outputs log if no task doesn t exist finally the error isn t very helpful if the dependencies field is not set it s the same as above expected behavior should work reproducible example yaml kind module name frontend type container tasks name echo hello args kind module name backend type container services name backend env test output runtime tasks echo hello runtime tasks echo hello outputs log default fixme test output runtime tasks echo hello outputs log default and fixme dependencies workaround n a
| 0
|
176,076
| 21,365,696,439
|
IssuesEvent
|
2022-04-20 01:11:05
|
phytomichael/KSA
|
https://api.github.com/repos/phytomichael/KSA
|
closed
|
CVE-2018-11039 (Medium) detected in spring-web-3.1.1.RELEASE.jar - autoclosed
|
security vulnerability
|
## CVE-2018-11039 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-3.1.1.RELEASE.jar</b></p></summary>
<p>Spring Framework Parent</p>
<p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_df7b546f-7ca1-4faf-a7cf-4b2796ca07e7/20190718203425_56332/git_depth_0/KSA/ksa/ksa/ksa-web-root/ksa-web/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/KSA/ksa/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-web-3.1.1.RELEASE.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Framework (versions 5.0.x prior to 5.0.7, versions 4.3.x prior to 4.3.18, and older unsupported versions) allow web applications to change the HTTP request method to any HTTP method (including TRACE) using the HiddenHttpMethodFilter in Spring MVC. If an application has a pre-existing XSS vulnerability, a malicious user (or attacker) can use this filter to escalate to an XST (Cross Site Tracing) attack.
<p>Publish Date: 2018-06-25
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11039>CVE-2018-11039</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-11039">https://nvd.nist.gov/vuln/detail/CVE-2018-11039</a></p>
<p>Release Date: 2018-06-25</p>
<p>Fix Resolution: 5.0.7,4.3.18</p>
</p>
</details>
<p></p>
|
True
|
CVE-2018-11039 (Medium) detected in spring-web-3.1.1.RELEASE.jar - autoclosed - ## CVE-2018-11039 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-3.1.1.RELEASE.jar</b></p></summary>
<p>Spring Framework Parent</p>
<p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_df7b546f-7ca1-4faf-a7cf-4b2796ca07e7/20190718203425_56332/git_depth_0/KSA/ksa/ksa/ksa-web-root/ksa-web/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/KSA/ksa/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar,/root/.m2/repository/org/springframework/spring-web/3.1.1.RELEASE/spring-web-3.1.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-web-3.1.1.RELEASE.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Framework (versions 5.0.x prior to 5.0.7, versions 4.3.x prior to 4.3.18, and older unsupported versions) allow web applications to change the HTTP request method to any HTTP method (including TRACE) using the HiddenHttpMethodFilter in Spring MVC. If an application has a pre-existing XSS vulnerability, a malicious user (or attacker) can use this filter to escalate to an XST (Cross Site Tracing) attack.
<p>Publish Date: 2018-06-25
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11039>CVE-2018-11039</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-11039">https://nvd.nist.gov/vuln/detail/CVE-2018-11039</a></p>
<p>Release Date: 2018-06-25</p>
<p>Fix Resolution: 5.0.7,4.3.18</p>
</p>
</details>
<p></p>
|
non_defect
|
cve medium detected in spring web release jar autoclosed cve medium severity vulnerability vulnerable library spring web release jar spring framework parent path to dependency file tmp whitesource archiveextractor git depth ksa ksa ksa ksa web root ksa web pom xml path to vulnerable library root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar repository org springframework spring web release spring web release jar ksa ksa web root ksa web target root web inf lib spring web release jar root repository org springframework spring web release spring web release jar repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar ksa ksa ksa ksa web root ksa web target root web inf lib spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar root repository org springframework spring web release spring web release jar dependency hierarchy x spring web release jar vulnerable library vulnerability details spring framework versions x prior to versions x prior to and older unsupported versions allow web applications to change the http request method to any http method including trace using the hiddenhttpmethodfilter in spring mvc if an application has a pre existing xss vulnerability a malicious user or attacker can use this filter to escalate to an xst cross site tracing attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
| 0
|
30,221
| 6,046,530,636
|
IssuesEvent
|
2017-06-12 12:23:09
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
Dropdown should scroll to selected item when overlay gets opened
|
defect
|
Hi,
I use dropdown component with bindings. When dropdown is opened for the first time, the scroll is not on selected item, I've to scroll manually to see it. Perhaps I'm missing something ?
Thanks
|
1.0
|
Dropdown should scroll to selected item when overlay gets opened - Hi,
I use dropdown component with bindings. When dropdown is opened for the first time, the scroll is not on selected item, I've to scroll manually to see it. Perhaps I'm missing something ?
Thanks
|
defect
|
dropdown should scroll to selected item when overlay gets opened hi i use dropdown component with bindings when dropdown is opened for the first time the scroll is not on selected item i ve to scroll manually to see it perhaps i m missing something thanks
| 1
|
56,133
| 14,940,921,747
|
IssuesEvent
|
2021-01-25 18:59:12
|
AeroScripts/QuestieDev
|
https://api.github.com/repos/AeroScripts/QuestieDev
|
closed
|
Auto Turn In Bugs on Argent Dawn Valor Token quests
|
Questie - Quest Issue Type - Defect
|
<!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 15M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @BreakBB / TheCrux#1702 (Discord) - Development
* @drejjmit / Drejjmit#8241 (Discord) - Testing/Research
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
<!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. -->
Auto Turn In gets hung up when turning in multiple AD rep items. It'll fail on the first quest, and just keep repeating it. If you hold shift you can sort of un-stick it.
* Corruptor's Scourgestones
* Invader's Scourgestones
* Minion's Scourgestones
* Argent Dawn Valor Token - Item - World of Warcraft || https://classic.wowhead.com/item=12844/argent-dawn-valor-token#reward-from-q:0+1+12
Happy to say this is a conflict with another addon, what can I do to help verify that? I don't think I have Auto Turn In enabled elsewhere, but hard to know for sure.
Hey @AeroScripts or @drejjmit, next time I have a bunch of Scourgestones do I just run profiler while I try and turn them in? What can I do to help pinpoint the issue?
## Screenshots
<!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. -->
## Questie version
<!--
Which version of Questie are you using? You can find it by:
- 1. Hovering over the Questie Minimap Icon
- 2. looking at your Questie.toc file (open it with any text editor).
It looks something like this: "v5.9.0" or "## Version: 5.9.0".
-->
6.0.4
|
1.0
|
Auto Turn In Bugs on Argent Dawn Valor Token quests - <!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 15M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @BreakBB / TheCrux#1702 (Discord) - Development
* @drejjmit / Drejjmit#8241 (Discord) - Testing/Research
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
<!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. -->
Auto Turn In gets hung up when turning in multiple AD rep items. It'll fail on the first quest, and just keep repeating it. If you hold shift you can sort of un-stick it.
* Corruptor's Scourgestones
* Invader's Scourgestones
* Minion's Scourgestones
* Argent Dawn Valor Token - Item - World of Warcraft || https://classic.wowhead.com/item=12844/argent-dawn-valor-token#reward-from-q:0+1+12
Happy to say this is a conflict with another addon, what can I do to help verify that? I don't think I have Auto Turn In enabled elsewhere, but hard to know for sure.
Hey @AeroScripts or @drejjmit, next time I have a bunch of Scourgestones do I just run profiler while I try and turn them in? What can I do to help pinpoint the issue?
## Screenshots
<!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. -->
## Questie version
<!--
Which version of Questie are you using? You can find it by:
- 1. Hovering over the Questie Minimap Icon
- 2. looking at your Questie.toc file (open it with any text editor).
It looks something like this: "v5.9.0" or "## Version: 5.9.0".
-->
6.0.4
|
defect
|
auto turn in bugs on argent dawn valor token quests read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is breakbb thecrux discord development drejjmit drejjmit discord testing research if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description auto turn in gets hung up when turning in multiple ad rep items it ll fail on the first quest and just keep repeating it if you hold shift you can sort of un stick it corruptor s scourgestones invader s scourgestones minion s scourgestones argent dawn valor token item world of warcraft happy to say this is a conflict with another addon what can i do to help verify that i don t think i have auto turn in enabled elsewhere but hard to know for sure hey aeroscripts or drejjmit next time i have a bunch of scourgestones do i just run profiler while i try and turn them in what can i do to help pinpoint the issue screenshots questie version which version of questie are you using you can find it by hovering over the questie minimap icon looking at your questie toc file open it with any text editor it looks something like this or version
| 1
|
41,017
| 10,265,853,154
|
IssuesEvent
|
2019-08-22 19:55:27
|
carbon-design-system/ibm-security
|
https://api.github.com/repos/carbon-design-system/ibm-security
|
closed
|
Tearsheet - wrong type size for Cancel button
|
Defect
|
@cameroncalder --
- [ ] Helper text on Cancel button `Information will not be saved upon exit` should be label-01 (size 12 not 16)
- [ ] Remove period after `exit` in helper text
|
1.0
|
Tearsheet - wrong type size for Cancel button - @cameroncalder --
- [ ] Helper text on Cancel button `Information will not be saved upon exit` should be label-01 (size 12 not 16)
- [ ] Remove period after `exit` in helper text
|
defect
|
tearsheet wrong type size for cancel button cameroncalder helper text on cancel button information will not be saved upon exit should be label size not remove period after exit in helper text
| 1
|
71,025
| 15,169,658,601
|
IssuesEvent
|
2021-02-12 21:34:19
|
gavarasana/ps-state
|
https://api.github.com/repos/gavarasana/ps-state
|
closed
|
CVE-2020-7693 (Medium) detected in sockjs-0.3.19.tgz
|
security vulnerability
|
## CVE-2020-7693 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sockjs-0.3.19.tgz</b></p></summary>
<p>SockJS-node is a server counterpart of SockJS-client a JavaScript library that provides a WebSocket-like object in the browser. SockJS gives you a coherent, cross-browser, Javascript API which creates a low latency, full duplex, cross-domain communication</p>
<p>Library home page: <a href="https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz">https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz</a></p>
<p>Path to dependency file: ps-state/package.json</p>
<p>Path to vulnerable library: ps-state/node_modules/sockjs/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.1.tgz (Root Library)
- webpack-dev-server-3.10.3.tgz
- :x: **sockjs-0.3.19.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gavarasana/ps-state/commit/0ff4648715ecf764b163d4c6a4b8ab91144b9bd0">0ff4648715ecf764b163d4c6a4b8ab91144b9bd0</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Incorrect handling of Upgrade header with the value websocket leads in crashing of containers hosting sockjs apps. This affects the package sockjs before 0.3.20.
<p>Publish Date: 2020-07-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693>CVE-2020-7693</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sockjs/sockjs-node/pull/265">https://github.com/sockjs/sockjs-node/pull/265</a></p>
<p>Release Date: 2020-07-09</p>
<p>Fix Resolution: sockjs - 0.3.20</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7693 (Medium) detected in sockjs-0.3.19.tgz - ## CVE-2020-7693 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sockjs-0.3.19.tgz</b></p></summary>
<p>SockJS-node is a server counterpart of SockJS-client a JavaScript library that provides a WebSocket-like object in the browser. SockJS gives you a coherent, cross-browser, Javascript API which creates a low latency, full duplex, cross-domain communication</p>
<p>Library home page: <a href="https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz">https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz</a></p>
<p>Path to dependency file: ps-state/package.json</p>
<p>Path to vulnerable library: ps-state/node_modules/sockjs/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.1.tgz (Root Library)
- webpack-dev-server-3.10.3.tgz
- :x: **sockjs-0.3.19.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gavarasana/ps-state/commit/0ff4648715ecf764b163d4c6a4b8ab91144b9bd0">0ff4648715ecf764b163d4c6a4b8ab91144b9bd0</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Incorrect handling of Upgrade header with the value websocket leads in crashing of containers hosting sockjs apps. This affects the package sockjs before 0.3.20.
<p>Publish Date: 2020-07-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693>CVE-2020-7693</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sockjs/sockjs-node/pull/265">https://github.com/sockjs/sockjs-node/pull/265</a></p>
<p>Release Date: 2020-07-09</p>
<p>Fix Resolution: sockjs - 0.3.20</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in sockjs tgz cve medium severity vulnerability vulnerable library sockjs tgz sockjs node is a server counterpart of sockjs client a javascript library that provides a websocket like object in the browser sockjs gives you a coherent cross browser javascript api which creates a low latency full duplex cross domain communication library home page a href path to dependency file ps state package json path to vulnerable library ps state node modules sockjs package json dependency hierarchy react scripts tgz root library webpack dev server tgz x sockjs tgz vulnerable library found in head commit a href found in base branch main vulnerability details incorrect handling of upgrade header with the value websocket leads in crashing of containers hosting sockjs apps this affects the package sockjs before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution sockjs step up your open source security game with whitesource
| 0
|
30,094
| 6,021,089,195
|
IssuesEvent
|
2017-06-07 17:55:01
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
opened
|
Cache eviction not fail fast.
|
Team: Core Type: Defect
|
When using the following hz xml only on retrieving the cache, I get a validation error. But the XML file could be validated when loaded.
````
<hazelcast xsi:schemaLocation="http://www.hazelcast.com/schema/config
http://www.hazelcast.com/schema/config/hazelcast-config-3.8.xsd"
xmlns="http://www.hazelcast.com/schema/config"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<group>
<name>workers</name>
</group>
<lite-member enabled="true"/>
<network>
<port port-count="200" auto-increment="true">5701</port>
<join>
<multicast enabled="false"/>
<tcp-ip enabled="true">
<member>10.212.40.101:5701</member>
<member>10.212.40.102:5701</member>
<member>10.212.40.103:5701</member>
<member>10.212.40.104:5701</member>
<member>10.212.40.105:5701</member>
<member>10.212.40.106:5701</member>
<member>10.212.40.107:5701</member>
<member>10.212.40.108:5701</member>
<member>10.212.40.109:5701</member>
<member>10.212.40.110:5701</member>
</tcp-ip>
</join>
<ssl enabled="false"/>
</network>
<properties>
<property name="hazelcast.phone.home.enabled">false</property>
</properties>
<license-key>...</license-key>
<!--MANAGEMENT_CENTER_CONFIG-->
<native-memory allocator-type="POOLED" enabled="true">
<size unit="GIGABYTES" value="2" />
<metadata-space-percentage>20</metadata-space-percentage>
</native-memory>
<cache name="cache">
<eviction size="10000000" max-size-policy="ENTRY_COUNT" eviction-policy="LFU"/>
<backup-count>1</backup-count>
<async-backup-count>0</async-backup-count>
<in-memory-format>NATIVE</in-memory-format>
</cache>
</hazelcast>
``
```
ERROR 2017-06-07 19:28:12,284 [Thread-5] com.hazelcast.simulator.worker.testcontainer.TestManager: --------------------------- global prepare of LongStringCacheTest FAILED ---------------------------
java.lang.IllegalArgumentException: Invalid max-size policy (ENTRY_COUNT) for com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore! Only USED_NATIVE_MEMORY_SIZE, USED_NATIVE_MEMORY_PERCENTAGE, FREE_NATIVE_MEMORY_SIZE, FREE_NATIVE_MEMORY_PERCENTAGE are supported.
at com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore.createCacheEvictionChecker(HiDensityNativeMemoryCacheRecordStore.java:121)
at com.hazelcast.cache.impl.AbstractCacheRecordStore.<init>(AbstractCacheRecordStore.java:150)
at com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore.<init>(HiDensityNativeMemoryCacheRecordStore.java:59)
at com.hazelcast.cache.EnterpriseCacheService.newNativeRecordStore(EnterpriseCacheService.java:243)
at com.hazelcast.cache.EnterpriseCacheService.createNewRecordStore(EnterpriseCacheService.java:217)
at com.hazelcast.cache.impl.CachePartitionSegment.createNew(CachePartitionSegment.java:51)
at com.hazelcast.cache.impl.CachePartitionSegment.createNew(CachePartitionSegment.java:37)
at com.hazelcast.util.ConcurrencyUtil.getOrPutSynchronized(ConcurrencyUtil.java:73)
at com.hazelcast.cache.impl.CachePartitionSegment.getOrCreateRecordStore(CachePartitionSegment.java:67)
at com.hazelcast.cache.impl.AbstractCacheService.getOrCreateRecordStore(AbstractCacheService.java:278)
```
|
1.0
|
Cache eviction not fail fast. - When using the following hz xml only on retrieving the cache, I get a validation error. But the XML file could be validated when loaded.
````
<hazelcast xsi:schemaLocation="http://www.hazelcast.com/schema/config
http://www.hazelcast.com/schema/config/hazelcast-config-3.8.xsd"
xmlns="http://www.hazelcast.com/schema/config"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<group>
<name>workers</name>
</group>
<lite-member enabled="true"/>
<network>
<port port-count="200" auto-increment="true">5701</port>
<join>
<multicast enabled="false"/>
<tcp-ip enabled="true">
<member>10.212.40.101:5701</member>
<member>10.212.40.102:5701</member>
<member>10.212.40.103:5701</member>
<member>10.212.40.104:5701</member>
<member>10.212.40.105:5701</member>
<member>10.212.40.106:5701</member>
<member>10.212.40.107:5701</member>
<member>10.212.40.108:5701</member>
<member>10.212.40.109:5701</member>
<member>10.212.40.110:5701</member>
</tcp-ip>
</join>
<ssl enabled="false"/>
</network>
<properties>
<property name="hazelcast.phone.home.enabled">false</property>
</properties>
<license-key>...</license-key>
<!--MANAGEMENT_CENTER_CONFIG-->
<native-memory allocator-type="POOLED" enabled="true">
<size unit="GIGABYTES" value="2" />
<metadata-space-percentage>20</metadata-space-percentage>
</native-memory>
<cache name="cache">
<eviction size="10000000" max-size-policy="ENTRY_COUNT" eviction-policy="LFU"/>
<backup-count>1</backup-count>
<async-backup-count>0</async-backup-count>
<in-memory-format>NATIVE</in-memory-format>
</cache>
</hazelcast>
``
```
ERROR 2017-06-07 19:28:12,284 [Thread-5] com.hazelcast.simulator.worker.testcontainer.TestManager: --------------------------- global prepare of LongStringCacheTest FAILED ---------------------------
java.lang.IllegalArgumentException: Invalid max-size policy (ENTRY_COUNT) for com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore! Only USED_NATIVE_MEMORY_SIZE, USED_NATIVE_MEMORY_PERCENTAGE, FREE_NATIVE_MEMORY_SIZE, FREE_NATIVE_MEMORY_PERCENTAGE are supported.
at com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore.createCacheEvictionChecker(HiDensityNativeMemoryCacheRecordStore.java:121)
at com.hazelcast.cache.impl.AbstractCacheRecordStore.<init>(AbstractCacheRecordStore.java:150)
at com.hazelcast.cache.hidensity.impl.nativememory.HiDensityNativeMemoryCacheRecordStore.<init>(HiDensityNativeMemoryCacheRecordStore.java:59)
at com.hazelcast.cache.EnterpriseCacheService.newNativeRecordStore(EnterpriseCacheService.java:243)
at com.hazelcast.cache.EnterpriseCacheService.createNewRecordStore(EnterpriseCacheService.java:217)
at com.hazelcast.cache.impl.CachePartitionSegment.createNew(CachePartitionSegment.java:51)
at com.hazelcast.cache.impl.CachePartitionSegment.createNew(CachePartitionSegment.java:37)
at com.hazelcast.util.ConcurrencyUtil.getOrPutSynchronized(ConcurrencyUtil.java:73)
at com.hazelcast.cache.impl.CachePartitionSegment.getOrCreateRecordStore(CachePartitionSegment.java:67)
at com.hazelcast.cache.impl.AbstractCacheService.getOrCreateRecordStore(AbstractCacheService.java:278)
```
|
defect
|
cache eviction not fail fast when using the following hz xml only on retrieving the cache i get a validation error but the xml file could be validated when loaded hazelcast xsi schemalocation xmlns xmlns xsi workers false native error com hazelcast simulator worker testcontainer testmanager global prepare of longstringcachetest failed java lang illegalargumentexception invalid max size policy entry count for com hazelcast cache hidensity impl nativememory hidensitynativememorycacherecordstore only used native memory size used native memory percentage free native memory size free native memory percentage are supported at com hazelcast cache hidensity impl nativememory hidensitynativememorycacherecordstore createcacheevictionchecker hidensitynativememorycacherecordstore java at com hazelcast cache impl abstractcacherecordstore abstractcacherecordstore java at com hazelcast cache hidensity impl nativememory hidensitynativememorycacherecordstore hidensitynativememorycacherecordstore java at com hazelcast cache enterprisecacheservice newnativerecordstore enterprisecacheservice java at com hazelcast cache enterprisecacheservice createnewrecordstore enterprisecacheservice java at com hazelcast cache impl cachepartitionsegment createnew cachepartitionsegment java at com hazelcast cache impl cachepartitionsegment createnew cachepartitionsegment java at com hazelcast util concurrencyutil getorputsynchronized concurrencyutil java at com hazelcast cache impl cachepartitionsegment getorcreaterecordstore cachepartitionsegment java at com hazelcast cache impl abstractcacheservice getorcreaterecordstore abstractcacheservice java
| 1
|
64,739
| 18,854,942,171
|
IssuesEvent
|
2021-11-12 04:11:21
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
TNC does not return optimal parameters
|
defect scipy.optimize
|
The truncated Newton (TNC) optimizer `scipy.optimize.minimize(method='TNC')` returns apparently not the _optimal_ parameter vector and function value, but the _last_ one, which often (depending on the dimension of the problem in 10-30% of cases in our applications) has a higher function value.
**Expected behavior**: It should return the optimal found parameters and function value.
This behavior was not observed for any other scipy optimizer considered.
#### Reproducing code example:
We encountered this problem in our parameter estimation tool [github.com/icb-dcm/pypesto](https://github.com/icb-dcm/pypesto), which employs an objective function which records a history of all function evaluations and can thus report if the best value in the trace is better than the value reported by the optimizer. See here for a discussion: https://github.com/ICB-DCM/pyPESTO/issues/327. The error can be reproduced in cell 5 of the notebook https://github.com/ICB-DCM/pyPESTO/blob/master/doc/example/rosenbrock.ipynb. The problem only has box constraints, which are not violated.
#### Error message:
```
Function values from history and optimizer do not match: 1.3168387678656086, 2.0179911928533514
Parameters obtained from history and optimizer do not match: [0.99553374 0.98995083 0.97544456 0.94768476 0.89550744 0.8019584
0.64185456 0.40180156 0.14794723 0.00844711], [0.9940831 0.98497745 0.96635783 0.93780681 0.87380595 0.74153588
0.53395112 0.25523609 0.03957006 0.00100739]
Function values from history and optimizer do not match: 1.6349167203078765, 2.336760892779954
```
#### Scipy/Numpy/Python version information:
```
1.4.1 / 1.18.4 / 3.7.4
```
|
1.0
|
TNC does not return optimal parameters - The truncated Newton (TNC) optimizer `scipy.optimize.minimize(method='TNC')` returns apparently not the _optimal_ parameter vector and function value, but the _last_ one, which often (depending on the dimension of the problem in 10-30% of cases in our applications) has a higher function value.
**Expected behavior**: It should return the optimal found parameters and function value.
This behavior was not observed for any other scipy optimizer considered.
#### Reproducing code example:
We encountered this problem in our parameter estimation tool [github.com/icb-dcm/pypesto](https://github.com/icb-dcm/pypesto), which employs an objective function which records a history of all function evaluations and can thus report if the best value in the trace is better than the value reported by the optimizer. See here for a discussion: https://github.com/ICB-DCM/pyPESTO/issues/327. The error can be reproduced in cell 5 of the notebook https://github.com/ICB-DCM/pyPESTO/blob/master/doc/example/rosenbrock.ipynb. The problem only has box constraints, which are not violated.
#### Error message:
```
Function values from history and optimizer do not match: 1.3168387678656086, 2.0179911928533514
Parameters obtained from history and optimizer do not match: [0.99553374 0.98995083 0.97544456 0.94768476 0.89550744 0.8019584
0.64185456 0.40180156 0.14794723 0.00844711], [0.9940831 0.98497745 0.96635783 0.93780681 0.87380595 0.74153588
0.53395112 0.25523609 0.03957006 0.00100739]
Function values from history and optimizer do not match: 1.6349167203078765, 2.336760892779954
```
#### Scipy/Numpy/Python version information:
```
1.4.1 / 1.18.4 / 3.7.4
```
|
defect
|
tnc does not return optimal parameters the truncated newton tnc optimizer scipy optimize minimize method tnc returns apparently not the optimal parameter vector and function value but the last one which often depending on the dimension of the problem in of cases in our applications has a higher function value expected behavior it should return the optimal found parameters and function value this behavior was not observed for any other scipy optimizer considered reproducing code example we encountered this problem in our parameter estimation tool which employs an objective function which records a history of all function evaluations and can thus report if the best value in the trace is better than the value reported by the optimizer see here for a discussion the error can be reproduced in cell of the notebook the problem only has box constraints which are not violated error message function values from history and optimizer do not match parameters obtained from history and optimizer do not match function values from history and optimizer do not match scipy numpy python version information
| 1
|
536,618
| 15,711,415,117
|
IssuesEvent
|
2021-03-27 07:07:10
|
nhcarrigan/code-snippets
|
https://api.github.com/repos/nhcarrigan/code-snippets
|
closed
|
[FEAT] - Add external integration notes
|
⭐ goal: addition 🏁 status: ready for dev 📄 aspect: text 🔒 staff only 🟧 priority: high
|
# Feature Request
## Is your feature request related to a problem? Please describe:
Need to document the external integrations used in repos.
<!--A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]-->
## Describe the solution you'd like
- LGTM
- - Repo is automatically added
- - Need to re-run build after initial project setup
- Code Climate
- - Need to add repo manually
- - Enable PR comments and status updates
- - Add `**/.spec.ts` to ignore settings (for Angular - because `tests` directory is already ignored)
<!--A clear and concise description of what you want to happen.-->
## Additional context
<!--Add any other context or screenshots about the feature request here.-->
|
1.0
|
[FEAT] - Add external integration notes - # Feature Request
## Is your feature request related to a problem? Please describe:
Need to document the external integrations used in repos.
<!--A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]-->
## Describe the solution you'd like
- LGTM
- - Repo is automatically added
- - Need to re-run build after initial project setup
- Code Climate
- - Need to add repo manually
- - Enable PR comments and status updates
- - Add `**/.spec.ts` to ignore settings (for Angular - because `tests` directory is already ignored)
<!--A clear and concise description of what you want to happen.-->
## Additional context
<!--Add any other context or screenshots about the feature request here.-->
|
non_defect
|
add external integration notes feature request is your feature request related to a problem please describe need to document the external integrations used in repos describe the solution you d like lgtm repo is automatically added need to re run build after initial project setup code climate need to add repo manually enable pr comments and status updates add spec ts to ignore settings for angular because tests directory is already ignored additional context
| 0
|
72,583
| 24,189,595,489
|
IssuesEvent
|
2022-09-23 16:09:50
|
vector-im/element-call
|
https://api.github.com/repos/vector-im/element-call
|
closed
|
Camera is left on after hanging up
|
T-Defect S-Minor O-Frequent X-Regression Privacy
|
### Steps to reproduce
1. Have a call
2. Press hangup button
3. Observe that camera light is still on
### Outcome
Looks like https://github.com/vector-im/element-call/issues/149 has regressed
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Will you send logs?
No
|
1.0
|
Camera is left on after hanging up - ### Steps to reproduce
1. Have a call
2. Press hangup button
3. Observe that camera light is still on
### Outcome
Looks like https://github.com/vector-im/element-call/issues/149 has regressed
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Will you send logs?
No
|
defect
|
camera is left on after hanging up steps to reproduce have a call press hangup button observe that camera light is still on outcome looks like has regressed operating system no response browser information no response url for webapp no response will you send logs no
| 1
|
103,052
| 16,601,965,029
|
IssuesEvent
|
2021-06-01 20:50:49
|
samq-ghdemo/SEARCH-NCJIS-nibrs
|
https://api.github.com/repos/samq-ghdemo/SEARCH-NCJIS-nibrs
|
opened
|
CVE-2020-26939 (Medium) detected in bcprov-jdk15on-1.54.jar
|
security vulnerability
|
## CVE-2020-26939 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.54.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p>
<p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p>
<p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-route/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar</p>
<p>
Dependency Hierarchy:
- :x: **bcprov-jdk15on-1.54.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/SEARCH-NCJIS-nibrs/commit/2643373aa9a184ff4ea81e98caf4009bf2ee8e91">2643373aa9a184ff4ea81e98caf4009bf2ee8e91</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Legion of the Bouncy Castle BC before 1.61 and BC-FJA before 1.0.1.2, attackers can obtain sensitive information about a private exponent because of Observable Differences in Behavior to Error Inputs. This occurs in org.bouncycastle.crypto.encodings.OAEPEncoding. Sending invalid ciphertext that decrypts to a short payload in the OAEP Decoder could result in the throwing of an early exception, potentially leaking some information about the private exponent of the RSA private key performing the encryption.
<p>Publish Date: 2020-11-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26939>CVE-2020-26939</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/bcgit/bc-java/wiki/CVE-2020-26939">https://github.com/bcgit/bc-java/wiki/CVE-2020-26939</a></p>
<p>Release Date: 2020-10-11</p>
<p>Fix Resolution: org.bouncycastle:bcprov-jdk14:1.61,org.bouncycastle:bcprov-ext-debug-jdk15on:1.61,org.bouncycastle:bcprov-debug-jdk15on:1.61,org.bouncycastle:bcprov-ext-jdk15on:1.61,org.bouncycastle:bcprov-jdk15on:1.61</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.bouncycastle","packageName":"bcprov-jdk15on","packageVersion":"1.54","packageFilePaths":["/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-validate-common/pom.xml","/tools/nibrs-fbi-service/pom.xml","/tools/nibrs-common/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-summary-report-common/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-flatfile/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-validation/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.bouncycastle:bcprov-jdk15on:1.54","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.bouncycastle:bcprov-jdk14:1.61,org.bouncycastle:bcprov-ext-debug-jdk15on:1.61,org.bouncycastle:bcprov-debug-jdk15on:1.61,org.bouncycastle:bcprov-ext-jdk15on:1.61,org.bouncycastle:bcprov-jdk15on:1.61"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-26939","vulnerabilityDetails":"In Legion of the Bouncy Castle BC before 1.61 and BC-FJA before 1.0.1.2, attackers can obtain sensitive information about a private exponent because of Observable Differences in Behavior to Error Inputs. This occurs in org.bouncycastle.crypto.encodings.OAEPEncoding. Sending invalid ciphertext that decrypts to a short payload in the OAEP Decoder could result in the throwing of an early exception, potentially leaking some information about the private exponent of the RSA private key performing the encryption.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26939","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-26939 (Medium) detected in bcprov-jdk15on-1.54.jar - ## CVE-2020-26939 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.54.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p>
<p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p>
<p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-route/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar</p>
<p>
Dependency Hierarchy:
- :x: **bcprov-jdk15on-1.54.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/SEARCH-NCJIS-nibrs/commit/2643373aa9a184ff4ea81e98caf4009bf2ee8e91">2643373aa9a184ff4ea81e98caf4009bf2ee8e91</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Legion of the Bouncy Castle BC before 1.61 and BC-FJA before 1.0.1.2, attackers can obtain sensitive information about a private exponent because of Observable Differences in Behavior to Error Inputs. This occurs in org.bouncycastle.crypto.encodings.OAEPEncoding. Sending invalid ciphertext that decrypts to a short payload in the OAEP Decoder could result in the throwing of an early exception, potentially leaking some information about the private exponent of the RSA private key performing the encryption.
<p>Publish Date: 2020-11-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26939>CVE-2020-26939</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/bcgit/bc-java/wiki/CVE-2020-26939">https://github.com/bcgit/bc-java/wiki/CVE-2020-26939</a></p>
<p>Release Date: 2020-10-11</p>
<p>Fix Resolution: org.bouncycastle:bcprov-jdk14:1.61,org.bouncycastle:bcprov-ext-debug-jdk15on:1.61,org.bouncycastle:bcprov-debug-jdk15on:1.61,org.bouncycastle:bcprov-ext-jdk15on:1.61,org.bouncycastle:bcprov-jdk15on:1.61</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.bouncycastle","packageName":"bcprov-jdk15on","packageVersion":"1.54","packageFilePaths":["/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-validate-common/pom.xml","/tools/nibrs-fbi-service/pom.xml","/tools/nibrs-common/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-summary-report-common/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-flatfile/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-validation/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.bouncycastle:bcprov-jdk15on:1.54","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.bouncycastle:bcprov-jdk14:1.61,org.bouncycastle:bcprov-ext-debug-jdk15on:1.61,org.bouncycastle:bcprov-debug-jdk15on:1.61,org.bouncycastle:bcprov-ext-jdk15on:1.61,org.bouncycastle:bcprov-jdk15on:1.61"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-26939","vulnerabilityDetails":"In Legion of the Bouncy Castle BC before 1.61 and BC-FJA before 1.0.1.2, attackers can obtain sensitive information about a private exponent because of Observable Differences in Behavior to Error Inputs. This occurs in org.bouncycastle.crypto.encodings.OAEPEncoding. Sending invalid ciphertext that decrypts to a short payload in the OAEP Decoder could result in the throwing of an early exception, potentially leaking some information about the private exponent of the RSA private key performing the encryption.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26939","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve medium detected in bcprov jar cve medium severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for jdk to jdk library home page a href path to dependency file search ncjis nibrs tools nibrs route pom xml path to vulnerable library home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar search ncjis nibrs web nibrs web target nibrs web web inf lib bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar dependency hierarchy x bcprov jar vulnerable library found in head commit a href found in base branch master vulnerability details in legion of the bouncy castle bc before and bc fja before attackers can obtain sensitive information about a private exponent because of observable differences in behavior to error inputs this occurs in org bouncycastle crypto encodings oaepencoding sending invalid ciphertext that decrypts to a short payload in the oaep decoder could result in the throwing of an early exception potentially leaking some information about the private exponent of the rsa private key performing the encryption publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org bouncycastle bcprov org bouncycastle bcprov ext debug org bouncycastle bcprov debug org bouncycastle bcprov ext org bouncycastle bcprov check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org bouncycastle bcprov isminimumfixversionavailable true minimumfixversion org bouncycastle bcprov org bouncycastle bcprov ext debug org bouncycastle bcprov debug org bouncycastle bcprov ext org bouncycastle bcprov basebranches vulnerabilityidentifier cve vulnerabilitydetails in legion of the bouncy castle bc before and bc fja before attackers can obtain sensitive information about a private exponent because of observable differences in behavior to error inputs this occurs in org bouncycastle crypto encodings oaepencoding sending invalid ciphertext that decrypts to a short payload in the oaep decoder could result in the throwing of an early exception potentially leaking some information about the private exponent of the rsa private key performing the encryption vulnerabilityurl
| 0
|
1,113
| 2,595,144,187
|
IssuesEvent
|
2015-02-20 11:52:31
|
keyboardsurfer/blinkendroid
|
https://api.github.com/repos/keyboardsurfer/blinkendroid
|
opened
|
kein ticketupdate&Wakelock nach back von playerview
|
auto-migrated Priority-Critical Type-Defect
|
```
wenn der server crashed oder beendet wird und man drückt den backbutton, dann
wird die loginactivity mit der ticketliste nicht mehr geupdated und auch das
wakelock geht nicht mehr
```
-----
Original issue reported on code.google.com by `lischke@gmail.com` on 31 Oct 2010 at 5:19
|
1.0
|
kein ticketupdate&Wakelock nach back von playerview - ```
wenn der server crashed oder beendet wird und man drückt den backbutton, dann
wird die loginactivity mit der ticketliste nicht mehr geupdated und auch das
wakelock geht nicht mehr
```
-----
Original issue reported on code.google.com by `lischke@gmail.com` on 31 Oct 2010 at 5:19
|
defect
|
kein ticketupdate wakelock nach back von playerview wenn der server crashed oder beendet wird und man drückt den backbutton dann wird die loginactivity mit der ticketliste nicht mehr geupdated und auch das wakelock geht nicht mehr original issue reported on code google com by lischke gmail com on oct at
| 1
|
55,537
| 6,485,247,796
|
IssuesEvent
|
2017-08-19 08:20:53
|
haskell-tools/haskell-tools
|
https://api.github.com/repos/haskell-tools/haskell-tools
|
opened
|
Flag configuration can prevent refactoring
|
category:bug origin:stackage-testing package:daemon type:project-handling
|
In the package NineP, there is flag that should be specified but somehow it seems it isn't. The dependencies look like:
```
if flag(bytestring-in-base)
-- bytestring was in base-2.0 and 2.1.1
build-depends: base >= 2.0 && < 2.2, binary >= 0.5.0.2
else
-- inbase 1.0 and 3.0 bytestring is a separate package
build-depends: base < 2.0 || >= 3, bytestring >= 0.9, binary >= 0.5.0.2
```
|
1.0
|
Flag configuration can prevent refactoring - In the package NineP, there is flag that should be specified but somehow it seems it isn't. The dependencies look like:
```
if flag(bytestring-in-base)
-- bytestring was in base-2.0 and 2.1.1
build-depends: base >= 2.0 && < 2.2, binary >= 0.5.0.2
else
-- inbase 1.0 and 3.0 bytestring is a separate package
build-depends: base < 2.0 || >= 3, bytestring >= 0.9, binary >= 0.5.0.2
```
|
non_defect
|
flag configuration can prevent refactoring in the package ninep there is flag that should be specified but somehow it seems it isn t the dependencies look like if flag bytestring in base bytestring was in base and build depends base else inbase and bytestring is a separate package build depends base bytestring binary
| 0
|
37,562
| 8,434,463,891
|
IssuesEvent
|
2018-10-17 10:15:24
|
scalameta/metals
|
https://api.github.com/repos/scalameta/metals
|
closed
|
Presentation compiler does not pick up new changes in dependent module
|
defect
|
* clean compile project with two modules: a and b
* open editor in module a
* add new `class A` in module a
* re-compile module a so that classfiles for `A` are generated
* open buffer in module b that depends on a
* write `new A`, compilation error
I tried doing `Global.askReset()` and it seems to fix this problem. The presentation compiler picked up the change, but I think we should try to figure out when is the best time to run `askReset`.
|
1.0
|
Presentation compiler does not pick up new changes in dependent module - * clean compile project with two modules: a and b
* open editor in module a
* add new `class A` in module a
* re-compile module a so that classfiles for `A` are generated
* open buffer in module b that depends on a
* write `new A`, compilation error
I tried doing `Global.askReset()` and it seems to fix this problem. The presentation compiler picked up the change, but I think we should try to figure out when is the best time to run `askReset`.
|
defect
|
presentation compiler does not pick up new changes in dependent module clean compile project with two modules a and b open editor in module a add new class a in module a re compile module a so that classfiles for a are generated open buffer in module b that depends on a write new a compilation error i tried doing global askreset and it seems to fix this problem the presentation compiler picked up the change but i think we should try to figure out when is the best time to run askreset
| 1
|
9,013
| 2,615,120,769
|
IssuesEvent
|
2015-03-01 05:47:09
|
chrsmith/google-api-java-client
|
https://api.github.com/repos/chrsmith/google-api-java-client
|
closed
|
samples repo has obsolete maven repo inside it with stall buzz library
|
auto-migrated Component-Google-APIs Priority-Medium Type-Defect
|
```
The mavenrepo tree should be pruned out of google-api-java-client-samples.
```
Original issue reported on code.google.com by `ai...@google.com` on 1 Jun 2011 at 9:18
|
1.0
|
samples repo has obsolete maven repo inside it with stall buzz library - ```
The mavenrepo tree should be pruned out of google-api-java-client-samples.
```
Original issue reported on code.google.com by `ai...@google.com` on 1 Jun 2011 at 9:18
|
defect
|
samples repo has obsolete maven repo inside it with stall buzz library the mavenrepo tree should be pruned out of google api java client samples original issue reported on code google com by ai google com on jun at
| 1
|
39,422
| 9,449,926,318
|
IssuesEvent
|
2019-04-16 04:10:57
|
supertuxkart/stk-code
|
https://api.github.com/repos/supertuxkart/stk-code
|
closed
|
Minor issues in Black Forest
|
C: 3D Modelling P4: minor R: fixed T: defect
|
A few very minor issues
- [x] Wrong zipper arrow direction in reverse. Apply to the reverse-only jump pad and to the texture in the bottom of the road leading to the village on the hill. ~~**Fixed for the jump pad in SVN but NOT for the zipper in the bottom of the road from the hill**~~ **EDIT: now fixed there too**

- [x] The houses of the village are not solid for physics, so you can go through their walls and see grass...
- [ ] Texture transitions are notoriously difficult, but ideally when the road split into two the "border" part of the texture should not go in the middle of the side road.

- [x] The names on the placards are not there. I think this is a texturing issue. See the screenshot above and look at the signs.
- [x] See the picture below.

- [x] At the end of the monastery path on the right, the tree with the light-colored trunk has a buggy hitbox which extends well beyond the visible trunk.
- [x] There is a tree floating in the air near the monastery path (drive in reverse to spot it, just right of the "stairs")
|
1.0
|
Minor issues in Black Forest - A few very minor issues
- [x] Wrong zipper arrow direction in reverse. Apply to the reverse-only jump pad and to the texture in the bottom of the road leading to the village on the hill. ~~**Fixed for the jump pad in SVN but NOT for the zipper in the bottom of the road from the hill**~~ **EDIT: now fixed there too**

- [x] The houses of the village are not solid for physics, so you can go through their walls and see grass...
- [ ] Texture transitions are notoriously difficult, but ideally when the road split into two the "border" part of the texture should not go in the middle of the side road.

- [x] The names on the placards are not there. I think this is a texturing issue. See the screenshot above and look at the signs.
- [x] See the picture below.

- [x] At the end of the monastery path on the right, the tree with the light-colored trunk has a buggy hitbox which extends well beyond the visible trunk.
- [x] There is a tree floating in the air near the monastery path (drive in reverse to spot it, just right of the "stairs")
|
defect
|
minor issues in black forest a few very minor issues wrong zipper arrow direction in reverse apply to the reverse only jump pad and to the texture in the bottom of the road leading to the village on the hill fixed for the jump pad in svn but not for the zipper in the bottom of the road from the hill edit now fixed there too the houses of the village are not solid for physics so you can go through their walls and see grass texture transitions are notoriously difficult but ideally when the road split into two the border part of the texture should not go in the middle of the side road the names on the placards are not there i think this is a texturing issue see the screenshot above and look at the signs see the picture below at the end of the monastery path on the right the tree with the light colored trunk has a buggy hitbox which extends well beyond the visible trunk there is a tree floating in the air near the monastery path drive in reverse to spot it just right of the stairs
| 1
|
6,435
| 2,610,243,111
|
IssuesEvent
|
2015-02-26 19:17:20
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
台州割包皮包茎手术哪家医院专业
|
auto-migrated Priority-Medium Type-Defect
|
```
台州割包皮包茎手术哪家医院专业【台州五洲生殖医院】24小
时健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院�
��址:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘�
��104、108、118、198及椒江一金清公交车直达枫南小区,乘坐107
、105、109、112、901、
902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 1:34
|
1.0
|
台州割包皮包茎手术哪家医院专业 - ```
台州割包皮包茎手术哪家医院专业【台州五洲生殖医院】24小
时健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院�
��址:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘�
��104、108、118、198及椒江一金清公交车直达枫南小区,乘坐107
、105、109、112、901、
902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 1:34
|
defect
|
台州割包皮包茎手术哪家医院专业 台州割包皮包茎手术哪家医院专业【台州五洲生殖医院】 时健康咨询热线 微信号tzwzszyy 医院� ��址 (枫南大转盘旁)乘车线路 乘� �� 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
47,709
| 13,066,127,238
|
IssuesEvent
|
2020-07-30 21:02:57
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade (Trac #1030)
|
Migrated from Trac combo core defect
|
When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.
An example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3
```text
In [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy
Out[43]: 9226.722570765558
In [44]: dataclasses.get_most_energetic_cascade(tree).energy
Out[44]: 16.27508624107153
```
This issue occured in icerec trunk, r133025.
Migrated from https://code.icecube.wisc.edu/ticket/1030
```json
{
"status": "closed",
"changetime": "2015-08-10T20:31:17",
"description": "When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.\n\nAn example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3\n\n\n{{{\nIn [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy\nOut[43]: 9226.722570765558\n\nIn [44]: dataclasses.get_most_energetic_cascade(tree).energy\nOut[44]: 16.27508624107153\n}}}\n\nThis issue occured in icerec trunk, r133025.\n",
"reporter": "hbretz",
"cc": "david.schultz@icecube.wisc.edu",
"resolution": "fixed",
"_ts": "1439238677802127",
"component": "combo core",
"summary": "dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade",
"priority": "blocker",
"keywords": "",
"time": "2015-06-24T15:41:58",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
|
1.0
|
dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade (Trac #1030) - When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.
An example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3
```text
In [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy
Out[43]: 9226.722570765558
In [44]: dataclasses.get_most_energetic_cascade(tree).energy
Out[44]: 16.27508624107153
```
This issue occured in icerec trunk, r133025.
Migrated from https://code.icecube.wisc.edu/ticket/1030
```json
{
"status": "closed",
"changetime": "2015-08-10T20:31:17",
"description": "When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.\n\nAn example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3\n\n\n{{{\nIn [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy\nOut[43]: 9226.722570765558\n\nIn [44]: dataclasses.get_most_energetic_cascade(tree).energy\nOut[44]: 16.27508624107153\n}}}\n\nThis issue occured in icerec trunk, r133025.\n",
"reporter": "hbretz",
"cc": "david.schultz@icecube.wisc.edu",
"resolution": "fixed",
"_ts": "1439238677802127",
"component": "combo core",
"summary": "dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade",
"priority": "blocker",
"keywords": "",
"time": "2015-06-24T15:41:58",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
|
defect
|
dataclasses get most energetic cascade returns cascade with lower energy than dataclasses get most energetic inice cascade trac when using dataclasses get most energetic cascade it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses get most energetic inice cascade an example file where this happens e g for the first event can be found at data user hbretz cascades vs inice cascades text in dataclasses get most energetic inice cascade tree energy out in dataclasses get most energetic cascade tree energy out this issue occured in icerec trunk migrated from json status closed changetime description when using dataclasses get most energetic cascade it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses get most energetic inice cascade n nan example file where this happens e g for the first event can be found at data user hbretz cascades vs inice cascades n n n nin dataclasses get most energetic inice cascade tree energy nout n nin dataclasses get most energetic cascade tree energy nout n n nthis issue occured in icerec trunk n reporter hbretz cc david schultz icecube wisc edu resolution fixed ts component combo core summary dataclasses get most energetic cascade returns cascade with lower energy than dataclasses get most energetic inice cascade priority blocker keywords time milestone owner olivas type defect
| 1
|
9,674
| 2,615,165,107
|
IssuesEvent
|
2015-03-01 06:45:15
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
Non accurate progression pourcentage
|
auto-migrated Priority-Triage Type-Defect
|
```
0. What version of Reaver are you using?
Reaver v1.4
1. What operating system are you using?
Backtrack 5r3 - Linux 3.2.6
2. Is your wireless card in monitor mode?
Chipset: Atheros AR927
Driver: ath9k
Monitor mode: YES
3. What is the signal strength of the Access Point you are trying to crack?
According to airodump-ng : -72
4. What is the manufacturer and model # of the device you are trying to crack?
Sagem F@st 3304
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b 7C:03:4C:49:CB:FA -e SAGEM_CBF9 -c 6 -N -S -A -v
(I use aireplay-ng for the association process and manualy set the channel
using:)
iwconfig mon0 channel 6
aireplay-ng -1 5000 -a 7C:03:4C:49:CB:FA -h 54:e6:fc:xx:xx:xx -e SAGEM_CBF9 mon0
6. Please describe what you think the issue is.
The progression pourcentage is incorrect
39.94%
39.98%
90.93%
90.95%
(see the output below)
7. Paste the output from Reaver below.
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[?] Restore previous session for 7C:03:4C:49:CB:FA? [n/Y] y
[+] Restored previous session
[+] Waiting for beacon from 7C:03:4C:49:CB:FA
[+] Associated with 7C:03:4C:49:CB:FA (ESSID: SAGEM_CBF9)
[+] 39.35% complete @ 2013-01-08 14:58:46 (2 seconds/pin)
[+] Trying pin 43225675
[+] Trying pin 43235674
[+] Trying pin 43245673
[+] Trying pin 43255672
[+] Trying pin 43265671
[+] 39.39% complete @ 2013-01-08 14:58:58 (2 seconds/pin)
[+] Trying pin 43275670
[+] Trying pin 43285679
[+] Trying pin 43295678
[+] Trying pin 43305674
[+] Trying pin 43315673
[+] 39.44% complete @ 2013-01-08 14:59:10 (2 seconds/pin)
[+] Trying pin 43325672
[+] Trying pin 43335671
[+] Trying pin 43345670
[+] Trying pin 43355679
[+] Trying pin 43365678
[+] 39.48% complete @ 2013-01-08 14:59:21 (2 seconds/pin)
[+] Trying pin 43375677
[+] Trying pin 43385676
[+] Trying pin 43395675
[+] Trying pin 43405671
[+] Trying pin 43415670
[+] 39.53% complete @ 2013-01-08 14:59:33 (2 seconds/pin)
[+] Trying pin 43425679
[+] Trying pin 43435678
[+] Trying pin 43445677
[+] Trying pin 43455676
[+] Trying pin 43465675
[+] 39.57% complete @ 2013-01-08 14:59:44 (2 seconds/pin)
[+] Trying pin 43475674
[+] Trying pin 43485673
[+] Trying pin 43495672
[+] Trying pin 43505678
[+] Trying pin 43515677
[+] 39.62% complete @ 2013-01-08 14:59:56 (2 seconds/pin)
[+] Trying pin 43525676
[+] Trying pin 43535675
[+] Trying pin 43545674
[+] Trying pin 43555673
[+] Trying pin 43565672
[+] 39.66% complete @ 2013-01-08 15:00:09 (2 seconds/pin)
[+] Trying pin 43575671
[+] Trying pin 43585670
[+] Trying pin 43595679
[+] Trying pin 43605675
[+] Trying pin 43615674
[+] 39.71% complete @ 2013-01-08 15:00:20 (2 seconds/pin)
[+] Trying pin 43625673
[+] Trying pin 43635672
[+] Trying pin 43645671
[+] Trying pin 43655670
[+] Trying pin 43665679
[+] 39.75% complete @ 2013-01-08 15:00:32 (2 seconds/pin)
[+] Trying pin 43675678
[+] Trying pin 43685677
[+] Trying pin 43695676
[+] Trying pin 43705672
[+] Trying pin 43715671
[+] 39.80% complete @ 2013-01-08 15:00:43 (2 seconds/pin)
[+] Trying pin 43725670
[+] Trying pin 43735679
[+] Trying pin 43745678
[+] Trying pin 43755677
[+] Trying pin 43765676
[+] 39.85% complete @ 2013-01-08 15:00:55 (2 seconds/pin)
[+] Trying pin 43775675
[+] Trying pin 43785674
[+] Trying pin 43795673
[+] Trying pin 43805679
[+] Trying pin 43815678
[+] 39.89% complete @ 2013-01-08 15:01:06 (2 seconds/pin)
[+] Trying pin 43825677
[+] Trying pin 43835676
[+] Trying pin 43845675
[+] Trying pin 43855674
[+] Trying pin 43865673
[+] 39.94% complete @ 2013-01-08 15:01:18 (2 seconds/pin)
[+] Trying pin 43875672
[+] Trying pin 43885671
[+] Trying pin 43895670
[+] Trying pin 43905676
[+] Trying pin 43915675
[+] 39.98% complete @ 2013-01-08 15:01:30 (2 seconds/pin)
[+] Trying pin 43925674
[+] Trying pin 43935673
[+] Trying pin 43945672
[+] Trying pin 43940004
[+] Trying pin 43940004
[+] 90.93% complete @ 2013-01-08 15:01:45 (2 seconds/pin)
[+] Trying pin 43942220
[+] Trying pin 43942220
[+] Trying pin 43943333
[+] Trying pin 43943333
[+] Trying pin 43944446
[+] 90.95% complete @ 2013-01-08 15:02:06 (2 seconds/pin)
[+] Trying pin 43944446
[+] Trying pin 43945559
[+] Trying pin 43945559
[+] Trying pin 43946662
[+] Trying pin 43946662
[+] 90.97% complete @ 2013-01-08 15:02:25 (2 seconds/pin)
[+] Trying pin 43947775
[+] Trying pin 43947775
[+] Trying pin 43948888
[+] Trying pin 43948888
[+] Trying pin 43949991
[+] 90.99% complete @ 2013-01-08 15:02:46 (2 seconds/pin)
[+] Trying pin 43949991
[+] Trying pin 43940011
[+] Trying pin 43940011
[+] Trying pin 43940028
[+] Trying pin 43940028
[+] 91.02% complete @ 2013-01-08 15:03:05 (2 seconds/pin)
[+] Trying pin 43940035
[+] Trying pin 43940035
[+] Trying pin 43940042
[+] Trying pin 43940042
[+] Trying pin 43940059
[+] 91.04% complete @ 2013-01-08 15:03:26 (2 seconds/pin)
[+] Trying pin 43940059
[+] Trying pin 43940066
[+] Trying pin 43940066
[+] Trying pin 43940073
[+] Trying pin 43940073
[+] 91.06% complete @ 2013-01-08 15:03:45 (2 seconds/pin)
[+] Trying pin 43940080
[+] Trying pin 43940080
[+] Trying pin 43940097
[+] Trying pin 43940097
[+] Trying pin 43940103
[+] 91.08% complete @ 2013-01-08 15:04:06 (2 seconds/pin)
[+] Trying pin 43940103
[+] Trying pin 43940110
[+] Trying pin 43940110
[+] Trying pin 43940127
[+] Trying pin 43940127
[+] 91.11% complete @ 2013-01-08 15:04:25 (2 seconds/pin)
[+] Trying pin 43940134
[+] Trying pin 43940134
[+] Trying pin 43940141
[+] Trying pin 43940141
[+] Trying pin 43940158
[+] 91.13% complete @ 2013-01-08 15:04:46 (3 seconds/pin)
[+] Trying pin 43940158
[+] Trying pin 439401
```
Original issue reported on code.google.com by `i.boud...@gmail.com` on 8 Jan 2013 at 8:35
|
1.0
|
Non accurate progression pourcentage - ```
0. What version of Reaver are you using?
Reaver v1.4
1. What operating system are you using?
Backtrack 5r3 - Linux 3.2.6
2. Is your wireless card in monitor mode?
Chipset: Atheros AR927
Driver: ath9k
Monitor mode: YES
3. What is the signal strength of the Access Point you are trying to crack?
According to airodump-ng : -72
4. What is the manufacturer and model # of the device you are trying to crack?
Sagem F@st 3304
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b 7C:03:4C:49:CB:FA -e SAGEM_CBF9 -c 6 -N -S -A -v
(I use aireplay-ng for the association process and manualy set the channel
using:)
iwconfig mon0 channel 6
aireplay-ng -1 5000 -a 7C:03:4C:49:CB:FA -h 54:e6:fc:xx:xx:xx -e SAGEM_CBF9 mon0
6. Please describe what you think the issue is.
The progression pourcentage is incorrect
39.94%
39.98%
90.93%
90.95%
(see the output below)
7. Paste the output from Reaver below.
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[?] Restore previous session for 7C:03:4C:49:CB:FA? [n/Y] y
[+] Restored previous session
[+] Waiting for beacon from 7C:03:4C:49:CB:FA
[+] Associated with 7C:03:4C:49:CB:FA (ESSID: SAGEM_CBF9)
[+] 39.35% complete @ 2013-01-08 14:58:46 (2 seconds/pin)
[+] Trying pin 43225675
[+] Trying pin 43235674
[+] Trying pin 43245673
[+] Trying pin 43255672
[+] Trying pin 43265671
[+] 39.39% complete @ 2013-01-08 14:58:58 (2 seconds/pin)
[+] Trying pin 43275670
[+] Trying pin 43285679
[+] Trying pin 43295678
[+] Trying pin 43305674
[+] Trying pin 43315673
[+] 39.44% complete @ 2013-01-08 14:59:10 (2 seconds/pin)
[+] Trying pin 43325672
[+] Trying pin 43335671
[+] Trying pin 43345670
[+] Trying pin 43355679
[+] Trying pin 43365678
[+] 39.48% complete @ 2013-01-08 14:59:21 (2 seconds/pin)
[+] Trying pin 43375677
[+] Trying pin 43385676
[+] Trying pin 43395675
[+] Trying pin 43405671
[+] Trying pin 43415670
[+] 39.53% complete @ 2013-01-08 14:59:33 (2 seconds/pin)
[+] Trying pin 43425679
[+] Trying pin 43435678
[+] Trying pin 43445677
[+] Trying pin 43455676
[+] Trying pin 43465675
[+] 39.57% complete @ 2013-01-08 14:59:44 (2 seconds/pin)
[+] Trying pin 43475674
[+] Trying pin 43485673
[+] Trying pin 43495672
[+] Trying pin 43505678
[+] Trying pin 43515677
[+] 39.62% complete @ 2013-01-08 14:59:56 (2 seconds/pin)
[+] Trying pin 43525676
[+] Trying pin 43535675
[+] Trying pin 43545674
[+] Trying pin 43555673
[+] Trying pin 43565672
[+] 39.66% complete @ 2013-01-08 15:00:09 (2 seconds/pin)
[+] Trying pin 43575671
[+] Trying pin 43585670
[+] Trying pin 43595679
[+] Trying pin 43605675
[+] Trying pin 43615674
[+] 39.71% complete @ 2013-01-08 15:00:20 (2 seconds/pin)
[+] Trying pin 43625673
[+] Trying pin 43635672
[+] Trying pin 43645671
[+] Trying pin 43655670
[+] Trying pin 43665679
[+] 39.75% complete @ 2013-01-08 15:00:32 (2 seconds/pin)
[+] Trying pin 43675678
[+] Trying pin 43685677
[+] Trying pin 43695676
[+] Trying pin 43705672
[+] Trying pin 43715671
[+] 39.80% complete @ 2013-01-08 15:00:43 (2 seconds/pin)
[+] Trying pin 43725670
[+] Trying pin 43735679
[+] Trying pin 43745678
[+] Trying pin 43755677
[+] Trying pin 43765676
[+] 39.85% complete @ 2013-01-08 15:00:55 (2 seconds/pin)
[+] Trying pin 43775675
[+] Trying pin 43785674
[+] Trying pin 43795673
[+] Trying pin 43805679
[+] Trying pin 43815678
[+] 39.89% complete @ 2013-01-08 15:01:06 (2 seconds/pin)
[+] Trying pin 43825677
[+] Trying pin 43835676
[+] Trying pin 43845675
[+] Trying pin 43855674
[+] Trying pin 43865673
[+] 39.94% complete @ 2013-01-08 15:01:18 (2 seconds/pin)
[+] Trying pin 43875672
[+] Trying pin 43885671
[+] Trying pin 43895670
[+] Trying pin 43905676
[+] Trying pin 43915675
[+] 39.98% complete @ 2013-01-08 15:01:30 (2 seconds/pin)
[+] Trying pin 43925674
[+] Trying pin 43935673
[+] Trying pin 43945672
[+] Trying pin 43940004
[+] Trying pin 43940004
[+] 90.93% complete @ 2013-01-08 15:01:45 (2 seconds/pin)
[+] Trying pin 43942220
[+] Trying pin 43942220
[+] Trying pin 43943333
[+] Trying pin 43943333
[+] Trying pin 43944446
[+] 90.95% complete @ 2013-01-08 15:02:06 (2 seconds/pin)
[+] Trying pin 43944446
[+] Trying pin 43945559
[+] Trying pin 43945559
[+] Trying pin 43946662
[+] Trying pin 43946662
[+] 90.97% complete @ 2013-01-08 15:02:25 (2 seconds/pin)
[+] Trying pin 43947775
[+] Trying pin 43947775
[+] Trying pin 43948888
[+] Trying pin 43948888
[+] Trying pin 43949991
[+] 90.99% complete @ 2013-01-08 15:02:46 (2 seconds/pin)
[+] Trying pin 43949991
[+] Trying pin 43940011
[+] Trying pin 43940011
[+] Trying pin 43940028
[+] Trying pin 43940028
[+] 91.02% complete @ 2013-01-08 15:03:05 (2 seconds/pin)
[+] Trying pin 43940035
[+] Trying pin 43940035
[+] Trying pin 43940042
[+] Trying pin 43940042
[+] Trying pin 43940059
[+] 91.04% complete @ 2013-01-08 15:03:26 (2 seconds/pin)
[+] Trying pin 43940059
[+] Trying pin 43940066
[+] Trying pin 43940066
[+] Trying pin 43940073
[+] Trying pin 43940073
[+] 91.06% complete @ 2013-01-08 15:03:45 (2 seconds/pin)
[+] Trying pin 43940080
[+] Trying pin 43940080
[+] Trying pin 43940097
[+] Trying pin 43940097
[+] Trying pin 43940103
[+] 91.08% complete @ 2013-01-08 15:04:06 (2 seconds/pin)
[+] Trying pin 43940103
[+] Trying pin 43940110
[+] Trying pin 43940110
[+] Trying pin 43940127
[+] Trying pin 43940127
[+] 91.11% complete @ 2013-01-08 15:04:25 (2 seconds/pin)
[+] Trying pin 43940134
[+] Trying pin 43940134
[+] Trying pin 43940141
[+] Trying pin 43940141
[+] Trying pin 43940158
[+] 91.13% complete @ 2013-01-08 15:04:46 (3 seconds/pin)
[+] Trying pin 43940158
[+] Trying pin 439401
```
Original issue reported on code.google.com by `i.boud...@gmail.com` on 8 Jan 2013 at 8:35
|
defect
|
non accurate progression pourcentage what version of reaver are you using reaver what operating system are you using backtrack linux is your wireless card in monitor mode chipset atheros driver monitor mode yes what is the signal strength of the access point you are trying to crack according to airodump ng what is the manufacturer and model of the device you are trying to crack sagem f st what is the entire command line string you are supplying to reaver reaver i b cb fa e sagem c n s a v i use aireplay ng for the association process and manualy set the channel using iwconfig channel aireplay ng a cb fa h fc xx xx xx e sagem please describe what you think the issue is the progression pourcentage is incorrect see the output below paste the output from reaver below reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner restore previous session for cb fa y restored previous session waiting for beacon from cb fa associated with cb fa essid sagem complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin trying pin trying pin trying pin complete seconds pin trying pin trying pin original issue reported on code google com by i boud gmail com on jan at
| 1
|
82,054
| 10,219,374,311
|
IssuesEvent
|
2019-08-15 18:24:59
|
unlock-protocol/unlock
|
https://api.github.com/repos/unlock-protocol/unlock
|
closed
|
We Need a Consistent Grid
|
design
|
## Description
Currently different pages use different grids and rules, which don't follow the design and creating unsightly inconsistencies. Let's discuss and figure out a way forward. A consistent grid will help us look better but also make dev easier.

|
1.0
|
We Need a Consistent Grid - ## Description
Currently different pages use different grids and rules, which don't follow the design and creating unsightly inconsistencies. Let's discuss and figure out a way forward. A consistent grid will help us look better but also make dev easier.

|
non_defect
|
we need a consistent grid description currently different pages use different grids and rules which don t follow the design and creating unsightly inconsistencies let s discuss and figure out a way forward a consistent grid will help us look better but also make dev easier
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.