Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
226,790
| 18,044,168,614
|
IssuesEvent
|
2021-09-18 15:41:18
|
logicmoo/logicmoo_workspace
|
https://api.github.com/repos/logicmoo/logicmoo_workspace
|
opened
|
logicmoo.base.fol.fiveof.NONMONOTONIC_TYPE_01 JUnit
|
logicmoo.base.fol.fiveof Test_9999 unit_test NONMONOTONIC_TYPE_01
|
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif nonmonotonic_type_01.pl)
GH_MASTER_ISSUE_FINFO=
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANONMONOTONIC_TYPE_01
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.fol.fiveof/NONMONOTONIC_TYPE_01/logicmoo_base_fol_fiveof_NONMONOTONIC_TYPE_01_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/66/testReport/logicmoo.base.fol.fiveof/NONMONOTONIC_TYPE_01/logicmoo_base_fol_fiveof_NONMONOTONIC_TYPE_01_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl'),
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
%~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:96
%~ this_test_might_need( :-( expects_dialect(pfc)))
% =============================================
% File 'mpred_builtin.pfc'
% Purpose: Agent Reactivity for SWI-Prolog
% Maintainer: Douglas Miles
% Contact: $Author: dmiles $@users.sourceforge.net %
% Version: 'interface' 1.0.0
% Revision: $Revision: 1.9 $
% Revised At: $Date: 2002/06/27 14:13:20 $
% =============================================
%
:- module(baseKB).
:- process_script_file.
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl:23
%~ kifm = leftof(h1,h2).
%~ kifm = leftof(h1,h2).
%~ debugm(common_logic_loader,show_success(common_logic_loader,common_logic_loader:ain(clif(leftof(h1,h2)))))
%~ kifm = leftof(h1,h2).
%~ kifm = leftof(h2,h3).
%~ kifm = leftof(h2,h3).
%~ debugm(common_logic_loader,show_success(common_logic_loader,common_logic_loader:ain(clif(leftof(h2,h3)))))
%~ kifm = leftof(h2,h3).
%~ kifm = leftof(h3,h4).
%~ kifm = leftof(h3,h4).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl:25
%~ debugm(common_logic_loader,show_success(common_logic_loader,common_logic_loader:ain(clif(leftof(h3,h4)))))
%~ kifm = leftof(h3,h4).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl:26
%~ kifm = leftof(h4,h5).
%~ kifm = leftof(h4,h5).
%~ debugm(common_logic_loader,show_success(common_logic_loader,common_logic_loader:ain(clif(leftof(h4,h5)))))
%~ kifm = leftof(h4,h5).
%~ kifm = ( leftof(House_Leftof,House_Leftof3) =>
%~ house(House_Leftof)&house(House_Leftof3)).
%~ kifm = ( leftof(House_Leftof8,House_Leftof9) =>
%~ house(House_Leftof8)&house(House_Leftof9)).
```
totalTime=1
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANONMONOTONIC_TYPE_01
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.fol.fiveof/NONMONOTONIC_TYPE_01/logicmoo_base_fol_fiveof_NONMONOTONIC_TYPE_01_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/66/testReport/logicmoo.base.fol.fiveof/NONMONOTONIC_TYPE_01/logicmoo_base_fol_fiveof_NONMONOTONIC_TYPE_01_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k nonmonotonic_type_01.pl (returned 141)
|
2.0
|
logicmoo.base.fol.fiveof.NONMONOTONIC_TYPE_01 JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif nonmonotonic_type_01.pl)
GH_MASTER_ISSUE_FINFO=
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANONMONOTONIC_TYPE_01
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.fol.fiveof/NONMONOTONIC_TYPE_01/logicmoo_base_fol_fiveof_NONMONOTONIC_TYPE_01_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/66/testReport/logicmoo.base.fol.fiveof/NONMONOTONIC_TYPE_01/logicmoo_base_fol_fiveof_NONMONOTONIC_TYPE_01_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl'),
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
%~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:96
%~ this_test_might_need( :-( expects_dialect(pfc)))
% =============================================
% File 'mpred_builtin.pfc'
% Purpose: Agent Reactivity for SWI-Prolog
% Maintainer: Douglas Miles
% Contact: $Author: dmiles $@users.sourceforge.net %
% Version: 'interface' 1.0.0
% Revision: $Revision: 1.9 $
% Revised At: $Date: 2002/06/27 14:13:20 $
% =============================================
%
:- module(baseKB).
:- process_script_file.
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl:23
%~ kifm = leftof(h1,h2).
%~ kifm = leftof(h1,h2).
%~ debugm(common_logic_loader,show_success(common_logic_loader,common_logic_loader:ain(clif(leftof(h1,h2)))))
%~ kifm = leftof(h1,h2).
%~ kifm = leftof(h2,h3).
%~ kifm = leftof(h2,h3).
%~ debugm(common_logic_loader,show_success(common_logic_loader,common_logic_loader:ain(clif(leftof(h2,h3)))))
%~ kifm = leftof(h2,h3).
%~ kifm = leftof(h3,h4).
%~ kifm = leftof(h3,h4).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl:25
%~ debugm(common_logic_loader,show_success(common_logic_loader,common_logic_loader:ain(clif(leftof(h3,h4)))))
%~ kifm = leftof(h3,h4).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl:26
%~ kifm = leftof(h4,h5).
%~ kifm = leftof(h4,h5).
%~ debugm(common_logic_loader,show_success(common_logic_loader,common_logic_loader:ain(clif(leftof(h4,h5)))))
%~ kifm = leftof(h4,h5).
%~ kifm = ( leftof(House_Leftof,House_Leftof3) =>
%~ house(House_Leftof)&house(House_Leftof3)).
%~ kifm = ( leftof(House_Leftof8,House_Leftof9) =>
%~ house(House_Leftof8)&house(House_Leftof9)).
```
totalTime=1
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANONMONOTONIC_TYPE_01
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.fol.fiveof/NONMONOTONIC_TYPE_01/logicmoo_base_fol_fiveof_NONMONOTONIC_TYPE_01_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/66/testReport/logicmoo.base.fol.fiveof/NONMONOTONIC_TYPE_01/logicmoo_base_fol_fiveof_NONMONOTONIC_TYPE_01_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/logicmoo_base/t/examples/fol/fiveof/nonmonotonic_type_01.pl
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k nonmonotonic_type_01.pl (returned 141)
|
non_defect
|
logicmoo base fol fiveof nonmonotonic type junit cd var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol fiveof timeout foreground preserve status s sigkill k lmoo clif nonmonotonic type pl gh master issue finfo issue search gitlab latest this build github running var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol fiveof nonmonotonic type pl this test might need use module library logicmoo plarkc var lib jenkins local share swi prolog pack logicmoo utils prolog logicmoo test header pl this test might need expects dialect pfc file mpred builtin pfc purpose agent reactivity for swi prolog maintainer douglas miles contact author dmiles users sourceforge net version interface revision revision revised at date module basekb process script file var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol fiveof nonmonotonic type pl kifm leftof kifm leftof debugm common logic loader show success common logic loader common logic loader ain clif leftof kifm leftof kifm leftof kifm leftof debugm common logic loader show success common logic loader common logic loader ain clif leftof kifm leftof kifm leftof kifm leftof var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol fiveof nonmonotonic type pl debugm common logic loader show success common logic loader common logic loader ain clif leftof kifm leftof var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol fiveof nonmonotonic type pl kifm leftof kifm leftof debugm common logic loader show success common logic loader common logic loader ain clif leftof kifm leftof kifm leftof house leftof house house house leftof house house kifm leftof house house house house house house totaltime issue search gitlab latest this build github failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k nonmonotonic type pl returned
| 0
|
53,093
| 13,260,883,566
|
IssuesEvent
|
2020-08-20 18:55:40
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
port photonics_1.73 does not build on SL5 64 bit (Trac #686)
|
Migrated from Trac booking defect
|
Scientific Linux release 5.8 (Boron)
```text
$ gcc --version
gcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)
Copyright (C) 2006 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
```
```text
$ ./bin/port -vd install photonics_1.73
DEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73
DEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73
DEBUG: Requested variant x86_64 is not provided by port photonics_1.73.
DEBUG: Executing variant linux provides linux
DEBUG: Executing com.apple.main (photonics_1.73)
DEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz
DEBUG: Skipping unarchive (photonics_1.73) since no archive found
DEBUG: Skipping completed com.apple.unarchive (photonics_1.73)
---> Fetching photonics_1.73
DEBUG: Executing com.apple.fetch (photonics_1.73)
---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73
---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M
---> Verifying checksum(s) for photonics_1.73
DEBUG: Executing com.apple.checksum (photonics_1.73)
---> Checksumming photonics-1.73.tar.bz2
DEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2
DEBUG: setting option extract.cmd to /usr/bin/bzip2
---> Extracting photonics_1.73
DEBUG: Executing com.apple.extract (photonics_1.73)
---> Extracting photonics-1.73.tar.bz2
DEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'
DEBUG: Executing com.apple.patch (photonics_1.73)
---> Configuring photonics_1.73
DEBUG: Executing com.apple.configure (photonics_1.73)
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'
checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... yes
checking for a thread-safe mkdir -p... /bin/mkdir -p
checking for gawk... gawk
checking whether make sets $(MAKE)... yes
checking for gawk... (cached) gawk
checking for gcc... gcc
checking for C compiler default output file name... a.out
checking whether the C compiler works... yes
checking whether we are cross compiling... no
checking for suffix of executables...
checking for suffix of object files... o
checking whether we are using the GNU C compiler... yes
checking whether gcc accepts -g... yes
checking for gcc option to accept ISO C89... none needed
checking for style of include used by make... GNU
checking dependency style of gcc... gcc3
checking how to run the C preprocessor... gcc -E
checking for g++... g++
checking whether we are using the GNU C++ compiler... yes
checking whether g++ accepts -g... yes
checking dependency style of g++... gcc3
checking how to run the C++ preprocessor... g++ -E
checking for a BSD-compatible install... /usr/bin/install -c
checking whether ln -s works... yes
checking whether make sets $(MAKE)... (cached) yes
checking build system type... x86_64-unknown-linux-gnu
checking host system type... x86_64-unknown-linux-gnu
checking for a sed that does not truncate output... /bin/sed
checking for grep that handles long lines and -e... /bin/grep
checking for egrep... /bin/grep -E
checking for fgrep... /bin/grep -F
checking for ld used by gcc... /usr/bin/ld
checking if the linker (/usr/bin/ld) is GNU ld... yes
checking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B
checking the name lister (/usr/bin/nm -B) interface... BSD nm
checking the maximum length of command line arguments... 98304
checking whether the shell understands some XSI constructs... yes
checking whether the shell understands "+="... yes
checking for /usr/bin/ld option to reload object files... -r
checking how to recognize dependent libraries... pass_all
checking for ar... ar
checking for strip... strip
checking for ranlib... ranlib
checking command to parse /usr/bin/nm -B output from gcc object... ok
checking for ANSI C header files... yes
checking for sys/types.h... yes
checking for sys/stat.h... yes
checking for stdlib.h... yes
checking for string.h... yes
checking for memory.h... yes
checking for strings.h... yes
checking for inttypes.h... yes
checking for stdint.h... yes
checking for unistd.h... yes
checking for dlfcn.h... yes
checking whether we are using the GNU C++ compiler... (cached) yes
checking whether g++ accepts -g... (cached) yes
checking dependency style of g++... (cached) gcc3
checking how to run the C++ preprocessor... g++ -E
checking for objdir... .libs
checking if gcc supports -fno-rtti -fno-exceptions... no
checking for gcc option to produce PIC... -fPIC -DPIC
checking if gcc PIC flag -fPIC -DPIC works... yes
checking if gcc static flag -static works... yes
checking if gcc supports -c -o file.o... yes
checking if gcc supports -c -o file.o... (cached) yes
checking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking whether -lc should be explicitly linked in... no
checking dynamic linker characteristics... GNU/Linux ld.so
checking how to hardcode library paths into programs... immediate
checking whether stripping libraries is possible... yes
checking if libtool supports shared libraries... yes
checking whether to build shared libraries... yes
checking whether to build static libraries... yes
checking for ld used by g++... /usr/bin/ld -m elf_x86_64
checking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes
checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking for g++ option to produce PIC... -fPIC -DPIC
checking if g++ PIC flag -fPIC -DPIC works... yes
checking if g++ static flag -static works... yes
checking if g++ supports -c -o file.o... yes
checking if g++ supports -c -o file.o... (cached) yes
checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking dynamic linker characteristics... GNU/Linux ld.so
checking how to hardcode library paths into programs... immediate
checking for ANSI C header files... (cached) yes
checking for stdbool.h that conforms to C99... yes
checking for _Bool... yes
checking limits.h usability... yes
checking limits.h presence... yes
checking for limits.h... yes
checking malloc.h usability... yes
checking malloc.h presence... yes
checking for malloc.h... yes
checking for an ANSI C-conforming const... yes
checking for size_t... yes
checking for int8_t... yes
checking for int16_t... yes
checking for int32_t... yes
checking for int64_t... yes
checking for off_t... yes
checking for stdlib.h... (cached) yes
checking for GNU libc compatible malloc... yes
checking for stdlib.h... (cached) yes
checking for unistd.h... (cached) yes
checking for getpagesize... yes
checking for working mmap... yes
checking for stdlib.h... (cached) yes
checking for GNU libc compatible realloc... yes
checking for working strtod... yes
checking for strstr... yes
checking for strtod... (cached) yes
checking for strtol... yes
checking for strerror... yes
checking for memset... yes
checking for floor... no
checking for library containing floor... -lm
checking for pow... yes
checking for sqrt... yes
checking whether to enable debug mode... yes
disabled cernlib dependent code... yes
configure: creating ./config.status
config.status: creating Makefile
config.status: creating lib/Makefile
config.status: creating src/Makefile
config.status: creating ice/Makefile
config.status: creating scripts/Makefile
config.status: creating amasim/Makefile
config.status: creating level2/Makefile
config.status: creating config.h
config.status: executing depfiles commands
config.status: executing libtool commands
-------------------------------------------------------------------
Photonics: "1.73: pyrosoma r4"
please refer to the 'INSTALL' file for further instructions.
Hints:
Building photonics......................'make'
Compiling tool directory................'make tool'
Performing post compile test............'make tests'
All of the above........................'make everything'
Clean objects and binaries..............'make clean'
Clean tool directory....................'make toolclean'
Remove traces of previous configure.....'make distclean'
Also consider trying 'scripts/install_in_icetray.sh --help'
---> Building photonics_1.73 with target all
DEBUG: Executing com.apple.build (photonics_1.73)
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all'
make all-recursive
make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
Making all in lib
make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c
libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o
boundary.c:1: error: bad value (native) for -march= switch
boundary.c:1: error: bad value (native) for -mtune= switch
make[2]: *** [boundary.lo] Error 1
make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
make: *** [all] Error 2
Error: Target com.apple.build returned: shell command "cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all" returned error 2
Command output: make all-recursive
make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
Making all in lib
make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c
libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o
boundary.c:1: error: bad value (native) for -march= switch
boundary.c:1: error: bad value (native) for -mtune= switch
make[2]: *** [boundary.lo] Error 1
make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
make: *** [all] Error 2
Warning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install
icecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/686">https://code.icecube.wisc.edu/projects/icecube/ticket/686</a>, reported by boersma</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-06-22T16:05:07",
"_ts": "1340381107000000",
"description": "Scientific Linux release 5.8 (Boron)\n\n{{{\n$ gcc --version\ngcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)\nCopyright (C) 2006 Free Software Foundation, Inc.\nThis is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n}}}\n\n{{{\n$ ./bin/port -vd install photonics_1.73\nDEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Requested variant x86_64 is not provided by port photonics_1.73.\nDEBUG: Executing variant linux provides linux\nDEBUG: Executing com.apple.main (photonics_1.73)\nDEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz\nDEBUG: Skipping unarchive (photonics_1.73) since no archive found\nDEBUG: Skipping completed com.apple.unarchive (photonics_1.73)\n---> Fetching photonics_1.73\nDEBUG: Executing com.apple.fetch (photonics_1.73)\n---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\n---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M\n---> Verifying checksum(s) for photonics_1.73\nDEBUG: Executing com.apple.checksum (photonics_1.73)\n---> Checksumming photonics-1.73.tar.bz2\nDEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2\nDEBUG: setting option extract.cmd to /usr/bin/bzip2\n---> Extracting photonics_1.73\nDEBUG: Executing com.apple.extract (photonics_1.73)\n---> Extracting photonics-1.73.tar.bz2\nDEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work\" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'\nDEBUG: Executing com.apple.patch (photonics_1.73)\n---> Configuring photonics_1.73\nDEBUG: Executing com.apple.configure (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether build environment is sane... yes\nchecking for a thread-safe mkdir -p... /bin/mkdir -p\nchecking for gawk... gawk\nchecking whether make sets $(MAKE)... yes\nchecking for gawk... (cached) gawk\nchecking for gcc... gcc\nchecking for C compiler default output file name... a.out\nchecking whether the C compiler works... yes\nchecking whether we are cross compiling... no\nchecking for suffix of executables... \nchecking for suffix of object files... o\nchecking whether we are using the GNU C compiler... yes\nchecking whether gcc accepts -g... yes\nchecking for gcc option to accept ISO C89... none needed\nchecking for style of include used by make... GNU\nchecking dependency style of gcc... gcc3\nchecking how to run the C preprocessor... gcc -E\nchecking for g++... g++\nchecking whether we are using the GNU C++ compiler... yes\nchecking whether g++ accepts -g... yes\nchecking dependency style of g++... gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether ln -s works... yes\nchecking whether make sets $(MAKE)... (cached) yes\nchecking build system type... x86_64-unknown-linux-gnu\nchecking host system type... x86_64-unknown-linux-gnu\nchecking for a sed that does not truncate output... /bin/sed\nchecking for grep that handles long lines and -e... /bin/grep\nchecking for egrep... /bin/grep -E\nchecking for fgrep... /bin/grep -F\nchecking for ld used by gcc... /usr/bin/ld\nchecking if the linker (/usr/bin/ld) is GNU ld... yes\nchecking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B\nchecking the name lister (/usr/bin/nm -B) interface... BSD nm\nchecking the maximum length of command line arguments... 98304\nchecking whether the shell understands some XSI constructs... yes\nchecking whether the shell understands \"+=\"... yes\nchecking for /usr/bin/ld option to reload object files... -r\nchecking how to recognize dependent libraries... pass_all\nchecking for ar... ar\nchecking for strip... strip\nchecking for ranlib... ranlib\nchecking command to parse /usr/bin/nm -B output from gcc object... ok\nchecking for ANSI C header files... yes\nchecking for sys/types.h... yes\nchecking for sys/stat.h... yes\nchecking for stdlib.h... yes\nchecking for string.h... yes\nchecking for memory.h... yes\nchecking for strings.h... yes\nchecking for inttypes.h... yes\nchecking for stdint.h... yes\nchecking for unistd.h... yes\nchecking for dlfcn.h... yes\nchecking whether we are using the GNU C++ compiler... (cached) yes\nchecking whether g++ accepts -g... (cached) yes\nchecking dependency style of g++... (cached) gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for objdir... .libs\nchecking if gcc supports -fno-rtti -fno-exceptions... no\nchecking for gcc option to produce PIC... -fPIC -DPIC\nchecking if gcc PIC flag -fPIC -DPIC works... yes\nchecking if gcc static flag -static works... yes\nchecking if gcc supports -c -o file.o... yes\nchecking if gcc supports -c -o file.o... (cached) yes\nchecking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking whether -lc should be explicitly linked in... no\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking whether stripping libraries is possible... yes\nchecking if libtool supports shared libraries... yes\nchecking whether to build shared libraries... yes\nchecking whether to build static libraries... yes\nchecking for ld used by g++... /usr/bin/ld -m elf_x86_64\nchecking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking for g++ option to produce PIC... -fPIC -DPIC\nchecking if g++ PIC flag -fPIC -DPIC works... yes\nchecking if g++ static flag -static works... yes\nchecking if g++ supports -c -o file.o... yes\nchecking if g++ supports -c -o file.o... (cached) yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking for ANSI C header files... (cached) yes\nchecking for stdbool.h that conforms to C99... yes\nchecking for _Bool... yes\nchecking limits.h usability... yes\nchecking limits.h presence... yes\nchecking for limits.h... yes\nchecking malloc.h usability... yes\nchecking malloc.h presence... yes\nchecking for malloc.h... yes\nchecking for an ANSI C-conforming const... yes\nchecking for size_t... yes\nchecking for int8_t... yes\nchecking for int16_t... yes\nchecking for int32_t... yes\nchecking for int64_t... yes\nchecking for off_t... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible malloc... yes\nchecking for stdlib.h... (cached) yes\nchecking for unistd.h... (cached) yes\nchecking for getpagesize... yes\nchecking for working mmap... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible realloc... yes\nchecking for working strtod... yes\nchecking for strstr... yes\nchecking for strtod... (cached) yes\nchecking for strtol... yes\nchecking for strerror... yes\nchecking for memset... yes\nchecking for floor... no\nchecking for library containing floor... -lm\nchecking for pow... yes\nchecking for sqrt... yes\nchecking whether to enable debug mode... yes\ndisabled cernlib dependent code... yes\nconfigure: creating ./config.status\nconfig.status: creating Makefile\nconfig.status: creating lib/Makefile\nconfig.status: creating src/Makefile\nconfig.status: creating ice/Makefile\nconfig.status: creating scripts/Makefile\nconfig.status: creating amasim/Makefile\nconfig.status: creating level2/Makefile\nconfig.status: creating config.h\nconfig.status: executing depfiles commands\nconfig.status: executing libtool commands\n-------------------------------------------------------------------\n Photonics: \"1.73: pyrosoma r4\"\n please refer to the 'INSTALL' file for further instructions.\n Hints:\n\n Building photonics......................'make'\n Compiling tool directory................'make tool'\n Performing post compile test............'make tests'\n All of the above........................'make everything'\n\n Clean objects and binaries..............'make clean'\n Clean tool directory....................'make toolclean'\n Remove traces of previous configure.....'make distclean'\n\n Also consider trying 'scripts/install_in_icetray.sh --help'\n\n\n---> Building photonics_1.73 with target all\nDEBUG: Executing com.apple.build (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all'\nmake all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\nError: Target com.apple.build returned: shell command \"cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all\" returned error 2\nCommand output: make all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\n\nWarning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install\nicecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS\n}}}",
"reporter": "boersma",
"cc": "",
"resolution": "worksforme",
"time": "2012-06-22T15:13:34",
"component": "booking",
"summary": "port photonics_1.73 does not build on SL5 64 bit",
"priority": "normal",
"keywords": "photonics I3_PORTS gcc",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
port photonics_1.73 does not build on SL5 64 bit (Trac #686) - Scientific Linux release 5.8 (Boron)
```text
$ gcc --version
gcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)
Copyright (C) 2006 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
```
```text
$ ./bin/port -vd install photonics_1.73
DEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73
DEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73
DEBUG: Requested variant x86_64 is not provided by port photonics_1.73.
DEBUG: Executing variant linux provides linux
DEBUG: Executing com.apple.main (photonics_1.73)
DEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz
DEBUG: Skipping unarchive (photonics_1.73) since no archive found
DEBUG: Skipping completed com.apple.unarchive (photonics_1.73)
---> Fetching photonics_1.73
DEBUG: Executing com.apple.fetch (photonics_1.73)
---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73
---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M
---> Verifying checksum(s) for photonics_1.73
DEBUG: Executing com.apple.checksum (photonics_1.73)
---> Checksumming photonics-1.73.tar.bz2
DEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2
DEBUG: setting option extract.cmd to /usr/bin/bzip2
---> Extracting photonics_1.73
DEBUG: Executing com.apple.extract (photonics_1.73)
---> Extracting photonics-1.73.tar.bz2
DEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'
DEBUG: Executing com.apple.patch (photonics_1.73)
---> Configuring photonics_1.73
DEBUG: Executing com.apple.configure (photonics_1.73)
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'
checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... yes
checking for a thread-safe mkdir -p... /bin/mkdir -p
checking for gawk... gawk
checking whether make sets $(MAKE)... yes
checking for gawk... (cached) gawk
checking for gcc... gcc
checking for C compiler default output file name... a.out
checking whether the C compiler works... yes
checking whether we are cross compiling... no
checking for suffix of executables...
checking for suffix of object files... o
checking whether we are using the GNU C compiler... yes
checking whether gcc accepts -g... yes
checking for gcc option to accept ISO C89... none needed
checking for style of include used by make... GNU
checking dependency style of gcc... gcc3
checking how to run the C preprocessor... gcc -E
checking for g++... g++
checking whether we are using the GNU C++ compiler... yes
checking whether g++ accepts -g... yes
checking dependency style of g++... gcc3
checking how to run the C++ preprocessor... g++ -E
checking for a BSD-compatible install... /usr/bin/install -c
checking whether ln -s works... yes
checking whether make sets $(MAKE)... (cached) yes
checking build system type... x86_64-unknown-linux-gnu
checking host system type... x86_64-unknown-linux-gnu
checking for a sed that does not truncate output... /bin/sed
checking for grep that handles long lines and -e... /bin/grep
checking for egrep... /bin/grep -E
checking for fgrep... /bin/grep -F
checking for ld used by gcc... /usr/bin/ld
checking if the linker (/usr/bin/ld) is GNU ld... yes
checking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B
checking the name lister (/usr/bin/nm -B) interface... BSD nm
checking the maximum length of command line arguments... 98304
checking whether the shell understands some XSI constructs... yes
checking whether the shell understands "+="... yes
checking for /usr/bin/ld option to reload object files... -r
checking how to recognize dependent libraries... pass_all
checking for ar... ar
checking for strip... strip
checking for ranlib... ranlib
checking command to parse /usr/bin/nm -B output from gcc object... ok
checking for ANSI C header files... yes
checking for sys/types.h... yes
checking for sys/stat.h... yes
checking for stdlib.h... yes
checking for string.h... yes
checking for memory.h... yes
checking for strings.h... yes
checking for inttypes.h... yes
checking for stdint.h... yes
checking for unistd.h... yes
checking for dlfcn.h... yes
checking whether we are using the GNU C++ compiler... (cached) yes
checking whether g++ accepts -g... (cached) yes
checking dependency style of g++... (cached) gcc3
checking how to run the C++ preprocessor... g++ -E
checking for objdir... .libs
checking if gcc supports -fno-rtti -fno-exceptions... no
checking for gcc option to produce PIC... -fPIC -DPIC
checking if gcc PIC flag -fPIC -DPIC works... yes
checking if gcc static flag -static works... yes
checking if gcc supports -c -o file.o... yes
checking if gcc supports -c -o file.o... (cached) yes
checking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking whether -lc should be explicitly linked in... no
checking dynamic linker characteristics... GNU/Linux ld.so
checking how to hardcode library paths into programs... immediate
checking whether stripping libraries is possible... yes
checking if libtool supports shared libraries... yes
checking whether to build shared libraries... yes
checking whether to build static libraries... yes
checking for ld used by g++... /usr/bin/ld -m elf_x86_64
checking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes
checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking for g++ option to produce PIC... -fPIC -DPIC
checking if g++ PIC flag -fPIC -DPIC works... yes
checking if g++ static flag -static works... yes
checking if g++ supports -c -o file.o... yes
checking if g++ supports -c -o file.o... (cached) yes
checking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes
checking dynamic linker characteristics... GNU/Linux ld.so
checking how to hardcode library paths into programs... immediate
checking for ANSI C header files... (cached) yes
checking for stdbool.h that conforms to C99... yes
checking for _Bool... yes
checking limits.h usability... yes
checking limits.h presence... yes
checking for limits.h... yes
checking malloc.h usability... yes
checking malloc.h presence... yes
checking for malloc.h... yes
checking for an ANSI C-conforming const... yes
checking for size_t... yes
checking for int8_t... yes
checking for int16_t... yes
checking for int32_t... yes
checking for int64_t... yes
checking for off_t... yes
checking for stdlib.h... (cached) yes
checking for GNU libc compatible malloc... yes
checking for stdlib.h... (cached) yes
checking for unistd.h... (cached) yes
checking for getpagesize... yes
checking for working mmap... yes
checking for stdlib.h... (cached) yes
checking for GNU libc compatible realloc... yes
checking for working strtod... yes
checking for strstr... yes
checking for strtod... (cached) yes
checking for strtol... yes
checking for strerror... yes
checking for memset... yes
checking for floor... no
checking for library containing floor... -lm
checking for pow... yes
checking for sqrt... yes
checking whether to enable debug mode... yes
disabled cernlib dependent code... yes
configure: creating ./config.status
config.status: creating Makefile
config.status: creating lib/Makefile
config.status: creating src/Makefile
config.status: creating ice/Makefile
config.status: creating scripts/Makefile
config.status: creating amasim/Makefile
config.status: creating level2/Makefile
config.status: creating config.h
config.status: executing depfiles commands
config.status: executing libtool commands
-------------------------------------------------------------------
Photonics: "1.73: pyrosoma r4"
please refer to the 'INSTALL' file for further instructions.
Hints:
Building photonics......................'make'
Compiling tool directory................'make tool'
Performing post compile test............'make tests'
All of the above........................'make everything'
Clean objects and binaries..............'make clean'
Clean tool directory....................'make toolclean'
Remove traces of previous configure.....'make distclean'
Also consider trying 'scripts/install_in_icetray.sh --help'
---> Building photonics_1.73 with target all
DEBUG: Executing com.apple.build (photonics_1.73)
DEBUG: Assembled command: 'cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all'
make all-recursive
make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
Making all in lib
make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c
libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o
boundary.c:1: error: bad value (native) for -march= switch
boundary.c:1: error: bad value (native) for -mtune= switch
make[2]: *** [boundary.lo] Error 1
make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
make: *** [all] Error 2
Error: Target com.apple.build returned: shell command "cd "/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73" && CFLAGS=-fPIC make all" returned error 2
Command output: make all-recursive
make[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
Making all in lib
make[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c
libtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o
boundary.c:1: error: bad value (native) for -march= switch
boundary.c:1: error: bad value (native) for -mtune= switch
make[2]: *** [boundary.lo] Error 1
make[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'
make: *** [all] Error 2
Warning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install
icecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/686">https://code.icecube.wisc.edu/projects/icecube/ticket/686</a>, reported by boersma</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-06-22T16:05:07",
"_ts": "1340381107000000",
"description": "Scientific Linux release 5.8 (Boron)\n\n{{{\n$ gcc --version\ngcc (GCC) 4.1.2 20080704 (Red Hat 4.1.2-52)\nCopyright (C) 2006 Free Software Foundation, Inc.\nThis is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n}}}\n\n{{{\n$ ./bin/port -vd install photonics_1.73\nDEBUG: Found port in file:///net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Changing to port directory: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/sources/rsync.code.icecube.wisc.edu_icecube-tools-ports/science/photonics_1.73\nDEBUG: Requested variant x86_64 is not provided by port photonics_1.73.\nDEBUG: Executing variant linux provides linux\nDEBUG: Executing com.apple.main (photonics_1.73)\nDEBUG: No TGZ archive: /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/packages/linux/x86_64/photonics_1.73-1.73_0.x86_64.tgz\nDEBUG: Skipping unarchive (photonics_1.73) since no archive found\nDEBUG: Skipping completed com.apple.unarchive (photonics_1.73)\n---> Fetching photonics_1.73\nDEBUG: Executing com.apple.fetch (photonics_1.73)\n---> photonics-1.73.tar.bz2 doesn't seem to exist in /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\n---> Attempting to fetch photonics-1.73.tar.bz2 from http://kent.dl.sourceforge.net/photonics\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73\" && curl -f -L -o photonics-1.73.tar.bz2.TMP http://kent.dl.sourceforge.net/photonics/photonics-1.73.tar.bz2'\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 485k 100 485k 0 0 299k 0 0:00:01 0:00:01 --:--:-- 10.4M\n---> Verifying checksum(s) for photonics_1.73\nDEBUG: Executing com.apple.checksum (photonics_1.73)\n---> Checksumming photonics-1.73.tar.bz2\nDEBUG: Correct (md5) checksum for photonics-1.73.tar.bz2\nDEBUG: setting option extract.cmd to /usr/bin/bzip2\n---> Extracting photonics_1.73\nDEBUG: Executing com.apple.extract (photonics_1.73)\n---> Extracting photonics-1.73.tar.bz2\nDEBUG: setting option extract.args to /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work\" && /usr/bin/bzip2 -dc /net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/distfiles/photonics_1.73/photonics-1.73.tar.bz2 | tar --no-same-owner -xf -'\nDEBUG: Executing com.apple.patch (photonics_1.73)\n---> Configuring photonics_1.73\nDEBUG: Executing com.apple.configure (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC ./configure --prefix=/net/software_icecube/SL5-py26/I3_PORTS --enable-optimize'\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether build environment is sane... yes\nchecking for a thread-safe mkdir -p... /bin/mkdir -p\nchecking for gawk... gawk\nchecking whether make sets $(MAKE)... yes\nchecking for gawk... (cached) gawk\nchecking for gcc... gcc\nchecking for C compiler default output file name... a.out\nchecking whether the C compiler works... yes\nchecking whether we are cross compiling... no\nchecking for suffix of executables... \nchecking for suffix of object files... o\nchecking whether we are using the GNU C compiler... yes\nchecking whether gcc accepts -g... yes\nchecking for gcc option to accept ISO C89... none needed\nchecking for style of include used by make... GNU\nchecking dependency style of gcc... gcc3\nchecking how to run the C preprocessor... gcc -E\nchecking for g++... g++\nchecking whether we are using the GNU C++ compiler... yes\nchecking whether g++ accepts -g... yes\nchecking dependency style of g++... gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for a BSD-compatible install... /usr/bin/install -c\nchecking whether ln -s works... yes\nchecking whether make sets $(MAKE)... (cached) yes\nchecking build system type... x86_64-unknown-linux-gnu\nchecking host system type... x86_64-unknown-linux-gnu\nchecking for a sed that does not truncate output... /bin/sed\nchecking for grep that handles long lines and -e... /bin/grep\nchecking for egrep... /bin/grep -E\nchecking for fgrep... /bin/grep -F\nchecking for ld used by gcc... /usr/bin/ld\nchecking if the linker (/usr/bin/ld) is GNU ld... yes\nchecking for BSD- or MS-compatible name lister (nm)... /usr/bin/nm -B\nchecking the name lister (/usr/bin/nm -B) interface... BSD nm\nchecking the maximum length of command line arguments... 98304\nchecking whether the shell understands some XSI constructs... yes\nchecking whether the shell understands \"+=\"... yes\nchecking for /usr/bin/ld option to reload object files... -r\nchecking how to recognize dependent libraries... pass_all\nchecking for ar... ar\nchecking for strip... strip\nchecking for ranlib... ranlib\nchecking command to parse /usr/bin/nm -B output from gcc object... ok\nchecking for ANSI C header files... yes\nchecking for sys/types.h... yes\nchecking for sys/stat.h... yes\nchecking for stdlib.h... yes\nchecking for string.h... yes\nchecking for memory.h... yes\nchecking for strings.h... yes\nchecking for inttypes.h... yes\nchecking for stdint.h... yes\nchecking for unistd.h... yes\nchecking for dlfcn.h... yes\nchecking whether we are using the GNU C++ compiler... (cached) yes\nchecking whether g++ accepts -g... (cached) yes\nchecking dependency style of g++... (cached) gcc3\nchecking how to run the C++ preprocessor... g++ -E\nchecking for objdir... .libs\nchecking if gcc supports -fno-rtti -fno-exceptions... no\nchecking for gcc option to produce PIC... -fPIC -DPIC\nchecking if gcc PIC flag -fPIC -DPIC works... yes\nchecking if gcc static flag -static works... yes\nchecking if gcc supports -c -o file.o... yes\nchecking if gcc supports -c -o file.o... (cached) yes\nchecking whether the gcc linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking whether -lc should be explicitly linked in... no\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking whether stripping libraries is possible... yes\nchecking if libtool supports shared libraries... yes\nchecking whether to build shared libraries... yes\nchecking whether to build static libraries... yes\nchecking for ld used by g++... /usr/bin/ld -m elf_x86_64\nchecking if the linker (/usr/bin/ld -m elf_x86_64) is GNU ld... yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking for g++ option to produce PIC... -fPIC -DPIC\nchecking if g++ PIC flag -fPIC -DPIC works... yes\nchecking if g++ static flag -static works... yes\nchecking if g++ supports -c -o file.o... yes\nchecking if g++ supports -c -o file.o... (cached) yes\nchecking whether the g++ linker (/usr/bin/ld -m elf_x86_64) supports shared libraries... yes\nchecking dynamic linker characteristics... GNU/Linux ld.so\nchecking how to hardcode library paths into programs... immediate\nchecking for ANSI C header files... (cached) yes\nchecking for stdbool.h that conforms to C99... yes\nchecking for _Bool... yes\nchecking limits.h usability... yes\nchecking limits.h presence... yes\nchecking for limits.h... yes\nchecking malloc.h usability... yes\nchecking malloc.h presence... yes\nchecking for malloc.h... yes\nchecking for an ANSI C-conforming const... yes\nchecking for size_t... yes\nchecking for int8_t... yes\nchecking for int16_t... yes\nchecking for int32_t... yes\nchecking for int64_t... yes\nchecking for off_t... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible malloc... yes\nchecking for stdlib.h... (cached) yes\nchecking for unistd.h... (cached) yes\nchecking for getpagesize... yes\nchecking for working mmap... yes\nchecking for stdlib.h... (cached) yes\nchecking for GNU libc compatible realloc... yes\nchecking for working strtod... yes\nchecking for strstr... yes\nchecking for strtod... (cached) yes\nchecking for strtol... yes\nchecking for strerror... yes\nchecking for memset... yes\nchecking for floor... no\nchecking for library containing floor... -lm\nchecking for pow... yes\nchecking for sqrt... yes\nchecking whether to enable debug mode... yes\ndisabled cernlib dependent code... yes\nconfigure: creating ./config.status\nconfig.status: creating Makefile\nconfig.status: creating lib/Makefile\nconfig.status: creating src/Makefile\nconfig.status: creating ice/Makefile\nconfig.status: creating scripts/Makefile\nconfig.status: creating amasim/Makefile\nconfig.status: creating level2/Makefile\nconfig.status: creating config.h\nconfig.status: executing depfiles commands\nconfig.status: executing libtool commands\n-------------------------------------------------------------------\n Photonics: \"1.73: pyrosoma r4\"\n please refer to the 'INSTALL' file for further instructions.\n Hints:\n\n Building photonics......................'make'\n Compiling tool directory................'make tool'\n Performing post compile test............'make tests'\n All of the above........................'make everything'\n\n Clean objects and binaries..............'make clean'\n Clean tool directory....................'make toolclean'\n Remove traces of previous configure.....'make distclean'\n\n Also consider trying 'scripts/install_in_icetray.sh --help'\n\n\n---> Building photonics_1.73 with target all\nDEBUG: Executing com.apple.build (photonics_1.73)\nDEBUG: Assembled command: 'cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all'\nmake all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\nError: Target com.apple.build returned: shell command \"cd \"/net/software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73\" && CFLAGS=-fPIC make all\" returned error 2\nCommand output: make all-recursive\nmake[1]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nMaking all in lib\nmake[2]: Entering directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\n/bin/sh ../libtool --tag=CC --mode=compile gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c -o boundary.lo boundary.c\nlibtool: compile: gcc -DHAVE_CONFIG_H -I. -I.. -O3 -funroll-loops -fmerge-all-constants -march=native -mtune=native -g -O2 -Wall -fno-inline -fPIC -MT boundary.lo -MD -MP -MF .deps/boundary.Tpo -c boundary.c -fPIC -DPIC -o .libs/boundary.o\nboundary.c:1: error: bad value (native) for -march= switch\nboundary.c:1: error: bad value (native) for -mtune= switch\nmake[2]: *** [boundary.lo] Error 1\nmake[2]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73/lib'\nmake[1]: *** [all-recursive] Error 1\nmake[1]: Leaving directory `/.automount/net_ro/net__software_icecube/SL5-py26/I3_PORTS/var/db/dports/build/file._net_software_icecube_SL5-py26_I3_PORTS_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_photonics_1.73/work/photonics-1.73'\nmake: *** [all] Error 2\n\nWarning: the following items did not execute (for photonics_1.73): com.apple.activate com.apple.build com.apple.destroot com.apple.archive com.apple.install\nicecubemgr@lx3b48:/net/software_icecube/SL5-py26/I3_PORTS\n}}}",
"reporter": "boersma",
"cc": "",
"resolution": "worksforme",
"time": "2012-06-22T15:13:34",
"component": "booking",
"summary": "port photonics_1.73 does not build on SL5 64 bit",
"priority": "normal",
"keywords": "photonics I3_PORTS gcc",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
defect
|
port photonics does not build on bit trac scientific linux release boron text gcc version gcc gcc red hat copyright c free software foundation inc this is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose text bin port vd install photonics debug found port in file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics debug changing to port directory net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics debug requested variant is not provided by port photonics debug executing variant linux provides linux debug executing com apple main photonics debug no tgz archive net software icecube ports var db dports packages linux photonics tgz debug skipping unarchive photonics since no archive found debug skipping completed com apple unarchive photonics fetching photonics debug executing com apple fetch photonics photonics tar doesn t seem to exist in net software icecube ports var db dports distfiles photonics attempting to fetch photonics tar from debug assembled command cd net software icecube ports var db dports distfiles photonics curl f l o photonics tar tmp total received xferd average speed time time time current dload upload total spent left speed verifying checksum s for photonics debug executing com apple checksum photonics checksumming photonics tar debug correct checksum for photonics tar debug setting option extract cmd to usr bin extracting photonics debug executing com apple extract photonics extracting photonics tar debug setting option extract args to net software icecube ports var db dports distfiles photonics photonics tar debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work usr bin dc net software icecube ports var db dports distfiles photonics photonics tar tar no same owner xf debug executing com apple patch photonics configuring photonics debug executing com apple configure photonics debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic configure prefix net software icecube ports enable optimize checking for a bsd compatible install usr bin install c checking whether build environment is sane yes checking for a thread safe mkdir p bin mkdir p checking for gawk gawk checking whether make sets make yes checking for gawk cached gawk checking for gcc gcc checking for c compiler default output file name a out checking whether the c compiler works yes checking whether we are cross compiling no checking for suffix of executables checking for suffix of object files o checking whether we are using the gnu c compiler yes checking whether gcc accepts g yes checking for gcc option to accept iso none needed checking for style of include used by make gnu checking dependency style of gcc checking how to run the c preprocessor gcc e checking for g g checking whether we are using the gnu c compiler yes checking whether g accepts g yes checking dependency style of g checking how to run the c preprocessor g e checking for a bsd compatible install usr bin install c checking whether ln s works yes checking whether make sets make cached yes checking build system type unknown linux gnu checking host system type unknown linux gnu checking for a sed that does not truncate output bin sed checking for grep that handles long lines and e bin grep checking for egrep bin grep e checking for fgrep bin grep f checking for ld used by gcc usr bin ld checking if the linker usr bin ld is gnu ld yes checking for bsd or ms compatible name lister nm usr bin nm b checking the name lister usr bin nm b interface bsd nm checking the maximum length of command line arguments checking whether the shell understands some xsi constructs yes checking whether the shell understands yes checking for usr bin ld option to reload object files r checking how to recognize dependent libraries pass all checking for ar ar checking for strip strip checking for ranlib ranlib checking command to parse usr bin nm b output from gcc object ok checking for ansi c header files yes checking for sys types h yes checking for sys stat h yes checking for stdlib h yes checking for string h yes checking for memory h yes checking for strings h yes checking for inttypes h yes checking for stdint h yes checking for unistd h yes checking for dlfcn h yes checking whether we are using the gnu c compiler cached yes checking whether g accepts g cached yes checking dependency style of g cached checking how to run the c preprocessor g e checking for objdir libs checking if gcc supports fno rtti fno exceptions no checking for gcc option to produce pic fpic dpic checking if gcc pic flag fpic dpic works yes checking if gcc static flag static works yes checking if gcc supports c o file o yes checking if gcc supports c o file o cached yes checking whether the gcc linker usr bin ld m elf supports shared libraries yes checking whether lc should be explicitly linked in no checking dynamic linker characteristics gnu linux ld so checking how to hardcode library paths into programs immediate checking whether stripping libraries is possible yes checking if libtool supports shared libraries yes checking whether to build shared libraries yes checking whether to build static libraries yes checking for ld used by g usr bin ld m elf checking if the linker usr bin ld m elf is gnu ld yes checking whether the g linker usr bin ld m elf supports shared libraries yes checking for g option to produce pic fpic dpic checking if g pic flag fpic dpic works yes checking if g static flag static works yes checking if g supports c o file o yes checking if g supports c o file o cached yes checking whether the g linker usr bin ld m elf supports shared libraries yes checking dynamic linker characteristics gnu linux ld so checking how to hardcode library paths into programs immediate checking for ansi c header files cached yes checking for stdbool h that conforms to yes checking for bool yes checking limits h usability yes checking limits h presence yes checking for limits h yes checking malloc h usability yes checking malloc h presence yes checking for malloc h yes checking for an ansi c conforming const yes checking for size t yes checking for t yes checking for t yes checking for t yes checking for t yes checking for off t yes checking for stdlib h cached yes checking for gnu libc compatible malloc yes checking for stdlib h cached yes checking for unistd h cached yes checking for getpagesize yes checking for working mmap yes checking for stdlib h cached yes checking for gnu libc compatible realloc yes checking for working strtod yes checking for strstr yes checking for strtod cached yes checking for strtol yes checking for strerror yes checking for memset yes checking for floor no checking for library containing floor lm checking for pow yes checking for sqrt yes checking whether to enable debug mode yes disabled cernlib dependent code yes configure creating config status config status creating makefile config status creating lib makefile config status creating src makefile config status creating ice makefile config status creating scripts makefile config status creating amasim makefile config status creating makefile config status creating config h config status executing depfiles commands config status executing libtool commands photonics pyrosoma please refer to the install file for further instructions hints building photonics make compiling tool directory make tool performing post compile test make tests all of the above make everything clean objects and binaries make clean clean tool directory make toolclean remove traces of previous configure make distclean also consider trying scripts install in icetray sh help building photonics with target all debug executing com apple build photonics debug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all make all recursive make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics making all in lib make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c libtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o boundary c error bad value native for march switch boundary c error bad value native for mtune switch make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics make error error target com apple build returned shell command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all returned error command output make all recursive make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics making all in lib make entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c libtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o boundary c error bad value native for march switch boundary c error bad value native for mtune switch make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib make error make leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics make error warning the following items did not execute for photonics com apple activate com apple build com apple destroot com apple archive com apple install icecubemgr net software icecube ports migrated from json status closed changetime ts description scientific linux release boron n n n gcc version ngcc gcc red hat ncopyright c free software foundation inc nthis is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose n n n n bin port vd install photonics ndebug found port in file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics ndebug changing to port directory net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics ndebug requested variant is not provided by port photonics ndebug executing variant linux provides linux ndebug executing com apple main photonics ndebug no tgz archive net software icecube ports var db dports packages linux photonics tgz ndebug skipping unarchive photonics since no archive found ndebug skipping completed com apple unarchive photonics n fetching photonics ndebug executing com apple fetch photonics n photonics tar doesn t seem to exist in net software icecube ports var db dports distfiles photonics n attempting to fetch photonics tar from assembled command cd net software icecube ports var db dports distfiles photonics curl f l o photonics tar tmp total received xferd average speed time time time current n dload upload total spent left speed n verifying checksum s for photonics ndebug executing com apple checksum photonics n checksumming photonics tar ndebug correct checksum for photonics tar ndebug setting option extract cmd to usr bin n extracting photonics ndebug executing com apple extract photonics n extracting photonics tar ndebug setting option extract args to net software icecube ports var db dports distfiles photonics photonics tar ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work usr bin dc net software icecube ports var db dports distfiles photonics photonics tar tar no same owner xf ndebug executing com apple patch photonics n configuring photonics ndebug executing com apple configure photonics ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic configure prefix net software icecube ports enable optimize nchecking for a bsd compatible install usr bin install c nchecking whether build environment is sane yes nchecking for a thread safe mkdir p bin mkdir p nchecking for gawk gawk nchecking whether make sets make yes nchecking for gawk cached gawk nchecking for gcc gcc nchecking for c compiler default output file name a out nchecking whether the c compiler works yes nchecking whether we are cross compiling no nchecking for suffix of executables nchecking for suffix of object files o nchecking whether we are using the gnu c compiler yes nchecking whether gcc accepts g yes nchecking for gcc option to accept iso none needed nchecking for style of include used by make gnu nchecking dependency style of gcc nchecking how to run the c preprocessor gcc e nchecking for g g nchecking whether we are using the gnu c compiler yes nchecking whether g accepts g yes nchecking dependency style of g nchecking how to run the c preprocessor g e nchecking for a bsd compatible install usr bin install c nchecking whether ln s works yes nchecking whether make sets make cached yes nchecking build system type unknown linux gnu nchecking host system type unknown linux gnu nchecking for a sed that does not truncate output bin sed nchecking for grep that handles long lines and e bin grep nchecking for egrep bin grep e nchecking for fgrep bin grep f nchecking for ld used by gcc usr bin ld nchecking if the linker usr bin ld is gnu ld yes nchecking for bsd or ms compatible name lister nm usr bin nm b nchecking the name lister usr bin nm b interface bsd nm nchecking the maximum length of command line arguments nchecking whether the shell understands some xsi constructs yes nchecking whether the shell understands yes nchecking for usr bin ld option to reload object files r nchecking how to recognize dependent libraries pass all nchecking for ar ar nchecking for strip strip nchecking for ranlib ranlib nchecking command to parse usr bin nm b output from gcc object ok nchecking for ansi c header files yes nchecking for sys types h yes nchecking for sys stat h yes nchecking for stdlib h yes nchecking for string h yes nchecking for memory h yes nchecking for strings h yes nchecking for inttypes h yes nchecking for stdint h yes nchecking for unistd h yes nchecking for dlfcn h yes nchecking whether we are using the gnu c compiler cached yes nchecking whether g accepts g cached yes nchecking dependency style of g cached nchecking how to run the c preprocessor g e nchecking for objdir libs nchecking if gcc supports fno rtti fno exceptions no nchecking for gcc option to produce pic fpic dpic nchecking if gcc pic flag fpic dpic works yes nchecking if gcc static flag static works yes nchecking if gcc supports c o file o yes nchecking if gcc supports c o file o cached yes nchecking whether the gcc linker usr bin ld m elf supports shared libraries yes nchecking whether lc should be explicitly linked in no nchecking dynamic linker characteristics gnu linux ld so nchecking how to hardcode library paths into programs immediate nchecking whether stripping libraries is possible yes nchecking if libtool supports shared libraries yes nchecking whether to build shared libraries yes nchecking whether to build static libraries yes nchecking for ld used by g usr bin ld m elf nchecking if the linker usr bin ld m elf is gnu ld yes nchecking whether the g linker usr bin ld m elf supports shared libraries yes nchecking for g option to produce pic fpic dpic nchecking if g pic flag fpic dpic works yes nchecking if g static flag static works yes nchecking if g supports c o file o yes nchecking if g supports c o file o cached yes nchecking whether the g linker usr bin ld m elf supports shared libraries yes nchecking dynamic linker characteristics gnu linux ld so nchecking how to hardcode library paths into programs immediate nchecking for ansi c header files cached yes nchecking for stdbool h that conforms to yes nchecking for bool yes nchecking limits h usability yes nchecking limits h presence yes nchecking for limits h yes nchecking malloc h usability yes nchecking malloc h presence yes nchecking for malloc h yes nchecking for an ansi c conforming const yes nchecking for size t yes nchecking for t yes nchecking for t yes nchecking for t yes nchecking for t yes nchecking for off t yes nchecking for stdlib h cached yes nchecking for gnu libc compatible malloc yes nchecking for stdlib h cached yes nchecking for unistd h cached yes nchecking for getpagesize yes nchecking for working mmap yes nchecking for stdlib h cached yes nchecking for gnu libc compatible realloc yes nchecking for working strtod yes nchecking for strstr yes nchecking for strtod cached yes nchecking for strtol yes nchecking for strerror yes nchecking for memset yes nchecking for floor no nchecking for library containing floor lm nchecking for pow yes nchecking for sqrt yes nchecking whether to enable debug mode yes ndisabled cernlib dependent code yes nconfigure creating config status nconfig status creating makefile nconfig status creating lib makefile nconfig status creating src makefile nconfig status creating ice makefile nconfig status creating scripts makefile nconfig status creating amasim makefile nconfig status creating makefile nconfig status creating config h nconfig status executing depfiles commands nconfig status executing libtool commands n n photonics pyrosoma n please refer to the install file for further instructions n hints n n building photonics make n compiling tool directory make tool n performing post compile test make tests n all of the above make everything n n clean objects and binaries make clean n clean tool directory make toolclean n remove traces of previous configure make distclean n n also consider trying scripts install in icetray sh help n n n building photonics with target all ndebug executing com apple build photonics ndebug assembled command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all nmake all recursive nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmaking all in lib nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib n bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c nlibtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o nboundary c error bad value native for march switch nboundary c error bad value native for mtune switch nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmake error nerror target com apple build returned shell command cd net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics cflags fpic make all returned error ncommand output make all recursive nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmaking all in lib nmake entering directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib n bin sh libtool tag cc mode compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c o boundary lo boundary c nlibtool compile gcc dhave config h i i funroll loops fmerge all constants march native mtune native g wall fno inline fpic mt boundary lo md mp mf deps boundary tpo c boundary c fpic dpic o libs boundary o nboundary c error bad value native for march switch nboundary c error bad value native for mtune switch nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics lib nmake error nmake leaving directory automount net ro net software icecube ports var db dports build file net software icecube ports var db dports sources rsync code icecube wisc edu icecube tools ports science photonics work photonics nmake error n nwarning the following items did not execute for photonics com apple activate com apple build com apple destroot com apple archive com apple install nicecubemgr net software icecube ports n reporter boersma cc resolution worksforme time component booking summary port photonics does not build on bit priority normal keywords photonics ports gcc milestone owner type defect
| 1
|
2,958
| 2,607,967,648
|
IssuesEvent
|
2015-02-26 00:43:06
|
chrsmithdemos/leveldb
|
https://api.github.com/repos/chrsmithdemos/leveldb
|
closed
|
build_detect_platform line ending error
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. git clone https://code.google.com/p/leveldb/
2. cd leveldb
3. make
What is the expected output? What do you see instead?
It should compile. I get an error:
$ make
/bin/sh: 1: ./build_detect_platform: not found
Makefile:18: build_config.mk: No such file or directory
make: *** No rule to make target `build_config.mk'. Stop.
./build_detect_platform actually exists, but when executed it prints:
$ ./build_detect_platform
bash: ./build_detect_platform: /bin/sh^M: bad interpreter: No such file or
directory
What version of the product are you using? On what operating system?
leveldb 1.9, Ubuntu 12.04.
Please provide any additional information below.
'dos2unix build_detect_platform' fixes the problem
```
-----
Original issue reported on code.google.com by `gmarg...@gmail.com` on 22 Jan 2013 at 1:41
|
1.0
|
build_detect_platform line ending error - ```
What steps will reproduce the problem?
1. git clone https://code.google.com/p/leveldb/
2. cd leveldb
3. make
What is the expected output? What do you see instead?
It should compile. I get an error:
$ make
/bin/sh: 1: ./build_detect_platform: not found
Makefile:18: build_config.mk: No such file or directory
make: *** No rule to make target `build_config.mk'. Stop.
./build_detect_platform actually exists, but when executed it prints:
$ ./build_detect_platform
bash: ./build_detect_platform: /bin/sh^M: bad interpreter: No such file or
directory
What version of the product are you using? On what operating system?
leveldb 1.9, Ubuntu 12.04.
Please provide any additional information below.
'dos2unix build_detect_platform' fixes the problem
```
-----
Original issue reported on code.google.com by `gmarg...@gmail.com` on 22 Jan 2013 at 1:41
|
defect
|
build detect platform line ending error what steps will reproduce the problem git clone cd leveldb make what is the expected output what do you see instead it should compile i get an error make bin sh build detect platform not found makefile build config mk no such file or directory make no rule to make target build config mk stop build detect platform actually exists but when executed it prints build detect platform bash build detect platform bin sh m bad interpreter no such file or directory what version of the product are you using on what operating system leveldb ubuntu please provide any additional information below build detect platform fixes the problem original issue reported on code google com by gmarg gmail com on jan at
| 1
|
81,556
| 31,018,920,406
|
IssuesEvent
|
2023-08-10 02:29:59
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
Data corruption since generic_file_splice_read -> filemap_splice_read change (6.5 compat, but occurs on 6.4 too)
|
Type: Defect
|
### System information
Type | Version/Name
--- | ---
Distribution Name | Arch
Distribution Version | Rolling release
Kernel Version | 6.4.8, 6.5rc1/2/3/4
Architecture | x86-64
OpenZFS Version | [commit 36261c8](https://github.com/openzfs/zfs/commit/36261c8238df462b214854ccea1df4f060cf0995)
### Describe the problem you're observing
After the recent changes to get OpenZFS compiling/running on 6.5, there appears to be a possible lingering data corruption bug. In the repeatable example below, it reliably inserts a long run of NULL bytes into a file, causing a build to fail (conveniently, the build of ZFS).
My expectation is that the bug probably exists for any kernel where `filemap_splice_read` exists, which recently has replaced `generic_file_splice_read` in other Linux filesystem code.
### Describe how to reproduce the problem
**Again - despite demonstrating the problem with the OpenZFS build, the problem only manifests itself when running on the ZFS branch at the commit listed above. It just so happens that I'm able to use our build to reproduce the bug.**
1. You need to be running ZFS patched up to the commit listed above. I have reproduced this on Kernel 6.4.8 and all 6.5 RC's up to rc4
2. `git clone https://github.com/openzfs/zfs.git`
3. `cd ./zfs`
4. `./autogen.sh`
5. `mkdir -p ../zfs-test`
6. `cd ../zfs-test`
7. `../zfs/configure --with-linux=/usr/src/linux` (or wherever your headers/source tree is)
Eventually, the `configure` will fail with the following message:
```
configure: error:
*** This kernel does not include the required loadable module
*** support!
***
*** To build OpenZFS as a loadable Linux kernel module
*** enable loadable module support by setting
*** `CONFIG_MODULES=y` in the kernel configuration and run
*** `make modules_prepare` in the Linux source tree.
***
*** If you don't intend to enable loadable kernel module
*** support, please compile OpenZFS as a Linux kernel built-in.
***
*** Prepare the Linux source tree by running `make prepare`,
*** use the OpenZFS `--enable-linux-builtin` configure option,
*** copy the OpenZFS sources into the Linux source tree using
*** `./copy-builtin <linux source directory>`,
*** set `CONFIG_ZFS=y` in the kernel configuration and compile
*** kernel as usual.
```
I enter the directory of the failing test:
```
cd build/config_modules
```
Looking at the `config_modules.c` file, which is resulting in the failure:
```c
/* confdefs.h */
#define PACKAGE_NAME "zfs"
#define PACKAGE_TARNAME "zfs"
#define PACKAGE_VERSION "2.2.99"
#define PACKAGE_STRING "zfs 2.2.99"
#define PACKAGE_BUGREPORT ""
#define PACKAGE_URL ""
#define ZFS_META_NAME "zfs"
#define ZFS_META_VERSION "2.2.99"
#define SPL_META_VERSION ZFS_META_VERSION
#define ZFS_META_RELEASE "1"
#define SPL_META_RELEASE ZFS_META_RELEASE
#define ZFS_META_LICENSE "CDDL"
#define ZFS_META_ALIAS "zfs-2.2.99-1"
#define SPL_META_ALIAS ZFS_META_ALIAS
#define ZFS_META_AUTHOR "OpenZFS"
#define ZFS_META_KVER_MIN "3.10"
#define ZFS_META_KVER_MAX "6.4"
#define PACKAGE "zfs"
#define VERSION "2.2.99"
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
#include <linux/module.h>
#if !defined(CONFIG_MODULES)
#error CONFIG_MODULES not defined
#endif
int
main (void)
{
;
return 0;
}
```
Curiously, this bug does not manifest frequently and the system largely appears to run stable in many other use cases.
The section of `configure` that writes that file looks like below. Including this snippet here as it might help illuminate what conditions need to be true to trigger the bug:
```bash
cat confdefs.h - <<_ACEOF >build/config_modules/config_modules.c
#include <linux/module.h>
#if !defined(CONFIG_MODULES)
#error CONFIG_MODULES not defined
#endif
int
main (void)
{
;
return 0;
}
MODULE_DESCRIPTION("conftest");
MODULE_AUTHOR(ZFS_META_AUTHOR);
MODULE_VERSION(ZFS_META_VERSION "-" ZFS_META_RELEASE);
MODULE_LICENSE("Dual BSD/GPL");
_ACEOF
```
### Include any warning/errors/backtraces from the system logs
There are no errors reported to the console or in kernel messages
|
1.0
|
Data corruption since generic_file_splice_read -> filemap_splice_read change (6.5 compat, but occurs on 6.4 too) - ### System information
Type | Version/Name
--- | ---
Distribution Name | Arch
Distribution Version | Rolling release
Kernel Version | 6.4.8, 6.5rc1/2/3/4
Architecture | x86-64
OpenZFS Version | [commit 36261c8](https://github.com/openzfs/zfs/commit/36261c8238df462b214854ccea1df4f060cf0995)
### Describe the problem you're observing
After the recent changes to get OpenZFS compiling/running on 6.5, there appears to be a possible lingering data corruption bug. In the repeatable example below, it reliably inserts a long run of NULL bytes into a file, causing a build to fail (conveniently, the build of ZFS).
My expectation is that the bug probably exists for any kernel where `filemap_splice_read` exists, which recently has replaced `generic_file_splice_read` in other Linux filesystem code.
### Describe how to reproduce the problem
**Again - despite demonstrating the problem with the OpenZFS build, the problem only manifests itself when running on the ZFS branch at the commit listed above. It just so happens that I'm able to use our build to reproduce the bug.**
1. You need to be running ZFS patched up to the commit listed above. I have reproduced this on Kernel 6.4.8 and all 6.5 RC's up to rc4
2. `git clone https://github.com/openzfs/zfs.git`
3. `cd ./zfs`
4. `./autogen.sh`
5. `mkdir -p ../zfs-test`
6. `cd ../zfs-test`
7. `../zfs/configure --with-linux=/usr/src/linux` (or wherever your headers/source tree is)
Eventually, the `configure` will fail with the following message:
```
configure: error:
*** This kernel does not include the required loadable module
*** support!
***
*** To build OpenZFS as a loadable Linux kernel module
*** enable loadable module support by setting
*** `CONFIG_MODULES=y` in the kernel configuration and run
*** `make modules_prepare` in the Linux source tree.
***
*** If you don't intend to enable loadable kernel module
*** support, please compile OpenZFS as a Linux kernel built-in.
***
*** Prepare the Linux source tree by running `make prepare`,
*** use the OpenZFS `--enable-linux-builtin` configure option,
*** copy the OpenZFS sources into the Linux source tree using
*** `./copy-builtin <linux source directory>`,
*** set `CONFIG_ZFS=y` in the kernel configuration and compile
*** kernel as usual.
```
I enter the directory of the failing test:
```
cd build/config_modules
```
Looking at the `config_modules.c` file, which is resulting in the failure:
```c
/* confdefs.h */
#define PACKAGE_NAME "zfs"
#define PACKAGE_TARNAME "zfs"
#define PACKAGE_VERSION "2.2.99"
#define PACKAGE_STRING "zfs 2.2.99"
#define PACKAGE_BUGREPORT ""
#define PACKAGE_URL ""
#define ZFS_META_NAME "zfs"
#define ZFS_META_VERSION "2.2.99"
#define SPL_META_VERSION ZFS_META_VERSION
#define ZFS_META_RELEASE "1"
#define SPL_META_RELEASE ZFS_META_RELEASE
#define ZFS_META_LICENSE "CDDL"
#define ZFS_META_ALIAS "zfs-2.2.99-1"
#define SPL_META_ALIAS ZFS_META_ALIAS
#define ZFS_META_AUTHOR "OpenZFS"
#define ZFS_META_KVER_MIN "3.10"
#define ZFS_META_KVER_MAX "6.4"
#define PACKAGE "zfs"
#define VERSION "2.2.99"
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
#include <linux/module.h>
#if !defined(CONFIG_MODULES)
#error CONFIG_MODULES not defined
#endif
int
main (void)
{
;
return 0;
}
```
Curiously, this bug does not manifest frequently and the system largely appears to run stable in many other use cases.
The section of `configure` that writes that file looks like below. Including this snippet here as it might help illuminate what conditions need to be true to trigger the bug:
```bash
cat confdefs.h - <<_ACEOF >build/config_modules/config_modules.c
#include <linux/module.h>
#if !defined(CONFIG_MODULES)
#error CONFIG_MODULES not defined
#endif
int
main (void)
{
;
return 0;
}
MODULE_DESCRIPTION("conftest");
MODULE_AUTHOR(ZFS_META_AUTHOR);
MODULE_VERSION(ZFS_META_VERSION "-" ZFS_META_RELEASE);
MODULE_LICENSE("Dual BSD/GPL");
_ACEOF
```
### Include any warning/errors/backtraces from the system logs
There are no errors reported to the console or in kernel messages
|
defect
|
data corruption since generic file splice read filemap splice read change compat but occurs on too system information type version name distribution name arch distribution version rolling release kernel version architecture openzfs version describe the problem you re observing after the recent changes to get openzfs compiling running on there appears to be a possible lingering data corruption bug in the repeatable example below it reliably inserts a long run of null bytes into a file causing a build to fail conveniently the build of zfs my expectation is that the bug probably exists for any kernel where filemap splice read exists which recently has replaced generic file splice read in other linux filesystem code describe how to reproduce the problem again despite demonstrating the problem with the openzfs build the problem only manifests itself when running on the zfs branch at the commit listed above it just so happens that i m able to use our build to reproduce the bug you need to be running zfs patched up to the commit listed above i have reproduced this on kernel and all rc s up to git clone cd zfs autogen sh mkdir p zfs test cd zfs test zfs configure with linux usr src linux or wherever your headers source tree is eventually the configure will fail with the following message configure error this kernel does not include the required loadable module support to build openzfs as a loadable linux kernel module enable loadable module support by setting config modules y in the kernel configuration and run make modules prepare in the linux source tree if you don t intend to enable loadable kernel module support please compile openzfs as a linux kernel built in prepare the linux source tree by running make prepare use the openzfs enable linux builtin configure option copy the openzfs sources into the linux source tree using copy builtin set config zfs y in the kernel configuration and compile kernel as usual i enter the directory of the failing test cd build config modules looking at the config modules c file which is resulting in the failure c confdefs h define package name zfs define package tarname zfs define package version define package string zfs define package bugreport define package url define zfs meta name zfs define zfs meta version define spl meta version zfs meta version define zfs meta release define spl meta release zfs meta release define zfs meta license cddl define zfs meta alias zfs define spl meta alias zfs meta alias define zfs meta author openzfs define zfs meta kver min define zfs meta kver max define package zfs define version include if defined config modules error config modules not defined endif int main void return curiously this bug does not manifest frequently and the system largely appears to run stable in many other use cases the section of configure that writes that file looks like below including this snippet here as it might help illuminate what conditions need to be true to trigger the bug bash cat confdefs h build config modules config modules c include if defined config modules error config modules not defined endif int main void return module description conftest module author zfs meta author module version zfs meta version zfs meta release module license dual bsd gpl aceof include any warning errors backtraces from the system logs there are no errors reported to the console or in kernel messages
| 1
|
2,658
| 4,877,466,440
|
IssuesEvent
|
2016-11-16 15:46:15
|
CartoDB/cartodb
|
https://api.github.com/repos/CartoDB/cartodb
|
closed
|
Error in dropping old table in a sync process leaves stale tables
|
Data-services
|
Related: https://github.com/CartoDB/cartodb/issues/6640
(Copied from the issue above):
This [piece of code](https://github.com/CartoDB/cartodb/blob/9b496fee7beac79c8d5f47570da8e505902a5b64/app/models/synchronization/adapter.rb#L143-L147) is silently failing, so what happens here is:
- A table gets downloaded and imported from the source
- The old table is renamed to be replaced by the new one (inside a transaction)
- The old table is dropped: in this case, this drop is failing silently
```
materialized view prueba_3_3_mymatview depends on table importer_591b0768054811e680c20e98b61680bf_to_be_deleted
Re HINT: Use DROP ... CASCADE to drop the dependent objects too.
```
- The exception is being rescued but ignored, so the table never gets dropped. As it occurs inside a transaction, the table doesn't get renamed either. The transaction is aborted.
- The sync is marked as success because no exception happened (it was silenced) but the data is not updated
This scenario is leaving stale tables in the database and also considering the process a "synchronization success".
|
1.0
|
Error in dropping old table in a sync process leaves stale tables - Related: https://github.com/CartoDB/cartodb/issues/6640
(Copied from the issue above):
This [piece of code](https://github.com/CartoDB/cartodb/blob/9b496fee7beac79c8d5f47570da8e505902a5b64/app/models/synchronization/adapter.rb#L143-L147) is silently failing, so what happens here is:
- A table gets downloaded and imported from the source
- The old table is renamed to be replaced by the new one (inside a transaction)
- The old table is dropped: in this case, this drop is failing silently
```
materialized view prueba_3_3_mymatview depends on table importer_591b0768054811e680c20e98b61680bf_to_be_deleted
Re HINT: Use DROP ... CASCADE to drop the dependent objects too.
```
- The exception is being rescued but ignored, so the table never gets dropped. As it occurs inside a transaction, the table doesn't get renamed either. The transaction is aborted.
- The sync is marked as success because no exception happened (it was silenced) but the data is not updated
This scenario is leaving stale tables in the database and also considering the process a "synchronization success".
|
non_defect
|
error in dropping old table in a sync process leaves stale tables related copied from the issue above this is silently failing so what happens here is a table gets downloaded and imported from the source the old table is renamed to be replaced by the new one inside a transaction the old table is dropped in this case this drop is failing silently materialized view prueba mymatview depends on table importer to be deleted re hint use drop cascade to drop the dependent objects too the exception is being rescued but ignored so the table never gets dropped as it occurs inside a transaction the table doesn t get renamed either the transaction is aborted the sync is marked as success because no exception happened it was silenced but the data is not updated this scenario is leaving stale tables in the database and also considering the process a synchronization success
| 0
|
21,925
| 3,587,215,053
|
IssuesEvent
|
2016-01-30 05:06:04
|
mash99/crypto-js
|
https://api.github.com/repos/mash99/crypto-js
|
closed
|
Error: Unable to get property 'createEncryptor' of undefined or null reference
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1.Just run the html file
2.I am trying to encypt using tripleDes
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `anithala...@gmail.com` on 22 Nov 2014 at 7:17
Attachments:
* [partner new.html](https://storage.googleapis.com/google-code-attachments/crypto-js/issue-144/comment-0/partner new.html)
|
1.0
|
Error: Unable to get property 'createEncryptor' of undefined or null reference - ```
What steps will reproduce the problem?
1.Just run the html file
2.I am trying to encypt using tripleDes
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `anithala...@gmail.com` on 22 Nov 2014 at 7:17
Attachments:
* [partner new.html](https://storage.googleapis.com/google-code-attachments/crypto-js/issue-144/comment-0/partner new.html)
|
defect
|
error unable to get property createencryptor of undefined or null reference what steps will reproduce the problem just run the html file i am trying to encypt using tripledes what is the expected output what do you see instead what version of the product are you using on what operating system please provide any additional information below original issue reported on code google com by anithala gmail com on nov at attachments new html
| 1
|
147,984
| 5,656,821,671
|
IssuesEvent
|
2017-04-10 03:50:26
|
elementary/switchboard-plug-pantheon-shell
|
https://api.github.com/repos/elementary/switchboard-plug-pantheon-shell
|
closed
|
wallpaper plug should have basic browsing options
|
Priority: Wishlist
|
"wallpaper folder" and "custom folder" are fine now, but "picture folders" could be more detailed. I think the user should be able to dig in successive folders within the picture folder and go back the "picture folder root". I believe most people use nested folders to sort their pictures and this renders the selection "picture folder" almost useless compared to "custom folder", even though they would essentially use the same folder.
Only in "picture folders": the folders should be shown, a folder ".." should also be here to go back to the "picture folder root".
Launchpad Details: [#LP1114258](https://bugs.launchpad.net/bugs/1114258) Damien - 2013-02-03 12:00:24 +0000
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/43027959-wallpaper-plug-should-have-basic-browsing-options?utm_campaign=plugin&utm_content=tracker%2F57462545&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F57462545&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
wallpaper plug should have basic browsing options - "wallpaper folder" and "custom folder" are fine now, but "picture folders" could be more detailed. I think the user should be able to dig in successive folders within the picture folder and go back the "picture folder root". I believe most people use nested folders to sort their pictures and this renders the selection "picture folder" almost useless compared to "custom folder", even though they would essentially use the same folder.
Only in "picture folders": the folders should be shown, a folder ".." should also be here to go back to the "picture folder root".
Launchpad Details: [#LP1114258](https://bugs.launchpad.net/bugs/1114258) Damien - 2013-02-03 12:00:24 +0000
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/43027959-wallpaper-plug-should-have-basic-browsing-options?utm_campaign=plugin&utm_content=tracker%2F57462545&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F57462545&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_defect
|
wallpaper plug should have basic browsing options wallpaper folder and custom folder are fine now but picture folders could be more detailed i think the user should be able to dig in successive folders within the picture folder and go back the picture folder root i believe most people use nested folders to sort their pictures and this renders the selection picture folder almost useless compared to custom folder even though they would essentially use the same folder only in picture folders the folders should be shown a folder should also be here to go back to the picture folder root launchpad details damien want to back this issue we accept bounties via
| 0
|
40,846
| 10,186,424,330
|
IssuesEvent
|
2019-08-10 13:21:12
|
tulir/mautrix-telegram
|
https://api.github.com/repos/tulir/mautrix-telegram
|
closed
|
Error preventing startup: readexactly() called while another coroutine is already waiting for incoming data
|
bug: defect
|
Happens to a random puppet during startup, preventing the bridge from running.
```
Jun 3 15:12:37 integrations python[863]: [2019-06-03 15:12:37,269] [ERROR@telethon.582redacted.network.connection.connection] Unexpected exception in the receive loop
Jun 3 15:12:37 integrations python[863]: Traceback (most recent call last):
Jun 3 15:12:37 integrations python[863]: File "/home/telegram/tgenv/lib/python3.6/site-packages/telethon/network/connection/connection.py", line 149, in _recv_loop
Jun 3 15:12:37 integrations python[863]: data = await self._recv()
Jun 3 15:12:37 integrations python[863]: File "/home/telegram/tgenv/lib/python3.6/site-packages/telethon/network/connection/tcpfull.py", line 33, in _recv
Jun 3 15:12:37 integrations python[863]: body = await self._reader.readexactly(packet_len - 8)
Jun 3 15:12:37 integrations python[863]: File "/usr/lib/python3.6/asyncio/streams.py", line 674, in readexactly
Jun 3 15:12:37 integrations python[863]: yield from self._wait_for_data('readexactly')
Jun 3 15:12:37 integrations python[863]: File "/usr/lib/python3.6/asyncio/streams.py", line 452, in _wait_for_data
Jun 3 15:12:37 integrations python[863]: 'already waiting for incoming data' % func_name)
Jun 3 15:12:37 integrations python[863]: RuntimeError: readexactly() called while another coroutine is already waiting for incoming data
```
|
1.0
|
Error preventing startup: readexactly() called while another coroutine is already waiting for incoming data - Happens to a random puppet during startup, preventing the bridge from running.
```
Jun 3 15:12:37 integrations python[863]: [2019-06-03 15:12:37,269] [ERROR@telethon.582redacted.network.connection.connection] Unexpected exception in the receive loop
Jun 3 15:12:37 integrations python[863]: Traceback (most recent call last):
Jun 3 15:12:37 integrations python[863]: File "/home/telegram/tgenv/lib/python3.6/site-packages/telethon/network/connection/connection.py", line 149, in _recv_loop
Jun 3 15:12:37 integrations python[863]: data = await self._recv()
Jun 3 15:12:37 integrations python[863]: File "/home/telegram/tgenv/lib/python3.6/site-packages/telethon/network/connection/tcpfull.py", line 33, in _recv
Jun 3 15:12:37 integrations python[863]: body = await self._reader.readexactly(packet_len - 8)
Jun 3 15:12:37 integrations python[863]: File "/usr/lib/python3.6/asyncio/streams.py", line 674, in readexactly
Jun 3 15:12:37 integrations python[863]: yield from self._wait_for_data('readexactly')
Jun 3 15:12:37 integrations python[863]: File "/usr/lib/python3.6/asyncio/streams.py", line 452, in _wait_for_data
Jun 3 15:12:37 integrations python[863]: 'already waiting for incoming data' % func_name)
Jun 3 15:12:37 integrations python[863]: RuntimeError: readexactly() called while another coroutine is already waiting for incoming data
```
|
defect
|
error preventing startup readexactly called while another coroutine is already waiting for incoming data happens to a random puppet during startup preventing the bridge from running jun integrations python unexpected exception in the receive loop jun integrations python traceback most recent call last jun integrations python file home telegram tgenv lib site packages telethon network connection connection py line in recv loop jun integrations python data await self recv jun integrations python file home telegram tgenv lib site packages telethon network connection tcpfull py line in recv jun integrations python body await self reader readexactly packet len jun integrations python file usr lib asyncio streams py line in readexactly jun integrations python yield from self wait for data readexactly jun integrations python file usr lib asyncio streams py line in wait for data jun integrations python already waiting for incoming data func name jun integrations python runtimeerror readexactly called while another coroutine is already waiting for incoming data
| 1
|
178,463
| 6,609,072,164
|
IssuesEvent
|
2017-09-19 13:25:56
|
eMoflon/emoflon-tool
|
https://api.github.com/repos/eMoflon/emoflon-tool
|
closed
|
Make /injection a source folder
|
feature-request low-priority
|
If /injection is configured as a source folder in Eclipse, its subfolder structure is nicely presented as package.
|
1.0
|
Make /injection a source folder - If /injection is configured as a source folder in Eclipse, its subfolder structure is nicely presented as package.
|
non_defect
|
make injection a source folder if injection is configured as a source folder in eclipse its subfolder structure is nicely presented as package
| 0
|
427,773
| 12,399,006,664
|
IssuesEvent
|
2020-05-21 03:43:10
|
orbeon/orbeon-forms
|
https://api.github.com/repos/orbeon/orbeon-forms
|
closed
|
Dynamic dropdown with search to evaluate `resource` in outer scope
|
Area: XBL Components Priority: Regression
|
This used to be the case with the autocomplete, so this is, in a way, a regression, plus, since the value of `resource` is provided by users and is taken as a VT, it just makes sense for XPath to be evaluated in the outer rather than the inner scope.
[+1 from customer](https://3.basecamp.com/3600924/buckets/2016476/messages/2646556535)
|
1.0
|
Dynamic dropdown with search to evaluate `resource` in outer scope - This used to be the case with the autocomplete, so this is, in a way, a regression, plus, since the value of `resource` is provided by users and is taken as a VT, it just makes sense for XPath to be evaluated in the outer rather than the inner scope.
[+1 from customer](https://3.basecamp.com/3600924/buckets/2016476/messages/2646556535)
|
non_defect
|
dynamic dropdown with search to evaluate resource in outer scope this used to be the case with the autocomplete so this is in a way a regression plus since the value of resource is provided by users and is taken as a vt it just makes sense for xpath to be evaluated in the outer rather than the inner scope
| 0
|
5,630
| 2,610,192,192
|
IssuesEvent
|
2015-02-26 19:00:43
|
chrsmith/quchuseban
|
https://api.github.com/repos/chrsmith/quchuseban
|
opened
|
咨询去除面部色斑的小妙招
|
auto-migrated Priority-Medium Type-Defect
|
```
《摘要》
愉快跟不愉快的回忆,比如一个硬币的两面,存在于我们的��
�一段情感里。就像那个著名的“蝴蝶效应”,如果你经常记�
��不愉快的人、不愉快的事,生活就跟着变得不愉快起来。相
反,有些女人却能在跟老公吵架的时候及其她求婚时的表情��
�他怀抱的温暖。这里的“吵”是一种乐观、积极的沟通方式�
��这样的女人即便是面临命运得不测风云,也不会唉声叹气,
而当它是动力。面带微笑、坦然自处,男人有乐观女人的相��
�,一生都将阳光灿烂。去除面部色斑的小妙招,
《客户案例》
我想知道一起拥有的回忆是浓了你还是醉了我,在这个��
�花飘落的季节里,我又想起了你,想起了那个雪地里雨伞下�
��长而又安静的拥抱,回想起你的眼泪曾伴随着雪花一起纷飞
,我想那是一种多么绝望而又撕心裂肺、痛彻心扉的的凄美��
�谢谢你曾用你的青春带走了我的忧伤,我在26岁那年,脸上��
�出了许多黄褐斑。看到满脸的黄褐斑,我心中甚是苦恼。于�
��,为了恢复年轻的美丽的容颜,我决定治疗黄褐斑。可是,
黄褐斑的治疗如何进行最有效呢? 怎样去除,面部黄褐斑</br>
后来为了去斑我开始依赖于网络,在网上搜索可以帮我��
�掉斑点的方法,只要是有点希望,我便会想法去试试,一个�
��然的机会让我遇见了黛芙薇尔去斑,遇见就不再错过,这句
话说的真好。于是我就深入的了解了一下黛芙薇尔,最后我��
�质询了他们的专家跟客服,他们说我的病情很适合他们的产�
��,于是我就定购了几个周期的黛芙薇尔。怎样去除,面部黄��
�斑</br>
我每天坚持使用,并保持一定量的户外运动,坚持了一��
�多月,就发现脸上的斑点真的慢慢变淡了。接下来的日子我�
��心情更加开朗乐观起来,过了半个月的时间,就发现脸上的
黄褐斑已经彻底去除了!我觉得那时候自己是多么自信的女人�
��,我的幸福终于被我牢牢抓住了!真的很感谢黛芙薇尔,是��
�让我摆脱了色斑的困扰,彻底去除了我脸上的色斑,现在着�
��觉得没斑的感觉真是太好啦。
阅读了去除面部色斑的小妙招,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
去除面部色斑的小妙招,同时为您分享祛斑小方法
1、买适量的黄豆、醋(我用的是米醋)
2、洗干净黄豆,把醋倒出半瓶。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:22
|
1.0
|
咨询去除面部色斑的小妙招 - ```
《摘要》
愉快跟不愉快的回忆,比如一个硬币的两面,存在于我们的��
�一段情感里。就像那个著名的“蝴蝶效应”,如果你经常记�
��不愉快的人、不愉快的事,生活就跟着变得不愉快起来。相
反,有些女人却能在跟老公吵架的时候及其她求婚时的表情��
�他怀抱的温暖。这里的“吵”是一种乐观、积极的沟通方式�
��这样的女人即便是面临命运得不测风云,也不会唉声叹气,
而当它是动力。面带微笑、坦然自处,男人有乐观女人的相��
�,一生都将阳光灿烂。去除面部色斑的小妙招,
《客户案例》
我想知道一起拥有的回忆是浓了你还是醉了我,在这个��
�花飘落的季节里,我又想起了你,想起了那个雪地里雨伞下�
��长而又安静的拥抱,回想起你的眼泪曾伴随着雪花一起纷飞
,我想那是一种多么绝望而又撕心裂肺、痛彻心扉的的凄美��
�谢谢你曾用你的青春带走了我的忧伤,我在26岁那年,脸上��
�出了许多黄褐斑。看到满脸的黄褐斑,我心中甚是苦恼。于�
��,为了恢复年轻的美丽的容颜,我决定治疗黄褐斑。可是,
黄褐斑的治疗如何进行最有效呢? 怎样去除,面部黄褐斑</br>
后来为了去斑我开始依赖于网络,在网上搜索可以帮我��
�掉斑点的方法,只要是有点希望,我便会想法去试试,一个�
��然的机会让我遇见了黛芙薇尔去斑,遇见就不再错过,这句
话说的真好。于是我就深入的了解了一下黛芙薇尔,最后我��
�质询了他们的专家跟客服,他们说我的病情很适合他们的产�
��,于是我就定购了几个周期的黛芙薇尔。怎样去除,面部黄��
�斑</br>
我每天坚持使用,并保持一定量的户外运动,坚持了一��
�多月,就发现脸上的斑点真的慢慢变淡了。接下来的日子我�
��心情更加开朗乐观起来,过了半个月的时间,就发现脸上的
黄褐斑已经彻底去除了!我觉得那时候自己是多么自信的女人�
��,我的幸福终于被我牢牢抓住了!真的很感谢黛芙薇尔,是��
�让我摆脱了色斑的困扰,彻底去除了我脸上的色斑,现在着�
��觉得没斑的感觉真是太好啦。
阅读了去除面部色斑的小妙招,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
去除面部色斑的小妙招,同时为您分享祛斑小方法
1、买适量的黄豆、醋(我用的是米醋)
2、洗干净黄豆,把醋倒出半瓶。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:22
|
defect
|
咨询去除面部色斑的小妙招 《摘要》 愉快跟不愉快的回忆,比如一个硬币的两面,存在于我们的�� �一段情感里。就像那个著名的“蝴蝶效应”,如果你经常记� ��不愉快的人、不愉快的事,生活就跟着变得不愉快起来。相 反,有些女人却能在跟老公吵架的时候及其她求婚时的表情�� �他怀抱的温暖。这里的“吵”是一种乐观、积极的沟通方式� ��这样的女人即便是面临命运得不测风云,也不会唉声叹气, 而当它是动力。面带微笑、坦然自处,男人有乐观女人的相�� �,一生都将阳光灿烂。去除面部色斑的小妙招, 《客户案例》 我想知道一起拥有的回忆是浓了你还是醉了我,在这个�� �花飘落的季节里,我又想起了你,想起了那个雪地里雨伞下� ��长而又安静的拥抱,回想起你的眼泪曾伴随着雪花一起纷飞 ,我想那是一种多么绝望而又撕心裂肺、痛彻心扉的的凄美�� �谢谢你曾用你的青春带走了我的忧伤, ,脸上�� �出了许多黄褐斑。看到满脸的黄褐斑,我心中甚是苦恼。于� ��,为了恢复年轻的美丽的容颜,我决定治疗黄褐斑。可是, 黄褐斑的治疗如何进行最有效呢 怎样去除 面部黄褐斑 后来为了去斑我开始依赖于网络,在网上搜索可以帮我�� �掉斑点的方法,只要是有点希望,我便会想法去试试,一个� ��然的机会让我遇见了黛芙薇尔去斑,遇见就不再错过,这句 话说的真好。于是我就深入的了解了一下黛芙薇尔,最后我�� �质询了他们的专家跟客服,他们说我的病情很适合他们的产� ��,于是我就定购了几个周期的黛芙薇尔。怎样去除 面部黄�� �斑 我每天坚持使用,并保持一定量的户外运动,坚持了一�� �多月,就发现脸上的斑点真的慢慢变淡了。接下来的日子我� ��心情更加开朗乐观起来,过了半个月的时间,就发现脸上的 黄褐斑已经彻底去除了 我觉得那时候自己是多么自信的女人� ��,我的幸福终于被我牢牢抓住了 真的很感谢黛芙薇尔,是�� �让我摆脱了色斑的困扰,彻底去除了我脸上的色斑,现在着� ��觉得没斑的感觉真是太好啦。 阅读了去除面部色斑的小妙招,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 去除面部色斑的小妙招,同时为您分享祛斑小方法 、买适量的黄豆、醋(我用的是米醋) 、洗干净黄豆,把醋倒出半瓶。 original issue reported on code google com by additive gmail com on jul at
| 1
|
34,205
| 7,395,775,651
|
IssuesEvent
|
2018-03-18 02:42:54
|
pyscripter/MustangpeakCommonLib
|
https://api.github.com/repos/pyscripter/MustangpeakCommonLib
|
closed
|
how to install to Delphi XE3?
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. download with svn
2. open is file with InnoSetup
3. error "IDE install.txt not found"
What is the expected output? What do you see instead?
require other application or package?
What version of the product are you using? On what operating system?
DelphiXE3 on Win7
```
Original issue reported on code.google.com by `yl0g...@gmail.com` on 23 Jul 2015 at 6:43
|
1.0
|
how to install to Delphi XE3? - ```
What steps will reproduce the problem?
1. download with svn
2. open is file with InnoSetup
3. error "IDE install.txt not found"
What is the expected output? What do you see instead?
require other application or package?
What version of the product are you using? On what operating system?
DelphiXE3 on Win7
```
Original issue reported on code.google.com by `yl0g...@gmail.com` on 23 Jul 2015 at 6:43
|
defect
|
how to install to delphi what steps will reproduce the problem download with svn open is file with innosetup error ide install txt not found what is the expected output what do you see instead require other application or package what version of the product are you using on what operating system on original issue reported on code google com by gmail com on jul at
| 1
|
1,771
| 3,369,260,855
|
IssuesEvent
|
2015-11-23 09:10:44
|
rhiot/rhiot
|
https://api.github.com/repos/rhiot/rhiot
|
closed
|
create camel-labs.github.io website via Github Pages
|
infrastructure
|
creation of camel-labs.github.io repository needed
|
1.0
|
create camel-labs.github.io website via Github Pages - creation of camel-labs.github.io repository needed
|
non_defect
|
create camel labs github io website via github pages creation of camel labs github io repository needed
| 0
|
398,623
| 11,742,044,091
|
IssuesEvent
|
2020-03-11 23:25:17
|
thaliawww/concrexit
|
https://api.github.com/repos/thaliawww/concrexit
|
closed
|
Newsletter ordering only possible for info or events seperately
|
bug newsletter priority: medium
|
In GitLab by @lscholten on Mar 29, 2017, 20:36
### One-sentence description
Newsletter ordering only possible for info or events seperately
### Current behaviour / Reproducing the bug
1. Add a newsletter
2. Add content and events
3. Try to mix them in the order
4. No mixing in the result
### Expected behaviour
1. Add a newsletter
2. Add content and events
3. Try to mix them in the order
4. Result is correct
|
1.0
|
Newsletter ordering only possible for info or events seperately - In GitLab by @lscholten on Mar 29, 2017, 20:36
### One-sentence description
Newsletter ordering only possible for info or events seperately
### Current behaviour / Reproducing the bug
1. Add a newsletter
2. Add content and events
3. Try to mix them in the order
4. No mixing in the result
### Expected behaviour
1. Add a newsletter
2. Add content and events
3. Try to mix them in the order
4. Result is correct
|
non_defect
|
newsletter ordering only possible for info or events seperately in gitlab by lscholten on mar one sentence description newsletter ordering only possible for info or events seperately current behaviour reproducing the bug add a newsletter add content and events try to mix them in the order no mixing in the result expected behaviour add a newsletter add content and events try to mix them in the order result is correct
| 0
|
12,325
| 2,691,591,655
|
IssuesEvent
|
2015-03-31 22:33:34
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
closed
|
Peacock temp input generation bug
|
C: Peacock P: normal T: defect
|
When running the input file
```
moose/modules/phase_field/examples/multiphase/DerivativeMultiPhaseMaterial.i
```
from peacock, MOOSE throws the following error:
```
Error: the following unidentified entries were found in your input file:
Variables/eta3/InitialCondition/sqrt(x^2+y^2);if(r>7,1,0)
```
The issue seems to be that the temp input that peacock writes out does not contain quotes around the function expression in the variable initial conditions:
```
function = r:=sqrt(x^2+y^2);if(r>7,1,0)
```
|
1.0
|
Peacock temp input generation bug - When running the input file
```
moose/modules/phase_field/examples/multiphase/DerivativeMultiPhaseMaterial.i
```
from peacock, MOOSE throws the following error:
```
Error: the following unidentified entries were found in your input file:
Variables/eta3/InitialCondition/sqrt(x^2+y^2);if(r>7,1,0)
```
The issue seems to be that the temp input that peacock writes out does not contain quotes around the function expression in the variable initial conditions:
```
function = r:=sqrt(x^2+y^2);if(r>7,1,0)
```
|
defect
|
peacock temp input generation bug when running the input file moose modules phase field examples multiphase derivativemultiphasematerial i from peacock moose throws the following error error the following unidentified entries were found in your input file variables initialcondition sqrt x y if r the issue seems to be that the temp input that peacock writes out does not contain quotes around the function expression in the variable initial conditions function r sqrt x y if r
| 1
|
62,415
| 17,023,918,819
|
IssuesEvent
|
2021-07-03 04:33:16
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Login using openid does not function after signup
|
Component: website Priority: minor Resolution: invalid Type: defect
|
**[Submitted to the original trac issue database at 1.36pm, Saturday, 14th March 2015]**
Problem:
I signed up/registered on the website at[https://www.openstreetmap.org/user/new] using google as the third party. After that whenever I Tried to use google as the third party for login, it says "Your ID is not associated with a OpenStreetMap account yet". and redirects me to the signup page instead of logging me in.
If I proceed with this signup form it says that my email address is already taken.
However if I try to login using my email address and password everything works fine.
Environment:
Firefox 33.1 on Fedora 21.
Google Chrome Version 39.0.2171.95 (64-bit) on Fedora 21.
Steps to Repoduce on the environment:
1.Signup using google as a third party for an open street maps account.
2.Logout of the account if you are loggedin.
3.Try to login again using google as the third party openid provider.
Is it just me or is anyone else experiencing the same issue?
|
1.0
|
Login using openid does not function after signup - **[Submitted to the original trac issue database at 1.36pm, Saturday, 14th March 2015]**
Problem:
I signed up/registered on the website at[https://www.openstreetmap.org/user/new] using google as the third party. After that whenever I Tried to use google as the third party for login, it says "Your ID is not associated with a OpenStreetMap account yet". and redirects me to the signup page instead of logging me in.
If I proceed with this signup form it says that my email address is already taken.
However if I try to login using my email address and password everything works fine.
Environment:
Firefox 33.1 on Fedora 21.
Google Chrome Version 39.0.2171.95 (64-bit) on Fedora 21.
Steps to Repoduce on the environment:
1.Signup using google as a third party for an open street maps account.
2.Logout of the account if you are loggedin.
3.Try to login again using google as the third party openid provider.
Is it just me or is anyone else experiencing the same issue?
|
defect
|
login using openid does not function after signup problem i signed up registered on the website at using google as the third party after that whenever i tried to use google as the third party for login it says your id is not associated with a openstreetmap account yet and redirects me to the signup page instead of logging me in if i proceed with this signup form it says that my email address is already taken however if i try to login using my email address and password everything works fine environment firefox on fedora google chrome version bit on fedora steps to repoduce on the environment signup using google as a third party for an open street maps account logout of the account if you are loggedin try to login again using google as the third party openid provider is it just me or is anyone else experiencing the same issue
| 1
|
507,955
| 14,685,303,149
|
IssuesEvent
|
2021-01-01 08:18:45
|
sujith-bhatt/InfectiousDiseaseModelling
|
https://api.github.com/repos/sujith-bhatt/InfectiousDiseaseModelling
|
opened
|
Obtain Data
|
high priority
|
Use the JHU dataset and graph the cumulative incidence of cases confirmed and recovered.
|
1.0
|
Obtain Data - Use the JHU dataset and graph the cumulative incidence of cases confirmed and recovered.
|
non_defect
|
obtain data use the jhu dataset and graph the cumulative incidence of cases confirmed and recovered
| 0
|
392,950
| 11,598,025,573
|
IssuesEvent
|
2020-02-24 22:07:42
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.wunderlist.com - Page is not loaded
|
browser-firefox engine-gecko form-v2-experiment priority-normal severity-critical
|
<!-- @browser: Firefox 74.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:74.0) Gecko/20100101 Firefox/74.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @extra_labels: form-v2-experiment -->
**URL**: https://www.wunderlist.com/webapp#/lists/inbox
**Browser / Version**: Firefox 74.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Firefox Developer v74 suddenly starting to give just a blank page for signed in lists or account settings pages on Wunderlist.com. Used to work fine. Looking at Page Sourcecode there is HTML being sent back to browser but BODY tag is completely empty of content. No such issue at the same time using Edge Chromium v80. I did clear cookies in Firefox and then re-login to Wunderlist, but still same behaviour.
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/2/94ecf378-bc62-4d9f-a068-605c22c53ea3.jpg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200214015126</li><li>channel: aurora</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/2/d43c8625-0c40-4312-9f33-5385f0554014)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.wunderlist.com - Page is not loaded - <!-- @browser: Firefox 74.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:74.0) Gecko/20100101 Firefox/74.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @extra_labels: form-v2-experiment -->
**URL**: https://www.wunderlist.com/webapp#/lists/inbox
**Browser / Version**: Firefox 74.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Firefox Developer v74 suddenly starting to give just a blank page for signed in lists or account settings pages on Wunderlist.com. Used to work fine. Looking at Page Sourcecode there is HTML being sent back to browser but BODY tag is completely empty of content. No such issue at the same time using Edge Chromium v80. I did clear cookies in Firefox and then re-login to Wunderlist, but still same behaviour.
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/2/94ecf378-bc62-4d9f-a068-605c22c53ea3.jpg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200214015126</li><li>channel: aurora</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/2/d43c8625-0c40-4312-9f33-5385f0554014)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
page is not loaded url browser version firefox operating system windows tested another browser yes edge problem type site is not usable description page not loading correctly steps to reproduce firefox developer suddenly starting to give just a blank page for signed in lists or account settings pages on wunderlist com used to work fine looking at page sourcecode there is html being sent back to browser but body tag is completely empty of content no such issue at the same time using edge chromium i did clear cookies in firefox and then re login to wunderlist but still same behaviour view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel aurora hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
6,386
| 2,610,241,953
|
IssuesEvent
|
2015-02-26 19:17:02
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
台州割包皮包茎哪家男科医院好
|
auto-migrated Priority-Medium Type-Defect
|
```
台州割包皮包茎哪家男科医院好【台州五洲生殖医院】24小时
健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地�
��:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐1
04、108、118、198及椒江一金清公交车直达枫南小区,乘坐107、
105、109、112、901、
902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 12:30
|
1.0
|
台州割包皮包茎哪家男科医院好 - ```
台州割包皮包茎哪家男科医院好【台州五洲生殖医院】24小时
健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地�
��:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐1
04、108、118、198及椒江一金清公交车直达枫南小区,乘坐107、
105、109、112、901、
902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 12:30
|
defect
|
台州割包皮包茎哪家男科医院好 台州割包皮包茎哪家男科医院好【台州五洲生殖医院】 健康咨询热线 微信号tzwzszyy 医院地� �� (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
4,590
| 2,610,120,718
|
IssuesEvent
|
2015-02-26 18:37:28
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
Can't add blog on livejournal
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
When I'm trying to add my Livejournal's blog I fill lines with login and
password but browser with a small new window tells me that i've entered wrong
login or password for area "lj" at www.livejournal.com:80.
What browser are you using?
Safari, the same problem in Firefox
What version of ScribeFire are you running?
latest. I've installed it just an hour ago.
```
-----
Original issue reported on code.google.com by `strigov....@gmail.com` on 1 Oct 2010 at 4:40
|
1.0
|
Can't add blog on livejournal - ```
What's the problem?
When I'm trying to add my Livejournal's blog I fill lines with login and
password but browser with a small new window tells me that i've entered wrong
login or password for area "lj" at www.livejournal.com:80.
What browser are you using?
Safari, the same problem in Firefox
What version of ScribeFire are you running?
latest. I've installed it just an hour ago.
```
-----
Original issue reported on code.google.com by `strigov....@gmail.com` on 1 Oct 2010 at 4:40
|
defect
|
can t add blog on livejournal what s the problem when i m trying to add my livejournal s blog i fill lines with login and password but browser with a small new window tells me that i ve entered wrong login or password for area lj at what browser are you using safari the same problem in firefox what version of scribefire are you running latest i ve installed it just an hour ago original issue reported on code google com by strigov gmail com on oct at
| 1
|
70,374
| 23,145,049,111
|
IssuesEvent
|
2022-07-28 23:11:00
|
NREL/EnergyPlus
|
https://api.github.com/repos/NREL/EnergyPlus
|
closed
|
CheckConvexity is not removing collinear points consistently, causing fatal error
|
Defect
|
Issue overview
--------------
In the [attached example file](https://github.com/TiejunWu/EnergyPlus-IDF-Files/blob/master/2Zones_Uncontrolled.idf) , there are 3 surfaces with 8 vertices each, 2 of the surfaces are boundary condition to each other. 4 of the 8 vertices are collinear points.
```
BuildingSurface:Detailed,
Zn001:Ceiling001, !- Name
Ceiling, !- Surface Type
FLOOR, !- Construction Name
ZONE ONE, !- Zone Name
, !- Space Name
Surface, !- Outside Boundary Condition
Zn002:Flr001, !- Outside Boundary Condition Object
NoSun, !- Sun Exposure
NoWind, !- Wind Exposure
, !- View Factor to Ground
, !- Number of Vertices
25.825, -17.323, 3.658, !- X,Y,Z 1 {m}
25.825, -12.15, 3.658, !- X,Y,Z 2 {m}
25.825, -7.051, 3.658, !- X,Y,Z 3 {m}
16.859, -7.051, 3.658, !- X,Y,Z 4 {m}
8.008, -7.051, 3.658, !- X,Y,Z 5 {m}
8.008, -12.15, 3.658, !- X,Y,Z 6 {m}
8.008, -17.323, 3.658, !- X,Y,Z 7 {m}
16.859, -17.323, 3.658; !- X,Y,Z 8 {m}
BuildingSurface:Detailed,
Zn002:Flr001, !- Name
Floor, !- Surface Type
FLOOR, !- Construction Name
ZONE TWO, !- Zone Name
, !- Space Name
Surface, !- Outside Boundary Condition
Zn001:Ceiling001, !- Outside Boundary Condition Object
NoSun, !- Sun Exposure
NoWind, !- Wind Exposure
, !- View Factor to Ground
, !- Number of Vertices
8.008, -17.323, 3.658, !- X,Y,Z 1 {m}
8.008, -12.15, 3.658, !- X,Y,Z 2 {m}
8.008, -7.051, 3.658, !- X,Y,Z 3 {m}
16.859, -7.051, 3.658, !- X,Y,Z 4 {m}
25.825, -7.051, 3.658, !- X,Y,Z 5 {m}
25.825, -12.15, 3.658, !- X,Y,Z 6 {m}
25.825, -17.323, 3.658, !- X,Y,Z 7 {m}
16.859, -17.323, 3.658; !- X,Y,Z 8 {m}
BuildingSurface:Detailed,
Zn002:Roof001, !- Name
Roof, !- Surface Type
ROOF31, !- Construction Name
ZONE TWO, !- Zone Name
, !- Space Name
Outdoors, !- Outside Boundary Condition
, !- Outside Boundary Condition Object
SunExposed, !- Sun Exposure
WindExposed, !- Wind Exposure
, !- View Factor to Ground
, !- Number of Vertices
16.859, -17.323, 7.315, !- X,Y,Z 1 {m}
25.825, -17.323, 7.315, !- X,Y,Z 2 {m}
25.825, -12.15, 7.315, !- X,Y,Z 3 {m}
25.825, -7.051, 7.315, !- X,Y,Z 4 {m}
16.859, -7.051, 7.315, !- X,Y,Z 5 {m}
8.008, -7.051, 7.315, !- X,Y,Z 6 {m}
8.008, -12.15, 7.315, !- X,Y,Z 7 {m}
8.008, -17.323, 7.315; !- X,Y,Z 8 {m}
```
When I run a simulation, it generates the following error message:
```
** Warning ** CheckConvexity: Surface="ZN001:CEILING001", Collinear points have been removed.
** Warning ** CheckConvexity: Surface="ZN001:CEILING001": The vertex points has been reprocessed as Sides = 7
** Warning ** CheckConvexity: Surface="ZN002:FLR001", Collinear points have been removed.
** Warning ** CheckConvexity: Surface="ZN002:FLR001": The vertex points has been reprocessed as Sides = 4
** Warning ** CheckConvexity: Surface="ZN002:ROOF001", Collinear points have been removed.
** Warning ** CheckConvexity: Surface="ZN002:ROOF001": The vertex points has been reprocessed as Sides = 4
** Severe ** RoofCeiling:Detailed="ZN001:CEILING001", Vertex size mismatch between base surface :ZN001:CEILING001 and outside boundary surface: ZN002:FLR001
** ~~~ ** The vertex sizes are 7 for base surface and 4 for outside boundary surface. Please check inputs.
** Severe ** RoofCeiling:Detailed="ZN002:FLR001", Vertex size mismatch between base surface :ZN002:FLR001 and outside boundary surface: ZN001:CEILING001
** ~~~ ** The vertex sizes are 4 for base surface and 7 for outside boundary surface. Please check inputs.
** Fatal ** GetSurfaceData: Errors discovered, program terminates.
```
Looks like CheckConvexity removed only 1 of the 4 collinear points from surface ZN001:CEILING001 and caused the fatal error. The X and Y coordinates of these vertices are the same except the sequence of vertices are different, not sure why the program did not figure out all the collinear points.
### Details
Some additional details for this issue (if relevant):
- Windows 10 64 Bit
- EnergyPlus 9.6
- N/A
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [x] Defect file added (list location of [defect file here](https://github.com/TiejunWu/EnergyPlus-IDF-Files/blob/master/2Zones_Uncontrolled.idf) )
- [x] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
1.0
|
CheckConvexity is not removing collinear points consistently, causing fatal error - Issue overview
--------------
In the [attached example file](https://github.com/TiejunWu/EnergyPlus-IDF-Files/blob/master/2Zones_Uncontrolled.idf) , there are 3 surfaces with 8 vertices each, 2 of the surfaces are boundary condition to each other. 4 of the 8 vertices are collinear points.
```
BuildingSurface:Detailed,
Zn001:Ceiling001, !- Name
Ceiling, !- Surface Type
FLOOR, !- Construction Name
ZONE ONE, !- Zone Name
, !- Space Name
Surface, !- Outside Boundary Condition
Zn002:Flr001, !- Outside Boundary Condition Object
NoSun, !- Sun Exposure
NoWind, !- Wind Exposure
, !- View Factor to Ground
, !- Number of Vertices
25.825, -17.323, 3.658, !- X,Y,Z 1 {m}
25.825, -12.15, 3.658, !- X,Y,Z 2 {m}
25.825, -7.051, 3.658, !- X,Y,Z 3 {m}
16.859, -7.051, 3.658, !- X,Y,Z 4 {m}
8.008, -7.051, 3.658, !- X,Y,Z 5 {m}
8.008, -12.15, 3.658, !- X,Y,Z 6 {m}
8.008, -17.323, 3.658, !- X,Y,Z 7 {m}
16.859, -17.323, 3.658; !- X,Y,Z 8 {m}
BuildingSurface:Detailed,
Zn002:Flr001, !- Name
Floor, !- Surface Type
FLOOR, !- Construction Name
ZONE TWO, !- Zone Name
, !- Space Name
Surface, !- Outside Boundary Condition
Zn001:Ceiling001, !- Outside Boundary Condition Object
NoSun, !- Sun Exposure
NoWind, !- Wind Exposure
, !- View Factor to Ground
, !- Number of Vertices
8.008, -17.323, 3.658, !- X,Y,Z 1 {m}
8.008, -12.15, 3.658, !- X,Y,Z 2 {m}
8.008, -7.051, 3.658, !- X,Y,Z 3 {m}
16.859, -7.051, 3.658, !- X,Y,Z 4 {m}
25.825, -7.051, 3.658, !- X,Y,Z 5 {m}
25.825, -12.15, 3.658, !- X,Y,Z 6 {m}
25.825, -17.323, 3.658, !- X,Y,Z 7 {m}
16.859, -17.323, 3.658; !- X,Y,Z 8 {m}
BuildingSurface:Detailed,
Zn002:Roof001, !- Name
Roof, !- Surface Type
ROOF31, !- Construction Name
ZONE TWO, !- Zone Name
, !- Space Name
Outdoors, !- Outside Boundary Condition
, !- Outside Boundary Condition Object
SunExposed, !- Sun Exposure
WindExposed, !- Wind Exposure
, !- View Factor to Ground
, !- Number of Vertices
16.859, -17.323, 7.315, !- X,Y,Z 1 {m}
25.825, -17.323, 7.315, !- X,Y,Z 2 {m}
25.825, -12.15, 7.315, !- X,Y,Z 3 {m}
25.825, -7.051, 7.315, !- X,Y,Z 4 {m}
16.859, -7.051, 7.315, !- X,Y,Z 5 {m}
8.008, -7.051, 7.315, !- X,Y,Z 6 {m}
8.008, -12.15, 7.315, !- X,Y,Z 7 {m}
8.008, -17.323, 7.315; !- X,Y,Z 8 {m}
```
When I run a simulation, it generates the following error message:
```
** Warning ** CheckConvexity: Surface="ZN001:CEILING001", Collinear points have been removed.
** Warning ** CheckConvexity: Surface="ZN001:CEILING001": The vertex points has been reprocessed as Sides = 7
** Warning ** CheckConvexity: Surface="ZN002:FLR001", Collinear points have been removed.
** Warning ** CheckConvexity: Surface="ZN002:FLR001": The vertex points has been reprocessed as Sides = 4
** Warning ** CheckConvexity: Surface="ZN002:ROOF001", Collinear points have been removed.
** Warning ** CheckConvexity: Surface="ZN002:ROOF001": The vertex points has been reprocessed as Sides = 4
** Severe ** RoofCeiling:Detailed="ZN001:CEILING001", Vertex size mismatch between base surface :ZN001:CEILING001 and outside boundary surface: ZN002:FLR001
** ~~~ ** The vertex sizes are 7 for base surface and 4 for outside boundary surface. Please check inputs.
** Severe ** RoofCeiling:Detailed="ZN002:FLR001", Vertex size mismatch between base surface :ZN002:FLR001 and outside boundary surface: ZN001:CEILING001
** ~~~ ** The vertex sizes are 4 for base surface and 7 for outside boundary surface. Please check inputs.
** Fatal ** GetSurfaceData: Errors discovered, program terminates.
```
Looks like CheckConvexity removed only 1 of the 4 collinear points from surface ZN001:CEILING001 and caused the fatal error. The X and Y coordinates of these vertices are the same except the sequence of vertices are different, not sure why the program did not figure out all the collinear points.
### Details
Some additional details for this issue (if relevant):
- Windows 10 64 Bit
- EnergyPlus 9.6
- N/A
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [x] Defect file added (list location of [defect file here](https://github.com/TiejunWu/EnergyPlus-IDF-Files/blob/master/2Zones_Uncontrolled.idf) )
- [x] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
defect
|
checkconvexity is not removing collinear points consistently causing fatal error issue overview in the there are surfaces with vertices each of the surfaces are boundary condition to each other of the vertices are collinear points buildingsurface detailed name ceiling surface type floor construction name zone one zone name space name surface outside boundary condition outside boundary condition object nosun sun exposure nowind wind exposure view factor to ground number of vertices x y z m x y z m x y z m x y z m x y z m x y z m x y z m x y z m buildingsurface detailed name floor surface type floor construction name zone two zone name space name surface outside boundary condition outside boundary condition object nosun sun exposure nowind wind exposure view factor to ground number of vertices x y z m x y z m x y z m x y z m x y z m x y z m x y z m x y z m buildingsurface detailed name roof surface type construction name zone two zone name space name outdoors outside boundary condition outside boundary condition object sunexposed sun exposure windexposed wind exposure view factor to ground number of vertices x y z m x y z m x y z m x y z m x y z m x y z m x y z m x y z m when i run a simulation it generates the following error message warning checkconvexity surface collinear points have been removed warning checkconvexity surface the vertex points has been reprocessed as sides warning checkconvexity surface collinear points have been removed warning checkconvexity surface the vertex points has been reprocessed as sides warning checkconvexity surface collinear points have been removed warning checkconvexity surface the vertex points has been reprocessed as sides severe roofceiling detailed vertex size mismatch between base surface and outside boundary surface the vertex sizes are for base surface and for outside boundary surface please check inputs severe roofceiling detailed vertex size mismatch between base surface and outside boundary surface the vertex sizes are for base surface and for outside boundary surface please check inputs fatal getsurfacedata errors discovered program terminates looks like checkconvexity removed only of the collinear points from surface and caused the fatal error the x and y coordinates of these vertices are the same except the sequence of vertices are different not sure why the program did not figure out all the collinear points details some additional details for this issue if relevant windows bit energyplus n a checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
| 1
|
15,368
| 2,850,673,473
|
IssuesEvent
|
2015-05-31 19:33:09
|
damonkohler/android-scripting
|
https://api.github.com/repos/damonkohler/android-scripting
|
closed
|
sl4a + no wireless = sudden death
|
auto-migrated Priority-Medium Type-Defect
|
```
What device(s) are you experiencing the problem on?
HTC Dream
What firmware version are you running on the device?
cyanogenmod 6.1.0
What steps will reproduce the problem?
1. start sl4a
2. turn off/lose wireless connection
3. launch any 'hello world' script
What is the expected output? What do you see instead?
sl4a crashes immediately
What version of the product are you using? On what operating system?
latest sl4a release/android 2.2 (froyo)
Please provide any additional information below.
Snip of logcat (not sure it says much tho):
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Interpreter discovered: com.googlecode.bshforandroid
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Binary: /system/bin/dalvikvm
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Interpreter discovered: com.googlecode.bshforandroid
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Binary: /system/bin/dalvikvm
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Interpreter discovered: com.googlecode.jrubyforandroid
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Binary: /system/bin/dalvikvm
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Interpreter discovered: com.googlecode.jrubyforandroid
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Binary: /system/bin/dalvikvm
D/org.jraf.android.nolock.NoLockService( 656): onLowMemory
D/org.jraf.android.nolock.NoLockService( 656): handleStart
mKeyguardLock != null: reenabling keyguard
D/dalvikvm( 656): GC_EXPLICIT freed 60 objects / 3768 bytes in 103ms
D/dalvikvm( 1353): GC_EXPLICIT freed 6 objects / 216 bytes in 105ms
D/dalvikvm( 1401): GC_EXPLICIT freed 7 objects / 248 bytes in 151ms
W/ActivityManager( 190): Timeout executing service:
ServiceRecord{4406fbf8
com.googlecode.android_scripting/.activity.ScriptingLayerService}
I/ActivityManager( 190): Crashing app skipping ANR:
ProcessRecord{43f54748 2035:com.googlecode.android_scripting/10064}
Executing service
com.googlecode.android_scripting/.activity.ScriptingLayerService
D/dalvikvm( 435): GC_EXPLICIT freed 7 objects / 256 bytes in 116ms
D/dalvikvm( 978): GC_EXPLICIT freed 1039 objects / 47768 bytes in
162ms
I/Process ( 2035): Sending signal. PID: 2035 SIG: 9
I/ActivityManager( 190): Process com.googlecode.android_scripting
(pid 2035) has died.
W/ActivityManager( 190): Service crashed 2 times, stopping:
ServiceRecord{4406fbf8
com.googlecode.android_scripting/.activity.ScriptingLayerService}
```
Original issue reported on code.google.com by `guy.berl...@gmail.com` on 8 Jan 2011 at 2:28
|
1.0
|
sl4a + no wireless = sudden death - ```
What device(s) are you experiencing the problem on?
HTC Dream
What firmware version are you running on the device?
cyanogenmod 6.1.0
What steps will reproduce the problem?
1. start sl4a
2. turn off/lose wireless connection
3. launch any 'hello world' script
What is the expected output? What do you see instead?
sl4a crashes immediately
What version of the product are you using? On what operating system?
latest sl4a release/android 2.2 (froyo)
Please provide any additional information below.
Snip of logcat (not sure it says much tho):
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Interpreter discovered: com.googlecode.bshforandroid
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Binary: /system/bin/dalvikvm
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Interpreter discovered: com.googlecode.bshforandroid
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Binary: /system/bin/dalvikvm
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Interpreter discovered: com.googlecode.jrubyforandroid
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Binary: /system/bin/dalvikvm
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Interpreter discovered: com.googlecode.jrubyforandroid
V/sl4a.InterpreterConfiguration$InterpreterListener:127( 2035):
Binary: /system/bin/dalvikvm
D/org.jraf.android.nolock.NoLockService( 656): onLowMemory
D/org.jraf.android.nolock.NoLockService( 656): handleStart
mKeyguardLock != null: reenabling keyguard
D/dalvikvm( 656): GC_EXPLICIT freed 60 objects / 3768 bytes in 103ms
D/dalvikvm( 1353): GC_EXPLICIT freed 6 objects / 216 bytes in 105ms
D/dalvikvm( 1401): GC_EXPLICIT freed 7 objects / 248 bytes in 151ms
W/ActivityManager( 190): Timeout executing service:
ServiceRecord{4406fbf8
com.googlecode.android_scripting/.activity.ScriptingLayerService}
I/ActivityManager( 190): Crashing app skipping ANR:
ProcessRecord{43f54748 2035:com.googlecode.android_scripting/10064}
Executing service
com.googlecode.android_scripting/.activity.ScriptingLayerService
D/dalvikvm( 435): GC_EXPLICIT freed 7 objects / 256 bytes in 116ms
D/dalvikvm( 978): GC_EXPLICIT freed 1039 objects / 47768 bytes in
162ms
I/Process ( 2035): Sending signal. PID: 2035 SIG: 9
I/ActivityManager( 190): Process com.googlecode.android_scripting
(pid 2035) has died.
W/ActivityManager( 190): Service crashed 2 times, stopping:
ServiceRecord{4406fbf8
com.googlecode.android_scripting/.activity.ScriptingLayerService}
```
Original issue reported on code.google.com by `guy.berl...@gmail.com` on 8 Jan 2011 at 2:28
|
defect
|
no wireless sudden death what device s are you experiencing the problem on htc dream what firmware version are you running on the device cyanogenmod what steps will reproduce the problem start turn off lose wireless connection launch any hello world script what is the expected output what do you see instead crashes immediately what version of the product are you using on what operating system latest release android froyo please provide any additional information below snip of logcat not sure it says much tho v interpreterconfiguration interpreterlistener interpreter discovered com googlecode bshforandroid v interpreterconfiguration interpreterlistener binary system bin dalvikvm v interpreterconfiguration interpreterlistener interpreter discovered com googlecode bshforandroid v interpreterconfiguration interpreterlistener binary system bin dalvikvm v interpreterconfiguration interpreterlistener interpreter discovered com googlecode jrubyforandroid v interpreterconfiguration interpreterlistener binary system bin dalvikvm v interpreterconfiguration interpreterlistener interpreter discovered com googlecode jrubyforandroid v interpreterconfiguration interpreterlistener binary system bin dalvikvm d org jraf android nolock nolockservice onlowmemory d org jraf android nolock nolockservice handlestart mkeyguardlock null reenabling keyguard d dalvikvm gc explicit freed objects bytes in d dalvikvm gc explicit freed objects bytes in d dalvikvm gc explicit freed objects bytes in w activitymanager timeout executing service servicerecord com googlecode android scripting activity scriptinglayerservice i activitymanager crashing app skipping anr processrecord com googlecode android scripting executing service com googlecode android scripting activity scriptinglayerservice d dalvikvm gc explicit freed objects bytes in d dalvikvm gc explicit freed objects bytes in i process sending signal pid sig i activitymanager process com googlecode android scripting pid has died w activitymanager service crashed times stopping servicerecord com googlecode android scripting activity scriptinglayerservice original issue reported on code google com by guy berl gmail com on jan at
| 1
|
18,910
| 24,848,829,807
|
IssuesEvent
|
2022-10-26 18:10:21
|
apache/arrow-rs
|
https://api.github.com/repos/apache/arrow-rs
|
closed
|
failed to pass test cases while enabling feature chrono-tz
|
bug development-process
|
```bash
cargo test --features chrono-tz
```
```bash
failures:
---- compute::kernels::cast::tests::test_cast_timestamp_to_string stdout ----
[arrow/src/compute/kernels/cast.rs:3852] &array = PrimitiveArray<Timestamp(Millisecond, None)>
[
1997-05-19T00:00:00.005,
2018-12-25T00:00:00.001,
null,
]
thread 'compute::kernels::cast::tests::test_cast_timestamp_to_string' panicked at 'assertion failed: `(left == right)`
left: `"1997-05-19 00:00:00.005"`,
right: `"1997-05-19 00:00:00.005 +00:00"`', arrow/src/compute/kernels/cast.rs:3856:9
---- compute::kernels::cast::tests::test_timestamp_cast_utf8 stdout ----
thread 'compute::kernels::cast::tests::test_timestamp_cast_utf8' panicked at 'assertion failed: `(left == right)`
left: `StringArray
[
"1970-01-01 20:30:00 +10:00",
null,
"1970-01-02 09:58:59 +10:00",
]`,
right: `StringArray
[
"1970-01-01 20:30:00",
null,
"1970-01-02 09:58:59",
]`', arrow/src/compute/kernels/cast.rs:5762:9
---- csv::writer::tests::test_export_csv_timestamps stdout ----
thread 'csv::writer::tests::test_export_csv_timestamps' panicked at 'assertion failed: `(left == right)`
left: `Some("c1,c2\n2019-04-18T20:54:47.378000000+10:00,2019-04-18T10:54:47.378000000+00:00\n2021-10-30T17:59:07.000000000+11:00,2021-10-30T06:59:07.000000000+00:00\n")`,
right: `Some("c1,c2\n2019-04-18T20:54:47.378000000+10:00,2019-04-18T10:54:47.378000000\n2021-10-30T17:59:07.000000000+11:00,2021-10-30T06:59:07.000000000\n")`', arrow/src/csv/writer.rs:652:9
---- csv::writer::tests::test_write_csv stdout ----
thread 'csv::writer::tests::test_write_csv' panicked at 'assertion failed: `(left == right)`
left: `"c1,c2,c3,c4,c5,c6,c7\nLorem ipsum dolor sit amet,123.564532,3,true,,00:20:34,cupcakes\nconsectetur adipiscing elit,,2,false,2019-04-18T10:54:47.378000000+00:00,06:51:20,cupcakes\nsed do eiusmod tempor,-556132.25,1,,2019-04-18T02:45:55.555000000+00:00,23:46:03,foo\nLorem ipsum dolor sit amet,123.564532,3,true,,00:20:34,cupcakes\nconsectetur adipiscing elit,,2,false,2019-04-18T10:54:47.378000000+00:00,06:51:20,cupcakes\nsed do eiusmod tempor,-556132.25,1,,2019-04-18T02:45:55.555000000+00:00,23:46:03,foo\n"`,
right: `"c1,c2,c3,c4,c5,c6,c7\nLorem ipsum dolor sit amet,123.564532,3,true,,00:20:34,cupcakes\nconsectetur adipiscing elit,,2,false,2019-04-18T10:54:47.378000000,06:51:20,cupcakes\nsed do eiusmod tempor,-556132.25,1,,2019-04-18T02:45:55.555000000,23:46:03,foo\nLorem ipsum dolor sit amet,123.564532,3,true,,00:20:34,cupcakes\nconsectetur adipiscing elit,,2,false,2019-04-18T10:54:47.378000000,06:51:20,cupcakes\nsed do eiusmod tempor,-556132.25,1,,2019-04-18T02:45:55.555000000,23:46:03,foo\n"`', arrow/src/csv/writer.rs:546:9
failures:
compute::kernels::cast::tests::test_cast_timestamp_to_string
compute::kernels::cast::tests::test_timestamp_cast_utf8
csv::writer::tests::test_export_csv_timestamps
csv::writer::tests::test_write_csv
test result: FAILED. 751 passed; 4 failed; 0 ignored; 0 measured; 0 filtered out; finished in 1.21s
```
|
1.0
|
failed to pass test cases while enabling feature chrono-tz - ```bash
cargo test --features chrono-tz
```
```bash
failures:
---- compute::kernels::cast::tests::test_cast_timestamp_to_string stdout ----
[arrow/src/compute/kernels/cast.rs:3852] &array = PrimitiveArray<Timestamp(Millisecond, None)>
[
1997-05-19T00:00:00.005,
2018-12-25T00:00:00.001,
null,
]
thread 'compute::kernels::cast::tests::test_cast_timestamp_to_string' panicked at 'assertion failed: `(left == right)`
left: `"1997-05-19 00:00:00.005"`,
right: `"1997-05-19 00:00:00.005 +00:00"`', arrow/src/compute/kernels/cast.rs:3856:9
---- compute::kernels::cast::tests::test_timestamp_cast_utf8 stdout ----
thread 'compute::kernels::cast::tests::test_timestamp_cast_utf8' panicked at 'assertion failed: `(left == right)`
left: `StringArray
[
"1970-01-01 20:30:00 +10:00",
null,
"1970-01-02 09:58:59 +10:00",
]`,
right: `StringArray
[
"1970-01-01 20:30:00",
null,
"1970-01-02 09:58:59",
]`', arrow/src/compute/kernels/cast.rs:5762:9
---- csv::writer::tests::test_export_csv_timestamps stdout ----
thread 'csv::writer::tests::test_export_csv_timestamps' panicked at 'assertion failed: `(left == right)`
left: `Some("c1,c2\n2019-04-18T20:54:47.378000000+10:00,2019-04-18T10:54:47.378000000+00:00\n2021-10-30T17:59:07.000000000+11:00,2021-10-30T06:59:07.000000000+00:00\n")`,
right: `Some("c1,c2\n2019-04-18T20:54:47.378000000+10:00,2019-04-18T10:54:47.378000000\n2021-10-30T17:59:07.000000000+11:00,2021-10-30T06:59:07.000000000\n")`', arrow/src/csv/writer.rs:652:9
---- csv::writer::tests::test_write_csv stdout ----
thread 'csv::writer::tests::test_write_csv' panicked at 'assertion failed: `(left == right)`
left: `"c1,c2,c3,c4,c5,c6,c7\nLorem ipsum dolor sit amet,123.564532,3,true,,00:20:34,cupcakes\nconsectetur adipiscing elit,,2,false,2019-04-18T10:54:47.378000000+00:00,06:51:20,cupcakes\nsed do eiusmod tempor,-556132.25,1,,2019-04-18T02:45:55.555000000+00:00,23:46:03,foo\nLorem ipsum dolor sit amet,123.564532,3,true,,00:20:34,cupcakes\nconsectetur adipiscing elit,,2,false,2019-04-18T10:54:47.378000000+00:00,06:51:20,cupcakes\nsed do eiusmod tempor,-556132.25,1,,2019-04-18T02:45:55.555000000+00:00,23:46:03,foo\n"`,
right: `"c1,c2,c3,c4,c5,c6,c7\nLorem ipsum dolor sit amet,123.564532,3,true,,00:20:34,cupcakes\nconsectetur adipiscing elit,,2,false,2019-04-18T10:54:47.378000000,06:51:20,cupcakes\nsed do eiusmod tempor,-556132.25,1,,2019-04-18T02:45:55.555000000,23:46:03,foo\nLorem ipsum dolor sit amet,123.564532,3,true,,00:20:34,cupcakes\nconsectetur adipiscing elit,,2,false,2019-04-18T10:54:47.378000000,06:51:20,cupcakes\nsed do eiusmod tempor,-556132.25,1,,2019-04-18T02:45:55.555000000,23:46:03,foo\n"`', arrow/src/csv/writer.rs:546:9
failures:
compute::kernels::cast::tests::test_cast_timestamp_to_string
compute::kernels::cast::tests::test_timestamp_cast_utf8
csv::writer::tests::test_export_csv_timestamps
csv::writer::tests::test_write_csv
test result: FAILED. 751 passed; 4 failed; 0 ignored; 0 measured; 0 filtered out; finished in 1.21s
```
|
non_defect
|
failed to pass test cases while enabling feature chrono tz bash cargo test features chrono tz bash failures compute kernels cast tests test cast timestamp to string stdout array primitivearray null thread compute kernels cast tests test cast timestamp to string panicked at assertion failed left right left right arrow src compute kernels cast rs compute kernels cast tests test timestamp cast stdout thread compute kernels cast tests test timestamp cast panicked at assertion failed left right left stringarray null right stringarray null arrow src compute kernels cast rs csv writer tests test export csv timestamps stdout thread csv writer tests test export csv timestamps panicked at assertion failed left right left some n right some n arrow src csv writer rs csv writer tests test write csv stdout thread csv writer tests test write csv panicked at assertion failed left right left nlorem ipsum dolor sit amet true cupcakes nconsectetur adipiscing elit false cupcakes nsed do eiusmod tempor foo nlorem ipsum dolor sit amet true cupcakes nconsectetur adipiscing elit false cupcakes nsed do eiusmod tempor foo n right nlorem ipsum dolor sit amet true cupcakes nconsectetur adipiscing elit false cupcakes nsed do eiusmod tempor foo nlorem ipsum dolor sit amet true cupcakes nconsectetur adipiscing elit false cupcakes nsed do eiusmod tempor foo n arrow src csv writer rs failures compute kernels cast tests test cast timestamp to string compute kernels cast tests test timestamp cast csv writer tests test export csv timestamps csv writer tests test write csv test result failed passed failed ignored measured filtered out finished in
| 0
|
7,491
| 2,610,390,384
|
IssuesEvent
|
2015-02-26 20:06:18
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
Hedgewars bug in graph
|
auto-migrated Priority-Low Type-Defect
|
```
What steps will reproduce the problem?
1. Create a new game
2. Win the match in a single play
What is the expected output? What do you see instead?
The graph is not displayed well
What version of the product are you using? On what operating system?
0.9.17
```
-----
Original issue reported on code.google.com by `costamag...@gmail.com` on 5 Mar 2012 at 9:09
Attachments:
* [Hedgewars.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-372/comment-0/Hedgewars.png)
|
1.0
|
Hedgewars bug in graph - ```
What steps will reproduce the problem?
1. Create a new game
2. Win the match in a single play
What is the expected output? What do you see instead?
The graph is not displayed well
What version of the product are you using? On what operating system?
0.9.17
```
-----
Original issue reported on code.google.com by `costamag...@gmail.com` on 5 Mar 2012 at 9:09
Attachments:
* [Hedgewars.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-372/comment-0/Hedgewars.png)
|
defect
|
hedgewars bug in graph what steps will reproduce the problem create a new game win the match in a single play what is the expected output what do you see instead the graph is not displayed well what version of the product are you using on what operating system original issue reported on code google com by costamag gmail com on mar at attachments
| 1
|
31,063
| 6,420,781,895
|
IssuesEvent
|
2017-08-09 01:36:15
|
junichi11/netbeans-github-issues-plugin
|
https://api.github.com/repos/junichi11/netbeans-github-issues-plugin
|
closed
|
Can't add a Milestone from within Netbeans
|
defect
|
While viewing an issue in Netbeans (v 8.2), I try to add a Milestone from within the issue. When I click OK from within the New Milestone dialog I get "Can't add a milestone". I'm also not seeing Milestones that I've created online in my GitHub repository being available in the Milestone dropdown in an issue's edit page within Netbeans.
|
1.0
|
Can't add a Milestone from within Netbeans - While viewing an issue in Netbeans (v 8.2), I try to add a Milestone from within the issue. When I click OK from within the New Milestone dialog I get "Can't add a milestone". I'm also not seeing Milestones that I've created online in my GitHub repository being available in the Milestone dropdown in an issue's edit page within Netbeans.
|
defect
|
can t add a milestone from within netbeans while viewing an issue in netbeans v i try to add a milestone from within the issue when i click ok from within the new milestone dialog i get can t add a milestone i m also not seeing milestones that i ve created online in my github repository being available in the milestone dropdown in an issue s edit page within netbeans
| 1
|
43,045
| 23,092,251,778
|
IssuesEvent
|
2022-07-26 16:07:29
|
rapidsai/cudf
|
https://api.github.com/repos/rapidsai/cudf
|
closed
|
[FEA] Replace `cuco::static_multimap` by `cuco::static_map` in semi-anti-join
|
feature request 0 - Backlog libcudf Performance helps: Spark non-breaking
|
The implementation of semi-anti-join was refactored in #11100. One of the changes was to use `cuco::static_multimap`, which was later discovered that it has performance issue when the input tables have too many duplicate rows (https://github.com/rapidsai/cudf/issues/11299).
We should use `cuco::static_map` to avoid the performance issue. However, this is not just a simple change in implementation but needs a new FEA from cuco which adds `pair_contains` into `static_map`: https://github.com/NVIDIA/cuCollections/pull/191.
|
True
|
[FEA] Replace `cuco::static_multimap` by `cuco::static_map` in semi-anti-join - The implementation of semi-anti-join was refactored in #11100. One of the changes was to use `cuco::static_multimap`, which was later discovered that it has performance issue when the input tables have too many duplicate rows (https://github.com/rapidsai/cudf/issues/11299).
We should use `cuco::static_map` to avoid the performance issue. However, this is not just a simple change in implementation but needs a new FEA from cuco which adds `pair_contains` into `static_map`: https://github.com/NVIDIA/cuCollections/pull/191.
|
non_defect
|
replace cuco static multimap by cuco static map in semi anti join the implementation of semi anti join was refactored in one of the changes was to use cuco static multimap which was later discovered that it has performance issue when the input tables have too many duplicate rows we should use cuco static map to avoid the performance issue however this is not just a simple change in implementation but needs a new fea from cuco which adds pair contains into static map
| 0
|
21,475
| 3,511,949,080
|
IssuesEvent
|
2016-01-10 17:32:48
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
A significant amount of CPU is spent in String.indexOf() because of identifier escaping
|
C: Functionality P: Medium T: Defect
|
When running benchmarks with JMH, a significant amount of time is spent in `String.indexOf()`:
```
15.7% 30.9% java.lang.String.indexOf
4.9% 9.6% org.apache.log4j.Category.getEffectiveLevel
4.8% 9.5% org.jooq.impl.AbstractBindContext.bindInternal
2.7% 5.2% org.jooq.impl.AbstractContext.visit0
2.2% 4.4% org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.setValue
2.0% 4.0% org.jooq.impl.TableImpl.accept0
1.7% 3.3% java.lang.ThreadLocal$ThreadLocalMap.cleanSomeSlots
1.7% 3.3% org.h2.jdbc.JdbcConnection.prepareCommand
1.6% 3.1% org.h2.engine.Session.isReconnectNeeded
1.3% 2.5% org.jooq.impl.SchemaImpl.accept
12.3% 24.2% <other>
```
This effort can be traced back to `DefaultRenderContext.literal(String)`, which escapes all identifiers (for backticks ` ` `, quotes `"`, or brackets `]`), regardless if they are really contained in the identifier string:
```java
sql(StringUtils.replace(literal, quotes[QUOTE_END_DELIMITER][0], quotes[QUOTE_END_DELIMITER_ESCAPED][0]));
```
A quick `indexOf(char)` check prior to the string replace would be much faster:
```java
if (literal.indexOf(quotes[QUOTE_END_DELIMITER][0].charAt(0)) > -1)
sql(StringUtils.replace(literal, quotes[QUOTE_END_DELIMITER][0], quotes[QUOTE_END_DELIMITER_ESCAPED][0]));
else
sql(literal);
```
The above improvement has the following effect. Before:
```
ComparingWithJDBCBenchmark.testQueryWithFewResultsJOOQDSLBuildQueryAfresh thrpt 50 72833.733 ▒ 2095.076 ops/s
ComparingWithJDBCBenchmark.testQueryWithFewResultsJOOQDSLReuseQuery thrpt 50 81241.158 ▒ 3090.201 ops/s
```
After:
```
ComparingWithJDBCBenchmark.testQueryWithFewResultsJOOQDSLBuildQueryAfresh thrpt 50 77750.571 ▒ 2050.013 ops/s
ComparingWithJDBCBenchmark.testQueryWithFewResultsJOOQDSLReuseQuery thrpt 50 87117.076 ▒ 2700.036 ops/s
```
|
1.0
|
A significant amount of CPU is spent in String.indexOf() because of identifier escaping - When running benchmarks with JMH, a significant amount of time is spent in `String.indexOf()`:
```
15.7% 30.9% java.lang.String.indexOf
4.9% 9.6% org.apache.log4j.Category.getEffectiveLevel
4.8% 9.5% org.jooq.impl.AbstractBindContext.bindInternal
2.7% 5.2% org.jooq.impl.AbstractContext.visit0
2.2% 4.4% org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.setValue
2.0% 4.0% org.jooq.impl.TableImpl.accept0
1.7% 3.3% java.lang.ThreadLocal$ThreadLocalMap.cleanSomeSlots
1.7% 3.3% org.h2.jdbc.JdbcConnection.prepareCommand
1.6% 3.1% org.h2.engine.Session.isReconnectNeeded
1.3% 2.5% org.jooq.impl.SchemaImpl.accept
12.3% 24.2% <other>
```
This effort can be traced back to `DefaultRenderContext.literal(String)`, which escapes all identifiers (for backticks ` ` `, quotes `"`, or brackets `]`), regardless if they are really contained in the identifier string:
```java
sql(StringUtils.replace(literal, quotes[QUOTE_END_DELIMITER][0], quotes[QUOTE_END_DELIMITER_ESCAPED][0]));
```
A quick `indexOf(char)` check prior to the string replace would be much faster:
```java
if (literal.indexOf(quotes[QUOTE_END_DELIMITER][0].charAt(0)) > -1)
sql(StringUtils.replace(literal, quotes[QUOTE_END_DELIMITER][0], quotes[QUOTE_END_DELIMITER_ESCAPED][0]));
else
sql(literal);
```
The above improvement has the following effect. Before:
```
ComparingWithJDBCBenchmark.testQueryWithFewResultsJOOQDSLBuildQueryAfresh thrpt 50 72833.733 ▒ 2095.076 ops/s
ComparingWithJDBCBenchmark.testQueryWithFewResultsJOOQDSLReuseQuery thrpt 50 81241.158 ▒ 3090.201 ops/s
```
After:
```
ComparingWithJDBCBenchmark.testQueryWithFewResultsJOOQDSLBuildQueryAfresh thrpt 50 77750.571 ▒ 2050.013 ops/s
ComparingWithJDBCBenchmark.testQueryWithFewResultsJOOQDSLReuseQuery thrpt 50 87117.076 ▒ 2700.036 ops/s
```
|
defect
|
a significant amount of cpu is spent in string indexof because of identifier escaping when running benchmarks with jmh a significant amount of time is spent in string indexof java lang string indexof org apache category geteffectivelevel org jooq impl abstractbindcontext bindinternal org jooq impl abstractcontext org jooq impl cursorimpl cursoriterator cursorrecordinitialiser setvalue org jooq impl tableimpl java lang threadlocal threadlocalmap cleansomeslots org jdbc jdbcconnection preparecommand org engine session isreconnectneeded org jooq impl schemaimpl accept this effort can be traced back to defaultrendercontext literal string which escapes all identifiers for backticks quotes or brackets regardless if they are really contained in the identifier string java sql stringutils replace literal quotes quotes a quick indexof char check prior to the string replace would be much faster java if literal indexof quotes charat sql stringutils replace literal quotes quotes else sql literal the above improvement has the following effect before comparingwithjdbcbenchmark testquerywithfewresultsjooqdslbuildqueryafresh thrpt ▒ ops s comparingwithjdbcbenchmark testquerywithfewresultsjooqdslreusequery thrpt ▒ ops s after comparingwithjdbcbenchmark testquerywithfewresultsjooqdslbuildqueryafresh thrpt ▒ ops s comparingwithjdbcbenchmark testquerywithfewresultsjooqdslreusequery thrpt ▒ ops s
| 1
|
82,655
| 15,679,655,369
|
IssuesEvent
|
2021-03-25 01:01:47
|
snowdensb/Leo
|
https://api.github.com/repos/snowdensb/Leo
|
opened
|
CVE-2021-21349 (Medium) detected in xstream-1.4.8.jar
|
security vulnerability
|
## CVE-2021-21349 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.8.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to dependency file: Leo/core/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.8/xstream-1.4.8.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.8/xstream-1.4.8.jar</p>
<p>
Dependency Hierarchy:
- uimaj-as-core-2.9.0.jar (Root Library)
- :x: **xstream-1.4.8.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21349>CVE-2021-21349</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-f6hm-88x3-mfjv">https://github.com/x-stream/xstream/security/advisories/GHSA-f6hm-88x3-mfjv</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.8","packageFilePaths":["/core/pom.xml","/service/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.uima:uimaj-as-core:2.9.0;com.thoughtworks.xstream:xstream:1.4.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21349","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21349","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Changed","C":"None","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-21349 (Medium) detected in xstream-1.4.8.jar - ## CVE-2021-21349 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.8.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to dependency file: Leo/core/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.8/xstream-1.4.8.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.8/xstream-1.4.8.jar</p>
<p>
Dependency Hierarchy:
- uimaj-as-core-2.9.0.jar (Root Library)
- :x: **xstream-1.4.8.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21349>CVE-2021-21349</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-f6hm-88x3-mfjv">https://github.com/x-stream/xstream/security/advisories/GHSA-f6hm-88x3-mfjv</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.8","packageFilePaths":["/core/pom.xml","/service/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.uima:uimaj-as-core:2.9.0;com.thoughtworks.xstream:xstream:1.4.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21349","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21349","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Changed","C":"None","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve medium detected in xstream jar cve medium severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back path to dependency file leo core pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy uimaj as core jar root library x xstream jar vulnerable library found in base branch master vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope changed impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache uima uimaj as core com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability which may allow a remote attacker to request data from internal resources that are not publicly available only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream security framework with a whitelist limited to the minimal required types if you rely on xstream default blacklist of the security framework you will have to use at least version vulnerabilityurl
| 0
|
276,943
| 21,006,439,558
|
IssuesEvent
|
2022-03-29 23:15:44
|
NASA-IMPACT/sddo
|
https://api.github.com/repos/NASA-IMPACT/sddo
|
reopened
|
Complete definitions for all SDDO terms
|
documentation
|
Many of the classes and properties in SDDO have yet to have definitions....
|
1.0
|
Complete definitions for all SDDO terms - Many of the classes and properties in SDDO have yet to have definitions....
|
non_defect
|
complete definitions for all sddo terms many of the classes and properties in sddo have yet to have definitions
| 0
|
257,815
| 19,532,334,288
|
IssuesEvent
|
2021-12-30 19:34:37
|
InfiniTimeOrg/InfiniTime
|
https://api.github.com/repos/InfiniTimeOrg/InfiniTime
|
closed
|
My beginner experience / improving onboarding process
|
documentation enhancement help wanted question/discussion
|
###
- [X] I searched for similar feature request and found none was relevant.
### Pitch us your idea!
Improving documentation to be more user-centric
### Description
Hi,
I love this watch and this is my way of documenting a typical first time user experience.
I'm sharing this because once you get into the project, you quickly forget beginner struggles.
During setup I made an effort to document everything I thought to improve the docs.
This is not intended to sound mean, the community is amazingly helpful and the people are really kind.
I just want to improve the onboarding process by providing typical beginner thoughts.
# Expectations
When I bought this watch, my expecation was to have a watch with basic functionalities which I can change some source code on.
Since the manual recommends updating the firmware, I went looking on how to do that exactly.
* What is the right OS for me?
* What device do I need for flashing? What app?
* Where can I download the latest firmware?
A Google search lead me to the GitHub README, which is an appropriate source of info for a hacker watch.
# README
As I was being greeted by a gigantic logo, I felt overwhelmed.
The README currently has lots of links, many of which contain similar information and they all focus on development and debugging straight away.
I got stuck in analysis paralysis because I was presented too many options and alternatives.
Since the sealed version aims at casual techies, I recommend a more user-centric README.
Developers will dig into the docs anyway.
Onboarding could be split into multiple steps:
1. This is a community watch, if you need any help at all contact us here
1. How to get your PineTime working for the first time
1. This is the latest stable firmware in case you are just a casual user
1. Here is how to flash it on Android, iOS and Linux
1. FAQ when things don't work out
The README is a place to get novices into the project.
It might even be a place to promote the watch to users on the verge of buying it.
Development documentation should be a level deeper, I always think of the README as a pamphlet.
# Development
If you get the users to a point where they have a working watch with the latest firmware, they might be interested in developent.
I still recommend keeping the friction to get into development as low as possible (don't install anything locally, get started using GitHub Actions).
Any step that might sound trivial to you but "install SDK on Linux" might be a showstopper for a beginner.
Possible steps for development onboarding:
1. Differences between sealed vs development kit
1. How to use GitHub actions to fork and get your first custom .zip loaded onto your watch (reduce effort to get going by not having to install anything)
1. How to introduct a simple change (e.g. font color) and let GitHub recompile
1. How to setup a local environment using the most simple method (Docker devcontainers probably)
1. How to debug in case anything breaks (development kit only)
1. How to "unbrick" your device if you flashed something really bad
# Problems I had
Not intended to sound mean, these can all be cleaned up to make it easier for people to join development.
## As a casual user
* README is quite long, I was only looking for a getting started but read everything in case it's important to not miss anything.
* Took long to find the relevant information to download and flash firmware. Lost in development docs.
* No idea which app is the best to flash new firmware -> went for Gadgedbridge worked fine. Recommended.
* Watch was not findable using BLE (stock firmware 1.0.0). Only restart fixed that for me - no documented.
* Docs say to use Gadgedbridge to flash but there is no option in the app. What you acutally do is open the zip in your phones file browser and select Gadgedbridge to open it. Confusing.
* Didn't tell me to download and flash a zip-file. Talks of binary made me expect a .bin file or something similar.
* Some information was almost identical and I didn't know which one to use
* Difference between sealed and development kid barely stated - I was very confused about the debugging part as that is not relevant to me.
## As a developer
* All of the above still apply - any dev starts out as a user, not the other way around
* The docs recommend to flash the firmware without any bootloader for debugging purposes - I have a sealed version, so no debugging available and if I flash a firmware that bricks my watch, I can't do anything about it.
* The docs tell you to install all dependencies, SDKs, set up debugging etc. manually only to mention in the end that there is a all-in-one solution if you use devcontainers
* I have a sealed watch and can't really debug anything it turns out. Guess I'll wait for more dev kits to become available.
* No emulator available, I have to have a development kit (don't know if that is even possible with embedded programming)
# Summary
I LOVE THIS COMMUNITY!
Everyone was welcoming and helpful when I had questions and my custom firmware has an uptime of 7 days already.
PineTime and InfiniTime get recomended to everyone I see - I love it.
But I also realize that if I wasn't a developer, I would have never flashed a custom version of InfiniTime on my watch.
Acutally I was very close to just leaving the default firmware on it after being confused in the beginning.
Hopefully my ramblings are of use to the team if the docs get an overhaul in the future, I really want
more people to be able to experience this amazing piece of hardware.
Thank you for your time and greetings from Germany
Tom
|
1.0
|
My beginner experience / improving onboarding process - ###
- [X] I searched for similar feature request and found none was relevant.
### Pitch us your idea!
Improving documentation to be more user-centric
### Description
Hi,
I love this watch and this is my way of documenting a typical first time user experience.
I'm sharing this because once you get into the project, you quickly forget beginner struggles.
During setup I made an effort to document everything I thought to improve the docs.
This is not intended to sound mean, the community is amazingly helpful and the people are really kind.
I just want to improve the onboarding process by providing typical beginner thoughts.
# Expectations
When I bought this watch, my expecation was to have a watch with basic functionalities which I can change some source code on.
Since the manual recommends updating the firmware, I went looking on how to do that exactly.
* What is the right OS for me?
* What device do I need for flashing? What app?
* Where can I download the latest firmware?
A Google search lead me to the GitHub README, which is an appropriate source of info for a hacker watch.
# README
As I was being greeted by a gigantic logo, I felt overwhelmed.
The README currently has lots of links, many of which contain similar information and they all focus on development and debugging straight away.
I got stuck in analysis paralysis because I was presented too many options and alternatives.
Since the sealed version aims at casual techies, I recommend a more user-centric README.
Developers will dig into the docs anyway.
Onboarding could be split into multiple steps:
1. This is a community watch, if you need any help at all contact us here
1. How to get your PineTime working for the first time
1. This is the latest stable firmware in case you are just a casual user
1. Here is how to flash it on Android, iOS and Linux
1. FAQ when things don't work out
The README is a place to get novices into the project.
It might even be a place to promote the watch to users on the verge of buying it.
Development documentation should be a level deeper, I always think of the README as a pamphlet.
# Development
If you get the users to a point where they have a working watch with the latest firmware, they might be interested in developent.
I still recommend keeping the friction to get into development as low as possible (don't install anything locally, get started using GitHub Actions).
Any step that might sound trivial to you but "install SDK on Linux" might be a showstopper for a beginner.
Possible steps for development onboarding:
1. Differences between sealed vs development kit
1. How to use GitHub actions to fork and get your first custom .zip loaded onto your watch (reduce effort to get going by not having to install anything)
1. How to introduct a simple change (e.g. font color) and let GitHub recompile
1. How to setup a local environment using the most simple method (Docker devcontainers probably)
1. How to debug in case anything breaks (development kit only)
1. How to "unbrick" your device if you flashed something really bad
# Problems I had
Not intended to sound mean, these can all be cleaned up to make it easier for people to join development.
## As a casual user
* README is quite long, I was only looking for a getting started but read everything in case it's important to not miss anything.
* Took long to find the relevant information to download and flash firmware. Lost in development docs.
* No idea which app is the best to flash new firmware -> went for Gadgedbridge worked fine. Recommended.
* Watch was not findable using BLE (stock firmware 1.0.0). Only restart fixed that for me - no documented.
* Docs say to use Gadgedbridge to flash but there is no option in the app. What you acutally do is open the zip in your phones file browser and select Gadgedbridge to open it. Confusing.
* Didn't tell me to download and flash a zip-file. Talks of binary made me expect a .bin file or something similar.
* Some information was almost identical and I didn't know which one to use
* Difference between sealed and development kid barely stated - I was very confused about the debugging part as that is not relevant to me.
## As a developer
* All of the above still apply - any dev starts out as a user, not the other way around
* The docs recommend to flash the firmware without any bootloader for debugging purposes - I have a sealed version, so no debugging available and if I flash a firmware that bricks my watch, I can't do anything about it.
* The docs tell you to install all dependencies, SDKs, set up debugging etc. manually only to mention in the end that there is a all-in-one solution if you use devcontainers
* I have a sealed watch and can't really debug anything it turns out. Guess I'll wait for more dev kits to become available.
* No emulator available, I have to have a development kit (don't know if that is even possible with embedded programming)
# Summary
I LOVE THIS COMMUNITY!
Everyone was welcoming and helpful when I had questions and my custom firmware has an uptime of 7 days already.
PineTime and InfiniTime get recomended to everyone I see - I love it.
But I also realize that if I wasn't a developer, I would have never flashed a custom version of InfiniTime on my watch.
Acutally I was very close to just leaving the default firmware on it after being confused in the beginning.
Hopefully my ramblings are of use to the team if the docs get an overhaul in the future, I really want
more people to be able to experience this amazing piece of hardware.
Thank you for your time and greetings from Germany
Tom
|
non_defect
|
my beginner experience improving onboarding process i searched for similar feature request and found none was relevant pitch us your idea improving documentation to be more user centric description hi i love this watch and this is my way of documenting a typical first time user experience i m sharing this because once you get into the project you quickly forget beginner struggles during setup i made an effort to document everything i thought to improve the docs this is not intended to sound mean the community is amazingly helpful and the people are really kind i just want to improve the onboarding process by providing typical beginner thoughts expectations when i bought this watch my expecation was to have a watch with basic functionalities which i can change some source code on since the manual recommends updating the firmware i went looking on how to do that exactly what is the right os for me what device do i need for flashing what app where can i download the latest firmware a google search lead me to the github readme which is an appropriate source of info for a hacker watch readme as i was being greeted by a gigantic logo i felt overwhelmed the readme currently has lots of links many of which contain similar information and they all focus on development and debugging straight away i got stuck in analysis paralysis because i was presented too many options and alternatives since the sealed version aims at casual techies i recommend a more user centric readme developers will dig into the docs anyway onboarding could be split into multiple steps this is a community watch if you need any help at all contact us here how to get your pinetime working for the first time this is the latest stable firmware in case you are just a casual user here is how to flash it on android ios and linux faq when things don t work out the readme is a place to get novices into the project it might even be a place to promote the watch to users on the verge of buying it development documentation should be a level deeper i always think of the readme as a pamphlet development if you get the users to a point where they have a working watch with the latest firmware they might be interested in developent i still recommend keeping the friction to get into development as low as possible don t install anything locally get started using github actions any step that might sound trivial to you but install sdk on linux might be a showstopper for a beginner possible steps for development onboarding differences between sealed vs development kit how to use github actions to fork and get your first custom zip loaded onto your watch reduce effort to get going by not having to install anything how to introduct a simple change e g font color and let github recompile how to setup a local environment using the most simple method docker devcontainers probably how to debug in case anything breaks development kit only how to unbrick your device if you flashed something really bad problems i had not intended to sound mean these can all be cleaned up to make it easier for people to join development as a casual user readme is quite long i was only looking for a getting started but read everything in case it s important to not miss anything took long to find the relevant information to download and flash firmware lost in development docs no idea which app is the best to flash new firmware went for gadgedbridge worked fine recommended watch was not findable using ble stock firmware only restart fixed that for me no documented docs say to use gadgedbridge to flash but there is no option in the app what you acutally do is open the zip in your phones file browser and select gadgedbridge to open it confusing didn t tell me to download and flash a zip file talks of binary made me expect a bin file or something similar some information was almost identical and i didn t know which one to use difference between sealed and development kid barely stated i was very confused about the debugging part as that is not relevant to me as a developer all of the above still apply any dev starts out as a user not the other way around the docs recommend to flash the firmware without any bootloader for debugging purposes i have a sealed version so no debugging available and if i flash a firmware that bricks my watch i can t do anything about it the docs tell you to install all dependencies sdks set up debugging etc manually only to mention in the end that there is a all in one solution if you use devcontainers i have a sealed watch and can t really debug anything it turns out guess i ll wait for more dev kits to become available no emulator available i have to have a development kit don t know if that is even possible with embedded programming summary i love this community everyone was welcoming and helpful when i had questions and my custom firmware has an uptime of days already pinetime and infinitime get recomended to everyone i see i love it but i also realize that if i wasn t a developer i would have never flashed a custom version of infinitime on my watch acutally i was very close to just leaving the default firmware on it after being confused in the beginning hopefully my ramblings are of use to the team if the docs get an overhaul in the future i really want more people to be able to experience this amazing piece of hardware thank you for your time and greetings from germany tom
| 0
|
346,763
| 10,419,566,664
|
IssuesEvent
|
2019-09-15 17:30:11
|
HW-PlayersPatch/Development
|
https://api.github.com/repos/HW-PlayersPatch/Development
|
opened
|
EMP Friendly Fire and vs Frigs
|
MP priority: medium status: in discussion
|
## hw2c
EMP Damage: 35 / 105 (per scout/squad)
### EMP Shields:
Scouts 0
Fighters/Corvs 75
Frigs 310
Friendly Fire 0%
## Current State
EMP Damage: 20 / 60 (per scout/squad)
### EMP Shields:
Scouts 100
hw2 Fighters/Corvs 75
hw1 Fighters 30
hw1 Corvs 45
Frigs 310
Friendly Fire 50%
When 4 scouts try to emp the same target, they emp each other...
Also mol impossible to EMP frigs
## Suggested Changes
EMP Damage: 35 / 105 (per scout/squad)
### EMP Shields:
Scouts 200
hw2 Fighters/Corvs 110
hw1 Fighters 70
hw1 Corvs 70
Frigs 310
Friendly Fire 45%
When 5 scouts try to emp the same target, they emp each other.
3 scouts EMP a frig like hw2c
|
1.0
|
EMP Friendly Fire and vs Frigs - ## hw2c
EMP Damage: 35 / 105 (per scout/squad)
### EMP Shields:
Scouts 0
Fighters/Corvs 75
Frigs 310
Friendly Fire 0%
## Current State
EMP Damage: 20 / 60 (per scout/squad)
### EMP Shields:
Scouts 100
hw2 Fighters/Corvs 75
hw1 Fighters 30
hw1 Corvs 45
Frigs 310
Friendly Fire 50%
When 4 scouts try to emp the same target, they emp each other...
Also mol impossible to EMP frigs
## Suggested Changes
EMP Damage: 35 / 105 (per scout/squad)
### EMP Shields:
Scouts 200
hw2 Fighters/Corvs 110
hw1 Fighters 70
hw1 Corvs 70
Frigs 310
Friendly Fire 45%
When 5 scouts try to emp the same target, they emp each other.
3 scouts EMP a frig like hw2c
|
non_defect
|
emp friendly fire and vs frigs emp damage per scout squad emp shields scouts fighters corvs frigs friendly fire current state emp damage per scout squad emp shields scouts fighters corvs fighters corvs frigs friendly fire when scouts try to emp the same target they emp each other also mol impossible to emp frigs suggested changes emp damage per scout squad emp shields scouts fighters corvs fighters corvs frigs friendly fire when scouts try to emp the same target they emp each other scouts emp a frig like
| 0
|
94,594
| 19,562,842,303
|
IssuesEvent
|
2022-01-03 18:46:09
|
Onelinerhub/onelinerhub
|
https://api.github.com/repos/Onelinerhub/onelinerhub
|
closed
|
Short solution needed: "What's Redis default port" (redis)
|
help wanted good first issue code redis
|
Please help us write most modern and shortest code solution for this issue:
**What's Redis default port** (technology: [redis](https://onelinerhub.com/redis))
### Fast way
Just write the code solution in the comments.
### Prefered way
1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox).
2. Don't forget to use comments to make solution explained.
3. Link to this issue in comments of pull request.
|
1.0
|
Short solution needed: "What's Redis default port" (redis) - Please help us write most modern and shortest code solution for this issue:
**What's Redis default port** (technology: [redis](https://onelinerhub.com/redis))
### Fast way
Just write the code solution in the comments.
### Prefered way
1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox).
2. Don't forget to use comments to make solution explained.
3. Link to this issue in comments of pull request.
|
non_defect
|
short solution needed what s redis default port redis please help us write most modern and shortest code solution for this issue what s redis default port technology fast way just write the code solution in the comments prefered way create pull request with a new code file inside don t forget to use comments to make solution explained link to this issue in comments of pull request
| 0
|
261,133
| 27,785,299,954
|
IssuesEvent
|
2023-03-17 02:18:15
|
SebastianDarie/portfolio
|
https://api.github.com/repos/SebastianDarie/portfolio
|
opened
|
CVE-2023-28155 (Medium) detected in request-2.88.2.tgz
|
Mend: dependency security vulnerability
|
## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- astro-0.18.8.tgz (Root Library)
- snowpack-3.8.3.tgz
- arborist-2.8.0.tgz
- run-script-1.8.5.tgz
- node-gyp-7.1.2.tgz
- :x: **request-2.88.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SebastianDarie/portfolio/commit/53f136bc9bef9660a0158912544563c40ce463ba">53f136bc9bef9660a0158912544563c40ce463ba</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-28155 (Medium) detected in request-2.88.2.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- astro-0.18.8.tgz (Root Library)
- snowpack-3.8.3.tgz
- arborist-2.8.0.tgz
- run-script-1.8.5.tgz
- node-gyp-7.1.2.tgz
- :x: **request-2.88.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SebastianDarie/portfolio/commit/53f136bc9bef9660a0158912544563c40ce463ba">53f136bc9bef9660a0158912544563c40ce463ba</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in request tgz cve medium severity vulnerability vulnerable library request tgz simplified http request client library home page a href path to dependency file package json path to vulnerable library node modules request package json dependency hierarchy astro tgz root library snowpack tgz arborist tgz run script tgz node gyp tgz x request tgz vulnerable library found in head commit a href found in base branch main vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
| 0
|
1,327
| 2,539,163,788
|
IssuesEvent
|
2015-01-27 13:30:40
|
richtermondt/inithub-web
|
https://api.github.com/repos/richtermondt/inithub-web
|
closed
|
Clean project checkout, configuration and startup
|
test
|
take detailed notes and update installation notes when complete
|
1.0
|
Clean project checkout, configuration and startup - take detailed notes and update installation notes when complete
|
non_defect
|
clean project checkout configuration and startup take detailed notes and update installation notes when complete
| 0
|
90,315
| 10,678,340,727
|
IssuesEvent
|
2019-10-21 17:04:51
|
Ishan-Gunaratne/advance-javascript
|
https://api.github.com/repos/Ishan-Gunaratne/advance-javascript
|
closed
|
Update README.md
|
documentation good first issue
|
Update the READEME.md to show all the sections we have covered so far.
There is only one section now. (**variables**)
|
1.0
|
Update README.md - Update the READEME.md to show all the sections we have covered so far.
There is only one section now. (**variables**)
|
non_defect
|
update readme md update the reademe md to show all the sections we have covered so far there is only one section now variables
| 0
|
49,821
| 13,187,277,159
|
IssuesEvent
|
2020-08-13 02:54:17
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
dataio-pyshovel crashes upon missing key (Trac #2138)
|
Incomplete Migration Migrated from Trac combo core defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2138">https://code.icecube.wisc.edu/ticket/2138</a>, reported by peller and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2018-02-16T20:51:30",
"description": "when skipping through frames in an i3 file while having a key opened, and the key doesn't exist in the next frame it crashes. Exception should instead rather be caught and displayed that this key does not exist for the frame\n\n\n{{{\nFile \"/cvmfs/icecube.opensciencegrid.org/py2-v2/Ubuntu_16_x86_64/metaprojects/combo/trunk/lib/icecube/dataio/fileadaptor.py\", line 41, in __getitem__\n return self.frame.__getitem__(key)\nKeyError: 'OnlineL2_SPE2itFit'\n}}}\n\n",
"reporter": "peller",
"cc": "",
"resolution": "wontfix",
"_ts": "1518814290487856",
"component": "combo core",
"summary": "dataio-pyshovel crashes upon missing key",
"priority": "normal",
"keywords": "dataio",
"time": "2018-02-16T20:49:46",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
dataio-pyshovel crashes upon missing key (Trac #2138) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2138">https://code.icecube.wisc.edu/ticket/2138</a>, reported by peller and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2018-02-16T20:51:30",
"description": "when skipping through frames in an i3 file while having a key opened, and the key doesn't exist in the next frame it crashes. Exception should instead rather be caught and displayed that this key does not exist for the frame\n\n\n{{{\nFile \"/cvmfs/icecube.opensciencegrid.org/py2-v2/Ubuntu_16_x86_64/metaprojects/combo/trunk/lib/icecube/dataio/fileadaptor.py\", line 41, in __getitem__\n return self.frame.__getitem__(key)\nKeyError: 'OnlineL2_SPE2itFit'\n}}}\n\n",
"reporter": "peller",
"cc": "",
"resolution": "wontfix",
"_ts": "1518814290487856",
"component": "combo core",
"summary": "dataio-pyshovel crashes upon missing key",
"priority": "normal",
"keywords": "dataio",
"time": "2018-02-16T20:49:46",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
defect
|
dataio pyshovel crashes upon missing key trac migrated from json status closed changetime description when skipping through frames in an file while having a key opened and the key doesn t exist in the next frame it crashes exception should instead rather be caught and displayed that this key does not exist for the frame n n n nfile cvmfs icecube opensciencegrid org ubuntu metaprojects combo trunk lib icecube dataio fileadaptor py line in getitem n return self frame getitem key nkeyerror n n n reporter peller cc resolution wontfix ts component combo core summary dataio pyshovel crashes upon missing key priority normal keywords dataio time milestone owner type defect
| 1
|
179,998
| 30,343,612,381
|
IssuesEvent
|
2023-07-11 14:09:59
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
Suggestion: Remove "nameof()" from references list
|
Area-IDE Concept-Continuous Improvement Need Design Review IDE-Navigation
|
_This issue has been moved from [a ticket on Developer Community](https://developercommunity2.visualstudio.com/t/Suggestion:-Remove-nameof-from-refer/1337827)._
---

When a references list of a method is opened, it also shows references where the method is used in a call to nameof(). Most of the times these do not interest me. I'm only concerned about the invocations of the method. So it would be a nice-to-have if the references to nameof() could be filtered out.
---
### Original Comments
#### Feedback Bot on 2/14/2021, 10:04 PM:
Thank you for taking the time to provide your suggestion. We will do some preliminary checks to make sure we can proceed further. You will hear from us in about a week on our next steps.
#### Feedback Bot on 2/24/2021, 10:46 AM:
<p>Your suggestion has been queued up for prioritization. Feature suggestions are prioritized based on the value to our broader developer community and the product roadmap. We may not be able to pursue this one immediately, but we will continue to monitor it for community input</p>
|
1.0
|
Suggestion: Remove "nameof()" from references list - _This issue has been moved from [a ticket on Developer Community](https://developercommunity2.visualstudio.com/t/Suggestion:-Remove-nameof-from-refer/1337827)._
---

When a references list of a method is opened, it also shows references where the method is used in a call to nameof(). Most of the times these do not interest me. I'm only concerned about the invocations of the method. So it would be a nice-to-have if the references to nameof() could be filtered out.
---
### Original Comments
#### Feedback Bot on 2/14/2021, 10:04 PM:
Thank you for taking the time to provide your suggestion. We will do some preliminary checks to make sure we can proceed further. You will hear from us in about a week on our next steps.
#### Feedback Bot on 2/24/2021, 10:46 AM:
<p>Your suggestion has been queued up for prioritization. Feature suggestions are prioritized based on the value to our broader developer community and the product roadmap. We may not be able to pursue this one immediately, but we will continue to monitor it for community input</p>
|
non_defect
|
suggestion remove nameof from references list this issue has been moved from when a references list of a method is opened it also shows references where the method is used in a call to nameof most of the times these do not interest me i m only concerned about the invocations of the method so it would be a nice to have if the references to nameof could be filtered out original comments feedback bot on pm thank you for taking the time to provide your suggestion we will do some preliminary checks to make sure we can proceed further you will hear from us in about a week on our next steps feedback bot on am your suggestion has been queued up for prioritization feature suggestions are prioritized based on the value to our broader developer community and the product roadmap we may not be able to pursue this one immediately but we will continue to monitor it for community input
| 0
|
31,526
| 6,544,872,173
|
IssuesEvent
|
2017-09-03 22:19:11
|
BugBuster1701/banner
|
https://api.github.com/repos/BugBuster1701/banner
|
closed
|
Externe Banner werden nicht angezeigt
|
Defect
|
Es fehlt durch die Picture Anpassung die `//fake the Picture::create` Zeilen
Bei getSingleBanner und getSingleBannerFirst, bei Multi passt es.
|
1.0
|
Externe Banner werden nicht angezeigt - Es fehlt durch die Picture Anpassung die `//fake the Picture::create` Zeilen
Bei getSingleBanner und getSingleBannerFirst, bei Multi passt es.
|
defect
|
externe banner werden nicht angezeigt es fehlt durch die picture anpassung die fake the picture create zeilen bei getsinglebanner und getsinglebannerfirst bei multi passt es
| 1
|
254,441
| 27,387,763,705
|
IssuesEvent
|
2023-02-28 14:25:24
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
SOCKS access for Gary Fallon
|
operations external-request ops-access-request platform-security
|
### Your Name
Gary Fallon
### Your Email
gary.fallon@oddball.io
### Your Role and Team
Security Engineer on Platform Security
### Product Manager (PM) name and email
Alyssa Gallion alyssa.gallion@oddball.io
### Product Owner (PO) name and email
Ray Wang gang.wang@va.gov
### Access Type Requesting
viewers
### Public SSH Key
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCugbZFH0/4h+y1NM5OhvnGdEYINNc/9CUxg+mzP2NTaCmWCtwYVlPIo3w++mAZ6Gd9KfCXiZCfETtjz/flqGrPnn3TlaTflYbvrvCQH4Wo+UV5mN0vXY+MK5dtrYUSka1VwHcvuBNpHTwPq3UCbK9TTh9mFLQ4uP+iggrNBVGxo6qs2tBt4VXh1jeyorJVGNZrwtcLyCbc6Uq+YftkZFDFURO1h6AhQ0J5oJTQRZAs/ci2jI5szxlDE4yZWzBeSXRize469XMvz29W0VClF0VQI8uUk0CJX+Nzyr0m6oBk7vQmoqditdb6YrprK8gdU2OWo9+dhzPghgYvOeZmLAzJ gfall@Garys-Air
### User must exist in a roster before SOCKS access can be granted
- [ ] Search for user on the VFS Team Roster: https://docs.google.com/spreadsheets/d/11dpCJjhs007uC6CWJI6djy3OAvjB8rHB65m0Yj8HXIw/edit?folder=0ALlyxurHpUilUk9PVA#gid=2042046665
- [ ] Or search for user on the Platform Team Roster: https://vfs.atlassian.net/wiki/spaces/AP/pages/1908834623/Platform+Roster
- [ ] If user is on a VFS team but not in the VFS Team Roster, add the 'NOT YET' label and instruct them to start the Platform orientation process https://depo-platform-documentation.scrollhelp.site/getting-started/Platform-Orientation.1877344532.html
- [ ] If a user is on a Platform team but not on the Platform Team Roster in Confluence, add the 'NOT YET' label and instruct them to reach out to their Product Manager to be added.
- [ ] Comment in this issue saying which roster the user is listed in.
|
True
|
SOCKS access for Gary Fallon - ### Your Name
Gary Fallon
### Your Email
gary.fallon@oddball.io
### Your Role and Team
Security Engineer on Platform Security
### Product Manager (PM) name and email
Alyssa Gallion alyssa.gallion@oddball.io
### Product Owner (PO) name and email
Ray Wang gang.wang@va.gov
### Access Type Requesting
viewers
### Public SSH Key
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCugbZFH0/4h+y1NM5OhvnGdEYINNc/9CUxg+mzP2NTaCmWCtwYVlPIo3w++mAZ6Gd9KfCXiZCfETtjz/flqGrPnn3TlaTflYbvrvCQH4Wo+UV5mN0vXY+MK5dtrYUSka1VwHcvuBNpHTwPq3UCbK9TTh9mFLQ4uP+iggrNBVGxo6qs2tBt4VXh1jeyorJVGNZrwtcLyCbc6Uq+YftkZFDFURO1h6AhQ0J5oJTQRZAs/ci2jI5szxlDE4yZWzBeSXRize469XMvz29W0VClF0VQI8uUk0CJX+Nzyr0m6oBk7vQmoqditdb6YrprK8gdU2OWo9+dhzPghgYvOeZmLAzJ gfall@Garys-Air
### User must exist in a roster before SOCKS access can be granted
- [ ] Search for user on the VFS Team Roster: https://docs.google.com/spreadsheets/d/11dpCJjhs007uC6CWJI6djy3OAvjB8rHB65m0Yj8HXIw/edit?folder=0ALlyxurHpUilUk9PVA#gid=2042046665
- [ ] Or search for user on the Platform Team Roster: https://vfs.atlassian.net/wiki/spaces/AP/pages/1908834623/Platform+Roster
- [ ] If user is on a VFS team but not in the VFS Team Roster, add the 'NOT YET' label and instruct them to start the Platform orientation process https://depo-platform-documentation.scrollhelp.site/getting-started/Platform-Orientation.1877344532.html
- [ ] If a user is on a Platform team but not on the Platform Team Roster in Confluence, add the 'NOT YET' label and instruct them to reach out to their Product Manager to be added.
- [ ] Comment in this issue saying which roster the user is listed in.
|
non_defect
|
socks access for gary fallon your name gary fallon your email gary fallon oddball io your role and team security engineer on platform security product manager pm name and email alyssa gallion alyssa gallion oddball io product owner po name and email ray wang gang wang va gov access type requesting viewers public ssh key ssh rsa dhzpghgyvoezmlazj gfall garys air user must exist in a roster before socks access can be granted search for user on the vfs team roster or search for user on the platform team roster if user is on a vfs team but not in the vfs team roster add the not yet label and instruct them to start the platform orientation process if a user is on a platform team but not on the platform team roster in confluence add the not yet label and instruct them to reach out to their product manager to be added comment in this issue saying which roster the user is listed in
| 0
|
22,504
| 3,655,281,093
|
IssuesEvent
|
2016-02-17 15:48:40
|
contao/core-bundle
|
https://api.github.com/repos/contao/core-bundle
|
reopened
|
Image::setTargetPath problem with /web/ folder
|
defect
|
Yet another problematic folder. When setting a target folder for an image, the path is relative to `TL_ROOT`. Which might not be available from the web directory.
|
1.0
|
Image::setTargetPath problem with /web/ folder - Yet another problematic folder. When setting a target folder for an image, the path is relative to `TL_ROOT`. Which might not be available from the web directory.
|
defect
|
image settargetpath problem with web folder yet another problematic folder when setting a target folder for an image the path is relative to tl root which might not be available from the web directory
| 1
|
50,232
| 13,187,390,702
|
IssuesEvent
|
2020-08-13 03:15:55
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'> (Trac #348)
|
Migrated from Trac combo simulation defect
|
I can't see the revision log for romeo trunk
==== How to Reproduce ====
While doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.
''(please provide additional details here)''
Request parameters:
```text
{'path': u'/projects/romeo/trunk'}
```
User Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`
==== System Information ====
|| '''Trac''' || `0.11.1` ||
|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||
|| '''setuptools''' || `0.6` ||
|| '''psycopg2''' || `2.2.1` ||
|| '''Genshi''' || `0.6` ||
|| '''mod_python''' || `3.3.1` ||
|| '''Pygments''' || `0.10` ||
|| '''Subversion''' || `1.5.1 (r32289)` ||
|| '''jQuery:''' || `1.2.6` ||
==== Python Traceback ====
```text
Traceback (most recent call last):
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py", line 423, in _dispatch_request
dispatcher.dispatch(req)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py", line 219, in dispatch
data, content_type)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 715, in render_template
return stream.render(method, doctype=doctype)
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 183, in render
return encode(generator, method=method, encoding=encoding, out=out)
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 57, in encode
return _encode(''.join(list(iterator)))
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 339, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 826, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 670, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 771, in __call__
for kind, data, pos in chain(stream, [(None, None, None)]):
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 586, in __call__
for ev in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 767, in _strip_accesskeys
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 756, in _generate
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 378, in _match
ctxt, start=idx + 1, **vars):
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 378, in _match
ctxt, start=idx + 1, **vars):
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 327, in _match
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 545, in _flatten
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/path.py", line 588, in _generate
subevent = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 316, in _strip
event = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 545, in _flatten
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/path.py", line 588, in _generate
subevent = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 316, in _strip
event = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 565, in _flatten
result = _eval_expr(data, ctxt, vars)
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 277, in _eval_expr
retval = expr.evaluate(ctxt)
File "/usr/lib/pymodules/python2.5/genshi/template/eval.py", line 178, in evaluate
return eval(self.code, _globals, {'__data__': data})
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html", line 134, in <Expression u'dateinfo(change.date)'>
<td class="date" py:content="dateinfo(change.date)" />
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py", line 230, in dateinfo
pretty_timedelta(date),
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py", line 71, in pretty_timedelta
time1 = to_datetime(time1)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py", line 51, in to_datetime
type(t))
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>
```
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/348
, reported by mase and owned by nega_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T20:08:32",
"description": "I can't see the revision log for romeo trunk\n\n\n==== How to Reproduce ====\n\nWhile doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.\n\n''(please provide additional details here)''\n\n\nRequest parameters:\n{{{\n{'path': u'/projects/romeo/trunk'}\n}}}\n\n\nUser Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`\n\n==== System Information ====\n\n|| '''Trac''' || `0.11.1` ||\n|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||\n|| '''setuptools''' || `0.6` ||\n|| '''psycopg2''' || `2.2.1` ||\n|| '''Genshi''' || `0.6` ||\n|| '''mod_python''' || `3.3.1` ||\n|| '''Pygments''' || `0.10` ||\n|| '''Subversion''' || `1.5.1 (r32289)` ||\n|| '''jQuery:''' || `1.2.6` ||\n\n==== Python Traceback ====\n{{{\nTraceback (most recent call last):\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 423, in _dispatch_request\n dispatcher.dispatch(req)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 219, in dispatch\n data, content_type)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 715, in render_template\n return stream.render(method, doctype=doctype)\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 183, in render\n return encode(generator, method=method, encoding=encoding, out=out)\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 57, in encode\n return _encode(''.join(list(iterator)))\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 339, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 826, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 670, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 771, in __call__\n for kind, data, pos in chain(stream, [(None, None, None)]):\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 586, in __call__\n for ev in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 767, in _strip_accesskeys\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 756, in _generate\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 327, in _match\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 565, in _flatten\n result = _eval_expr(data, ctxt, vars)\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 277, in _eval_expr\n retval = expr.evaluate(ctxt)\n File \"/usr/lib/pymodules/python2.5/genshi/template/eval.py\", line 178, in evaluate\n return eval(self.code, _globals, {'__data__': data})\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html\", line 134, in <Expression u'dateinfo(change.date)'>\n <td class=\"date\" py:content=\"dateinfo(change.date)\" />\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py\", line 230, in dateinfo\n pretty_timedelta(date),\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 71, in pretty_timedelta\n time1 = to_datetime(time1)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 51, in to_datetime\n type(t))\nTypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>\n\n}}}\n ",
"reporter": "mase",
"cc": "",
"resolution": "fixed",
"_ts": "1547237312321270",
"component": "combo simulation",
"summary": "TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>",
"priority": "normal",
"keywords": "",
"time": "2012-01-30T03:41:32",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'> (Trac #348) - I can't see the revision log for romeo trunk
==== How to Reproduce ====
While doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.
''(please provide additional details here)''
Request parameters:
```text
{'path': u'/projects/romeo/trunk'}
```
User Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`
==== System Information ====
|| '''Trac''' || `0.11.1` ||
|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||
|| '''setuptools''' || `0.6` ||
|| '''psycopg2''' || `2.2.1` ||
|| '''Genshi''' || `0.6` ||
|| '''mod_python''' || `3.3.1` ||
|| '''Pygments''' || `0.10` ||
|| '''Subversion''' || `1.5.1 (r32289)` ||
|| '''jQuery:''' || `1.2.6` ||
==== Python Traceback ====
```text
Traceback (most recent call last):
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py", line 423, in _dispatch_request
dispatcher.dispatch(req)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py", line 219, in dispatch
data, content_type)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 715, in render_template
return stream.render(method, doctype=doctype)
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 183, in render
return encode(generator, method=method, encoding=encoding, out=out)
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 57, in encode
return _encode(''.join(list(iterator)))
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 339, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 826, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 670, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 771, in __call__
for kind, data, pos in chain(stream, [(None, None, None)]):
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 586, in __call__
for ev in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 767, in _strip_accesskeys
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 756, in _generate
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 378, in _match
ctxt, start=idx + 1, **vars):
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 378, in _match
ctxt, start=idx + 1, **vars):
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 327, in _match
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 545, in _flatten
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/path.py", line 588, in _generate
subevent = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 316, in _strip
event = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 545, in _flatten
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/path.py", line 588, in _generate
subevent = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 316, in _strip
event = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 565, in _flatten
result = _eval_expr(data, ctxt, vars)
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 277, in _eval_expr
retval = expr.evaluate(ctxt)
File "/usr/lib/pymodules/python2.5/genshi/template/eval.py", line 178, in evaluate
return eval(self.code, _globals, {'__data__': data})
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html", line 134, in <Expression u'dateinfo(change.date)'>
<td class="date" py:content="dateinfo(change.date)" />
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py", line 230, in dateinfo
pretty_timedelta(date),
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py", line 71, in pretty_timedelta
time1 = to_datetime(time1)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py", line 51, in to_datetime
type(t))
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>
```
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/348
, reported by mase and owned by nega_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T20:08:32",
"description": "I can't see the revision log for romeo trunk\n\n\n==== How to Reproduce ====\n\nWhile doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.\n\n''(please provide additional details here)''\n\n\nRequest parameters:\n{{{\n{'path': u'/projects/romeo/trunk'}\n}}}\n\n\nUser Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`\n\n==== System Information ====\n\n|| '''Trac''' || `0.11.1` ||\n|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||\n|| '''setuptools''' || `0.6` ||\n|| '''psycopg2''' || `2.2.1` ||\n|| '''Genshi''' || `0.6` ||\n|| '''mod_python''' || `3.3.1` ||\n|| '''Pygments''' || `0.10` ||\n|| '''Subversion''' || `1.5.1 (r32289)` ||\n|| '''jQuery:''' || `1.2.6` ||\n\n==== Python Traceback ====\n{{{\nTraceback (most recent call last):\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 423, in _dispatch_request\n dispatcher.dispatch(req)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 219, in dispatch\n data, content_type)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 715, in render_template\n return stream.render(method, doctype=doctype)\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 183, in render\n return encode(generator, method=method, encoding=encoding, out=out)\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 57, in encode\n return _encode(''.join(list(iterator)))\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 339, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 826, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 670, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 771, in __call__\n for kind, data, pos in chain(stream, [(None, None, None)]):\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 586, in __call__\n for ev in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 767, in _strip_accesskeys\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 756, in _generate\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 327, in _match\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 565, in _flatten\n result = _eval_expr(data, ctxt, vars)\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 277, in _eval_expr\n retval = expr.evaluate(ctxt)\n File \"/usr/lib/pymodules/python2.5/genshi/template/eval.py\", line 178, in evaluate\n return eval(self.code, _globals, {'__data__': data})\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html\", line 134, in <Expression u'dateinfo(change.date)'>\n <td class=\"date\" py:content=\"dateinfo(change.date)\" />\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py\", line 230, in dateinfo\n pretty_timedelta(date),\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 71, in pretty_timedelta\n time1 = to_datetime(time1)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 51, in to_datetime\n type(t))\nTypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>\n\n}}}\n ",
"reporter": "mase",
"cc": "",
"resolution": "fixed",
"_ts": "1547237312321270",
"component": "combo simulation",
"summary": "TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>",
"priority": "normal",
"keywords": "",
"time": "2012-01-30T03:41:32",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
typeerror expecting datetime int long float or none got trac i can t see the revision log for romeo trunk how to reproduce while doing a get operation on log projects romeo trunk trac issued an internal error please provide additional details here request parameters text path u projects romeo trunk user agent was mozilla u linux ja rv gecko ubuntu maverick firefox system information trac python nov setuptools genshi mod python pygments subversion jquery python traceback text traceback most recent call last file opt trac local lib site packages trac egg trac web main py line in dispatch request dispatcher dispatch req file opt trac local lib site packages trac egg trac web main py line in dispatch data content type file opt trac local lib site packages trac egg trac web chrome py line in render template return stream render method doctype doctype file usr lib pymodules genshi core py line in render return encode generator method method encoding encoding out out file usr lib pymodules genshi output py line in encode return encode join list iterator file usr lib pymodules genshi output py line in call for kind data pos in stream file usr lib pymodules genshi output py line in call for kind data pos in stream file usr lib pymodules genshi output py line in call for kind data pos in stream file usr lib pymodules genshi output py line in call for kind data pos in chain stream file usr lib pymodules genshi output py line in call for ev in stream file usr lib pymodules genshi core py line in ensure for event in stream file usr lib pymodules genshi core py line in ensure for event in stream file opt trac local lib site packages trac egg trac web chrome py line in strip accesskeys for kind data pos in stream file usr lib pymodules genshi core py line in ensure for event in stream file opt trac local lib site packages trac egg trac web chrome py line in generate for kind data pos in stream file usr lib pymodules genshi template base py line in include for event in stream file usr lib pymodules genshi template markup py line in match ctxt start idx vars file usr lib pymodules genshi template markup py line in match ctxt start idx vars file usr lib pymodules genshi template markup py line in match for event in stream file usr lib pymodules genshi template base py line in flatten for kind data pos in stream file usr lib pymodules genshi core py line in ensure for event in stream file usr lib pymodules genshi path py line in generate subevent next file usr lib pymodules genshi template base py line in include for event in stream file usr lib pymodules genshi template markup py line in strip event next file usr lib pymodules genshi template base py line in flatten for kind data pos in stream file usr lib pymodules genshi core py line in ensure for event in stream file usr lib pymodules genshi path py line in generate subevent next file usr lib pymodules genshi template base py line in include for event in stream file usr lib pymodules genshi template markup py line in strip event next file usr lib pymodules genshi template base py line in flatten result eval expr data ctxt vars file usr lib pymodules genshi template base py line in eval expr retval expr evaluate ctxt file usr lib pymodules genshi template eval py line in evaluate return eval self code globals data data file opt trac local lib site packages trac egg trac versioncontrol templates revisionlog html line in file opt trac local lib site packages trac egg trac timeline web ui py line in dateinfo pretty timedelta date file opt trac local lib site packages trac egg trac util datefmt py line in pretty timedelta to datetime file opt trac local lib site packages trac egg trac util datefmt py line in to datetime type t typeerror expecting datetime int long float or none got migrated from reported by mase and owned by nega json status closed changetime description i can t see the revision log for romeo trunk n n n how to reproduce n nwhile doing a get operation on log projects romeo trunk trac issued an internal error n n please provide additional details here n n nrequest parameters n n path u projects romeo trunk n n n nuser agent was mozilla u linux ja rv gecko ubuntu maverick firefox n n system information n n trac n python nov n setuptools n n genshi n mod python n pygments n subversion n jquery n n python traceback n ntraceback most recent call last n file opt trac local lib site packages trac egg trac web main py line in dispatch request n dispatcher dispatch req n file opt trac local lib site packages trac egg trac web main py line in dispatch n data content type n file opt trac local lib site packages trac egg trac web chrome py line in render template n return stream render method doctype doctype n file usr lib pymodules genshi core py line in render n return encode generator method method encoding encoding out out n file usr lib pymodules genshi output py line in encode n return encode join list iterator n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in chain stream n file usr lib pymodules genshi output py line in call n for ev in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file opt trac local lib site packages trac egg trac web chrome py line in strip accesskeys n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file opt trac local lib site packages trac egg trac web chrome py line in generate n for kind data pos in stream n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in match n ctxt start idx vars n file usr lib pymodules genshi template markup py line in match n ctxt start idx vars n file usr lib pymodules genshi template markup py line in match n for event in stream n file usr lib pymodules genshi template base py line in flatten n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi path py line in generate n subevent next n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in strip n event next n file usr lib pymodules genshi template base py line in flatten n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi path py line in generate n subevent next n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in strip n event next n file usr lib pymodules genshi template base py line in flatten n result eval expr data ctxt vars n file usr lib pymodules genshi template base py line in eval expr n retval expr evaluate ctxt n file usr lib pymodules genshi template eval py line in evaluate n return eval self code globals data data n file opt trac local lib site packages trac egg trac versioncontrol templates revisionlog html line in n n file opt trac local lib site packages trac egg trac timeline web ui py line in dateinfo n pretty timedelta date n file opt trac local lib site packages trac egg trac util datefmt py line in pretty timedelta n to datetime n file opt trac local lib site packages trac egg trac util datefmt py line in to datetime n type t ntypeerror expecting datetime int long float or none got n n n reporter mase cc resolution fixed ts component combo simulation summary typeerror expecting datetime int long float or none got priority normal keywords time milestone owner nega type defect
| 1
|
439,761
| 30,713,322,778
|
IssuesEvent
|
2023-07-27 11:21:14
|
Stephen-Hamilton-C/timecard-lib
|
https://api.github.com/repos/Stephen-Hamilton-C/timecard-lib
|
closed
|
Generate API Documentation with Dokka
|
documentation enhancement
|
See https://kotlinlang.org/docs/dokka-introduction.html
Look into placing this generated documentation into the Github Wiki for this repo
|
1.0
|
Generate API Documentation with Dokka - See https://kotlinlang.org/docs/dokka-introduction.html
Look into placing this generated documentation into the Github Wiki for this repo
|
non_defect
|
generate api documentation with dokka see look into placing this generated documentation into the github wiki for this repo
| 0
|
41,979
| 10,732,134,888
|
IssuesEvent
|
2019-10-28 21:07:38
|
jccastillo0007/eFacturaT
|
https://api.github.com/repos/jccastillo0007/eFacturaT
|
opened
|
Condominios - Incluir en el nombre de la SP el id CASA
|
defect
|
Hoy aparece null, supongo que porque no tienen RFC asociado.
Y por el simple nombre no se conoce de que casa es.
|
1.0
|
Condominios - Incluir en el nombre de la SP el id CASA - Hoy aparece null, supongo que porque no tienen RFC asociado.
Y por el simple nombre no se conoce de que casa es.
|
defect
|
condominios incluir en el nombre de la sp el id casa hoy aparece null supongo que porque no tienen rfc asociado y por el simple nombre no se conoce de que casa es
| 1
|
8,247
| 2,611,473,326
|
IssuesEvent
|
2015-02-27 05:17:29
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
opened
|
Hog swap too fast, plus "hotseat" issues.
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Start any game with any number of players (2~8), preferentially with at
least 2 sharing the same computer to be able to reproduce both problems.
2. Try to play normally
3. ???
4. Profit
What is the expected output? What do you see instead?
During the game, the next hog change is so fast a lot of the times you don't
even get to read the total damage dealt.
Another issue is the "hot-seat" mode, where more than 1 human player play on
the same computer by swapping seats. More often than desired you'll end up
accidentally jumping or double-jumping with the other team hog by accident,
unless everyone is VERY careful about their jump spam, which usually never
happens... xD And end up wasting the other players time.
What version of the product are you using? On what operating system?
Hedgewars 0.9.16
Windows 7 Professional 64bit
Please provide any additional information below.
Was not an issue on 0.9.15.
The quick change of hog is a good addition, but only for specific turn ends,
like when a hog dies.
```
Original issue reported on code.google.com by `inufa...@gmail.com` on 18 Sep 2011 at 7:59
|
1.0
|
Hog swap too fast, plus "hotseat" issues. - ```
What steps will reproduce the problem?
1. Start any game with any number of players (2~8), preferentially with at
least 2 sharing the same computer to be able to reproduce both problems.
2. Try to play normally
3. ???
4. Profit
What is the expected output? What do you see instead?
During the game, the next hog change is so fast a lot of the times you don't
even get to read the total damage dealt.
Another issue is the "hot-seat" mode, where more than 1 human player play on
the same computer by swapping seats. More often than desired you'll end up
accidentally jumping or double-jumping with the other team hog by accident,
unless everyone is VERY careful about their jump spam, which usually never
happens... xD And end up wasting the other players time.
What version of the product are you using? On what operating system?
Hedgewars 0.9.16
Windows 7 Professional 64bit
Please provide any additional information below.
Was not an issue on 0.9.15.
The quick change of hog is a good addition, but only for specific turn ends,
like when a hog dies.
```
Original issue reported on code.google.com by `inufa...@gmail.com` on 18 Sep 2011 at 7:59
|
defect
|
hog swap too fast plus hotseat issues what steps will reproduce the problem start any game with any number of players preferentially with at least sharing the same computer to be able to reproduce both problems try to play normally profit what is the expected output what do you see instead during the game the next hog change is so fast a lot of the times you don t even get to read the total damage dealt another issue is the hot seat mode where more than human player play on the same computer by swapping seats more often than desired you ll end up accidentally jumping or double jumping with the other team hog by accident unless everyone is very careful about their jump spam which usually never happens xd and end up wasting the other players time what version of the product are you using on what operating system hedgewars windows professional please provide any additional information below was not an issue on the quick change of hog is a good addition but only for specific turn ends like when a hog dies original issue reported on code google com by inufa gmail com on sep at
| 1
|
177,978
| 29,368,357,978
|
IssuesEvent
|
2023-05-29 00:01:41
|
ManageIQ/manageiq
|
https://api.github.com/repos/ManageIQ/manageiq
|
closed
|
Features dependency for an API-driven UI
|
enhancement stale redesign size/xl
|
We're moving the classic UI to be much more an API-driven and this is already causing problems for roles with limited access to certain resources. For [example](https://bugzilla.redhat.com/show_bug.cgi?id=1703967) the about modal would need at least read-only access to the information about the current Zone and Region. However, exposing the related features (if they exist) to the given role would also expose parts of the UI that we might not want to display. Also it's a bad user experience to select 1+N features in the features tree if we just need that one single feature.
There will be a similar problem with the Set Ownership form for VMs/Services if it [becomes](https://github.com/ManageIQ/manageiq-ui-classic/pull/5863) API driven as the role needs additional access to the `/api/users/` and `/api/groups`. As the matching read features are tied to the UI and live under `Configuration -> Access Control` we might not want to expose them in the menu.
For now we don't have many places where we're using the API, but the problem will be growing in the future as our plan is to use the API more. Therefore, it is inevitable to make some changes on our current RBAC architecture.
I'm proposing an additional field in the `miq_features.yml` for each API-feature, let's call it `dependent_api_features` where we could define a list of features that would be exposed through the API when enabling the given feature. This would not expose anything unwanted on the UI and also prevent the *1+N UX problem* from happening.
```yaml
- :name: Edit
:description: Edit a VM
:feature_type: admin
:identifier: vm_edit
:dependent_api_features:
- rbac_user_show_list
- rbac_group_show_list
```
If the above feature is selected, it would be allowed strictly via API only to also list users and groups. So when editing the ownership of the VM, we are able to build the dropdown for selecting the user/group for a limited role without exposing the `Settings -> Access` control.
I think we need this feature before we start having a lot of API-driven forms so it's easier to fill out the yaml with the dependencies when we're working on the same stuff in the UI area.
@kbrock @lpichler @martinpovolny @himdel @Fryguy
|
1.0
|
Features dependency for an API-driven UI - We're moving the classic UI to be much more an API-driven and this is already causing problems for roles with limited access to certain resources. For [example](https://bugzilla.redhat.com/show_bug.cgi?id=1703967) the about modal would need at least read-only access to the information about the current Zone and Region. However, exposing the related features (if they exist) to the given role would also expose parts of the UI that we might not want to display. Also it's a bad user experience to select 1+N features in the features tree if we just need that one single feature.
There will be a similar problem with the Set Ownership form for VMs/Services if it [becomes](https://github.com/ManageIQ/manageiq-ui-classic/pull/5863) API driven as the role needs additional access to the `/api/users/` and `/api/groups`. As the matching read features are tied to the UI and live under `Configuration -> Access Control` we might not want to expose them in the menu.
For now we don't have many places where we're using the API, but the problem will be growing in the future as our plan is to use the API more. Therefore, it is inevitable to make some changes on our current RBAC architecture.
I'm proposing an additional field in the `miq_features.yml` for each API-feature, let's call it `dependent_api_features` where we could define a list of features that would be exposed through the API when enabling the given feature. This would not expose anything unwanted on the UI and also prevent the *1+N UX problem* from happening.
```yaml
- :name: Edit
:description: Edit a VM
:feature_type: admin
:identifier: vm_edit
:dependent_api_features:
- rbac_user_show_list
- rbac_group_show_list
```
If the above feature is selected, it would be allowed strictly via API only to also list users and groups. So when editing the ownership of the VM, we are able to build the dropdown for selecting the user/group for a limited role without exposing the `Settings -> Access` control.
I think we need this feature before we start having a lot of API-driven forms so it's easier to fill out the yaml with the dependencies when we're working on the same stuff in the UI area.
@kbrock @lpichler @martinpovolny @himdel @Fryguy
|
non_defect
|
features dependency for an api driven ui we re moving the classic ui to be much more an api driven and this is already causing problems for roles with limited access to certain resources for the about modal would need at least read only access to the information about the current zone and region however exposing the related features if they exist to the given role would also expose parts of the ui that we might not want to display also it s a bad user experience to select n features in the features tree if we just need that one single feature there will be a similar problem with the set ownership form for vms services if it api driven as the role needs additional access to the api users and api groups as the matching read features are tied to the ui and live under configuration access control we might not want to expose them in the menu for now we don t have many places where we re using the api but the problem will be growing in the future as our plan is to use the api more therefore it is inevitable to make some changes on our current rbac architecture i m proposing an additional field in the miq features yml for each api feature let s call it dependent api features where we could define a list of features that would be exposed through the api when enabling the given feature this would not expose anything unwanted on the ui and also prevent the n ux problem from happening yaml name edit description edit a vm feature type admin identifier vm edit dependent api features rbac user show list rbac group show list if the above feature is selected it would be allowed strictly via api only to also list users and groups so when editing the ownership of the vm we are able to build the dropdown for selecting the user group for a limited role without exposing the settings access control i think we need this feature before we start having a lot of api driven forms so it s easier to fill out the yaml with the dependencies when we re working on the same stuff in the ui area kbrock lpichler martinpovolny himdel fryguy
| 0
|
39,766
| 9,651,735,188
|
IssuesEvent
|
2019-05-18 10:45:06
|
boxbackup/boxbackup
|
https://api.github.com/repos/boxbackup/boxbackup
|
closed
|
Windows client 1662 has 3 extra files? (Trac #29)
|
Migrated from Trac bbackupd ben defect
|
Found in boxbackup-chris_general_1662-backup-client-mingw32.zip
I don't think are of any use, are they?:
```text
bbackupd-config
install-backup-client
installer.iss
```
Migrated from https://www.boxbackup.org/ticket/29
```json
{
"status": "closed",
"changetime": "2007-09-19T22:11:07",
"description": "Found in boxbackup-chris_general_1662-backup-client-mingw32.zip\n\nI don't think are of any use, are they?:\n\n{{{\nbbackupd-config\ninstall-backup-client\ninstaller.iss\n}}}\n",
"reporter": "petej",
"cc": "",
"resolution": "duplicate",
"_ts": "1190239867000000",
"component": "bbackupd",
"summary": "Windows client 1662 has 3 extra files?",
"priority": "trivial",
"keywords": "",
"version": "0.10",
"time": "2007-05-17T04:09:47",
"milestone": "",
"owner": "ben",
"type": "defect"
}
```
|
1.0
|
Windows client 1662 has 3 extra files? (Trac #29) - Found in boxbackup-chris_general_1662-backup-client-mingw32.zip
I don't think are of any use, are they?:
```text
bbackupd-config
install-backup-client
installer.iss
```
Migrated from https://www.boxbackup.org/ticket/29
```json
{
"status": "closed",
"changetime": "2007-09-19T22:11:07",
"description": "Found in boxbackup-chris_general_1662-backup-client-mingw32.zip\n\nI don't think are of any use, are they?:\n\n{{{\nbbackupd-config\ninstall-backup-client\ninstaller.iss\n}}}\n",
"reporter": "petej",
"cc": "",
"resolution": "duplicate",
"_ts": "1190239867000000",
"component": "bbackupd",
"summary": "Windows client 1662 has 3 extra files?",
"priority": "trivial",
"keywords": "",
"version": "0.10",
"time": "2007-05-17T04:09:47",
"milestone": "",
"owner": "ben",
"type": "defect"
}
```
|
defect
|
windows client has extra files trac found in boxbackup chris general backup client zip i don t think are of any use are they text bbackupd config install backup client installer iss migrated from json status closed changetime description found in boxbackup chris general backup client zip n ni don t think are of any use are they n n nbbackupd config ninstall backup client ninstaller iss n n reporter petej cc resolution duplicate ts component bbackupd summary windows client has extra files priority trivial keywords version time milestone owner ben type defect
| 1
|
11,233
| 2,641,947,100
|
IssuesEvent
|
2015-03-11 20:40:47
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
closed
|
Does't work with SeaMonkey
|
Priority-Medium Type-Defect
|
Original [issue 148](https://code.google.com/p/html5rocks/issues/detail?id=148) created by chrsmith on 2010-08-14T01:23:27.000Z:
<b>Please describe the issue:</b>
I'm using SeaMonkey, build Mozilla/5.0 (X11; Linux i686; rv:2.0b4pre) Gecko/20100811
which sould work on any page Firefox beta 3 works. But clicking on any link doesn't work.
<b>Please provide any additional information below.</b>
The better way would be checking for gecko, instead of checking for Firefox.
It doesn't even raise an error.
|
1.0
|
Does't work with SeaMonkey - Original [issue 148](https://code.google.com/p/html5rocks/issues/detail?id=148) created by chrsmith on 2010-08-14T01:23:27.000Z:
<b>Please describe the issue:</b>
I'm using SeaMonkey, build Mozilla/5.0 (X11; Linux i686; rv:2.0b4pre) Gecko/20100811
which sould work on any page Firefox beta 3 works. But clicking on any link doesn't work.
<b>Please provide any additional information below.</b>
The better way would be checking for gecko, instead of checking for Firefox.
It doesn't even raise an error.
|
defect
|
does t work with seamonkey original created by chrsmith on please describe the issue i m using seamonkey build mozilla linux rv gecko which sould work on any page firefox beta works but clicking on any link doesn t work please provide any additional information below the better way would be checking for gecko instead of checking for firefox it doesn t even raise an error
| 1
|
21,636
| 11,660,461,003
|
IssuesEvent
|
2020-03-03 03:25:13
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
opened
|
VZD | Update Units Table to include `atd_mode_category`
|
Need: 1-Must Have Product: Vision Zero Crash Data System Project: Vision Zero Crash Data System Service: Dev Type: Enhancement Workgroup: VZ migrated
|
- [x] Add new column `atd_mode_category` to Unit Table
- [x] On Unit record creation, assign an `atd_mode_category` from the Mode Categories Lookup table (https://github.com/cityofaustin/atd-vz-data/issues/615) based on the record's `unit_desc_id` & `veh_body_styl_id`
*Migrated from [atd-vz-data #656](https://github.com/cityofaustin/atd-vz-data/issues/656)*
|
1.0
|
VZD | Update Units Table to include `atd_mode_category` - - [x] Add new column `atd_mode_category` to Unit Table
- [x] On Unit record creation, assign an `atd_mode_category` from the Mode Categories Lookup table (https://github.com/cityofaustin/atd-vz-data/issues/615) based on the record's `unit_desc_id` & `veh_body_styl_id`
*Migrated from [atd-vz-data #656](https://github.com/cityofaustin/atd-vz-data/issues/656)*
|
non_defect
|
vzd update units table to include atd mode category add new column atd mode category to unit table on unit record creation assign an atd mode category from the mode categories lookup table based on the record s unit desc id veh body styl id migrated from
| 0
|
75,142
| 25,551,646,207
|
IssuesEvent
|
2022-11-30 00:35:53
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
closed
|
[🐛 Bug]: RemoteWebDriver.findElement works weirdly
|
R-awaiting answer I-defect needs-triaging
|
### What happened?
When I tried to find an element by xpath, it works.
The current call stack of method was done and then jvm went into other method call stack.
I tried to do that again at there, but it doesn't work.
There is no action for ChromeDriver. The changes are only call stack.
I don't understand why.
```java
driver.findElement(By.xpath("//script[not(@src) and contains(text(), '__LZ_ME__')]"));
```
### How can we reproduce the issue?
```shell
1. Clone https://github.com/ImSejin/lezhin-comics-downloader on branch dev.
2. Write down your login info of lezhin comics to config.ini
3. Run io.github.imsejin.dl.lezhin.Application on debug mode.
4. Compare LoginProcessor.getAccessToken() with AccessTokenProcessor.getAccessToken().
```
### Relevant log output
```shell
Starting ChromeDriver 107.0.5304.62 (1eec40d3a5764881c92085aaee66d25075c159aa-refs/branch-heads/5304@{#942}) on port 5280
Only local connections are allowed.
Please see https://chromedriver.chromium.org/security-considerations for suggestions on keeping ChromeDriver safe.
ChromeDriver was started successfully.
Exception in thread "main" java.lang.RuntimeException: Cannot find access token
at io.github.imsejin.dl.lezhin.process.impl.AccessTokenProcessor.getAccessToken(AccessTokenProcessor.java:117)
at io.github.imsejin.dl.lezhin.process.impl.AccessTokenProcessor.process(AccessTokenProcessor.java:80)
at io.github.imsejin.dl.lezhin.process.impl.AccessTokenProcessor.process(AccessTokenProcessor.java:75)
at io.github.imsejin.dl.lezhin.Application.main(Application.java:73)
Caused by: org.openqa.selenium.NoSuchElementException: no such element: Unable to locate element: {"method":"xpath","selector":"//script[not(@src) and contains(text(), '__LZ_ME__')]"}
(Session info: headless chrome=107.0.5304.108)
For documentation on this error, please visit: https://selenium.dev/exceptions/#no_such_element
Build info: version: '4.6.0', revision: '79f1c02ae20'
System info: os.name: 'Windows 10', os.arch: 'amd64', os.version: '10.0', java.version: '15.0.7'
Driver info: org.openqa.selenium.chrome.ChromeDriver
Command: [de5624873d2b4a72ae1bdaca9b137fda, findElement {using=xpath, value=//script[not(@src) and contains(text(), '__LZ_ME__')]}]
Capabilities {acceptInsecureCerts: false, browserName: chrome, browserVersion: 107.0.5304.108, chrome: {chromedriverVersion: 107.0.5304.62
```
### Operating System
Windows 10
### Selenium version
Java 4.6.0
### What are the browser(s) and version(s) where you see this issue?
Chrome 107.0.5304.108
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 107.0.5304.62
### Are you using Selenium Grid?
No
|
1.0
|
[🐛 Bug]: RemoteWebDriver.findElement works weirdly - ### What happened?
When I tried to find an element by xpath, it works.
The current call stack of method was done and then jvm went into other method call stack.
I tried to do that again at there, but it doesn't work.
There is no action for ChromeDriver. The changes are only call stack.
I don't understand why.
```java
driver.findElement(By.xpath("//script[not(@src) and contains(text(), '__LZ_ME__')]"));
```
### How can we reproduce the issue?
```shell
1. Clone https://github.com/ImSejin/lezhin-comics-downloader on branch dev.
2. Write down your login info of lezhin comics to config.ini
3. Run io.github.imsejin.dl.lezhin.Application on debug mode.
4. Compare LoginProcessor.getAccessToken() with AccessTokenProcessor.getAccessToken().
```
### Relevant log output
```shell
Starting ChromeDriver 107.0.5304.62 (1eec40d3a5764881c92085aaee66d25075c159aa-refs/branch-heads/5304@{#942}) on port 5280
Only local connections are allowed.
Please see https://chromedriver.chromium.org/security-considerations for suggestions on keeping ChromeDriver safe.
ChromeDriver was started successfully.
Exception in thread "main" java.lang.RuntimeException: Cannot find access token
at io.github.imsejin.dl.lezhin.process.impl.AccessTokenProcessor.getAccessToken(AccessTokenProcessor.java:117)
at io.github.imsejin.dl.lezhin.process.impl.AccessTokenProcessor.process(AccessTokenProcessor.java:80)
at io.github.imsejin.dl.lezhin.process.impl.AccessTokenProcessor.process(AccessTokenProcessor.java:75)
at io.github.imsejin.dl.lezhin.Application.main(Application.java:73)
Caused by: org.openqa.selenium.NoSuchElementException: no such element: Unable to locate element: {"method":"xpath","selector":"//script[not(@src) and contains(text(), '__LZ_ME__')]"}
(Session info: headless chrome=107.0.5304.108)
For documentation on this error, please visit: https://selenium.dev/exceptions/#no_such_element
Build info: version: '4.6.0', revision: '79f1c02ae20'
System info: os.name: 'Windows 10', os.arch: 'amd64', os.version: '10.0', java.version: '15.0.7'
Driver info: org.openqa.selenium.chrome.ChromeDriver
Command: [de5624873d2b4a72ae1bdaca9b137fda, findElement {using=xpath, value=//script[not(@src) and contains(text(), '__LZ_ME__')]}]
Capabilities {acceptInsecureCerts: false, browserName: chrome, browserVersion: 107.0.5304.108, chrome: {chromedriverVersion: 107.0.5304.62
```
### Operating System
Windows 10
### Selenium version
Java 4.6.0
### What are the browser(s) and version(s) where you see this issue?
Chrome 107.0.5304.108
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 107.0.5304.62
### Are you using Selenium Grid?
No
|
defect
|
remotewebdriver findelement works weirdly what happened when i tried to find an element by xpath it works the current call stack of method was done and then jvm went into other method call stack i tried to do that again at there but it doesn t work there is no action for chromedriver the changes are only call stack i don t understand why java driver findelement by xpath script how can we reproduce the issue shell clone on branch dev write down your login info of lezhin comics to config ini run io github imsejin dl lezhin application on debug mode compare loginprocessor getaccesstoken with accesstokenprocessor getaccesstoken relevant log output shell starting chromedriver refs branch heads on port only local connections are allowed please see for suggestions on keeping chromedriver safe chromedriver was started successfully exception in thread main java lang runtimeexception cannot find access token at io github imsejin dl lezhin process impl accesstokenprocessor getaccesstoken accesstokenprocessor java at io github imsejin dl lezhin process impl accesstokenprocessor process accesstokenprocessor java at io github imsejin dl lezhin process impl accesstokenprocessor process accesstokenprocessor java at io github imsejin dl lezhin application main application java caused by org openqa selenium nosuchelementexception no such element unable to locate element method xpath selector script session info headless chrome for documentation on this error please visit build info version revision system info os name windows os arch os version java version driver info org openqa selenium chrome chromedriver command capabilities acceptinsecurecerts false browsername chrome browserversion chrome chromedriverversion operating system windows selenium version java what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver are you using selenium grid no
| 1
|
350,196
| 10,479,997,577
|
IssuesEvent
|
2019-09-24 06:22:11
|
wso2/product-apim
|
https://api.github.com/repos/wso2/product-apim
|
opened
|
[3.0.0] URL based swagger import is missing in API definitions page
|
3.0.0-alpha Priority/Highest Publisher Severity/Critical
|
URL and File-based swagger import option should be provided in API definitions page

|
1.0
|
[3.0.0] URL based swagger import is missing in API definitions page - URL and File-based swagger import option should be provided in API definitions page

|
non_defect
|
url based swagger import is missing in api definitions page url and file based swagger import option should be provided in api definitions page
| 0
|
13,523
| 3,737,267,144
|
IssuesEvent
|
2016-03-08 18:41:07
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Doc Fixit: node.js update website doc reference
|
documentation node
|
http://www.grpc.io/grpc/node/ still says 0.12.0
Navigation: grpc.io -> Documentation -> References -> Node.js API
|
1.0
|
Doc Fixit: node.js update website doc reference - http://www.grpc.io/grpc/node/ still says 0.12.0
Navigation: grpc.io -> Documentation -> References -> Node.js API
|
non_defect
|
doc fixit node js update website doc reference still says navigation grpc io documentation references node js api
| 0
|
422,301
| 28,433,801,489
|
IssuesEvent
|
2023-04-15 04:06:51
|
erritis/mongodb-macro
|
https://api.github.com/repos/erritis/mongodb-macro
|
opened
|
Add multiple factory examples to resources
|
documentation
|
Add examples of creating multiple factories to different servers, databases and collections to better demonstrate the purpose of creating this crate.
|
1.0
|
Add multiple factory examples to resources - Add examples of creating multiple factories to different servers, databases and collections to better demonstrate the purpose of creating this crate.
|
non_defect
|
add multiple factory examples to resources add examples of creating multiple factories to different servers databases and collections to better demonstrate the purpose of creating this crate
| 0
|
663,139
| 22,162,783,683
|
IssuesEvent
|
2022-06-04 19:05:24
|
TuSimple/naive-ui
|
https://api.github.com/repos/TuSimple/naive-ui
|
closed
|
Suggestions for improving Input Number component
|
feature request priority: low
|
<!-- generated by issue-helper DO NOT REMOVE __FEATURE_REQUEST__ -->
### This function solves the problem (这个功能解决的问题)
Thanks a lot for developing this fantastic UI library :)
Currently the Input Number component is a bit limited in its feature range imho.
It does not accept decimal separators for other locales (e.g. in Germany a comma is used for separating decimal values instead of a dot).
At the moment the component uses a text input under the hood. An input of type number would detect the correct locale automatically if it's set on the html tag (e.g. `<html lang="de">`) and a comma as decimal separator would be accepted.
I think it's safe to assume that most people would expect the "Input Number" to behave like an input of type number. If there is a technical reason that makes it necessary to use a text input, I would suggest making it possible to pass a locale or a custom decimal separator / a list of custom decimal separators to the component.
Also it would be great if it was possible to limit the decimal places to a fixed number, so that the user can't enter decimals with longer decimal places. This would be very useful for price inputs especially. In that case it would also be great to add a zero to the decimal place. E.g. 34.5 would become 34.50.
### Expected API (期望的 API)
decimal-separators `string[]` - custom separators allowed for decimal numbers
decimal-places `number` - the maximum length of decimal places for the number (zeroes will be added automatically to fill the decimal places)
`<n-input-number ... :decimal-separators="[',']" :decimal-places="2" />`
<!-- generated by issue-helper DO NOT REMOVE __FEATURE_REQUEST__ -->
|
1.0
|
Suggestions for improving Input Number component - <!-- generated by issue-helper DO NOT REMOVE __FEATURE_REQUEST__ -->
### This function solves the problem (这个功能解决的问题)
Thanks a lot for developing this fantastic UI library :)
Currently the Input Number component is a bit limited in its feature range imho.
It does not accept decimal separators for other locales (e.g. in Germany a comma is used for separating decimal values instead of a dot).
At the moment the component uses a text input under the hood. An input of type number would detect the correct locale automatically if it's set on the html tag (e.g. `<html lang="de">`) and a comma as decimal separator would be accepted.
I think it's safe to assume that most people would expect the "Input Number" to behave like an input of type number. If there is a technical reason that makes it necessary to use a text input, I would suggest making it possible to pass a locale or a custom decimal separator / a list of custom decimal separators to the component.
Also it would be great if it was possible to limit the decimal places to a fixed number, so that the user can't enter decimals with longer decimal places. This would be very useful for price inputs especially. In that case it would also be great to add a zero to the decimal place. E.g. 34.5 would become 34.50.
### Expected API (期望的 API)
decimal-separators `string[]` - custom separators allowed for decimal numbers
decimal-places `number` - the maximum length of decimal places for the number (zeroes will be added automatically to fill the decimal places)
`<n-input-number ... :decimal-separators="[',']" :decimal-places="2" />`
<!-- generated by issue-helper DO NOT REMOVE __FEATURE_REQUEST__ -->
|
non_defect
|
suggestions for improving input number component this function solves the problem 这个功能解决的问题 thanks a lot for developing this fantastic ui library currently the input number component is a bit limited in its feature range imho it does not accept decimal separators for other locales e g in germany a comma is used for separating decimal values instead of a dot at the moment the component uses a text input under the hood an input of type number would detect the correct locale automatically if it s set on the html tag e g and a comma as decimal separator would be accepted i think it s safe to assume that most people would expect the input number to behave like an input of type number if there is a technical reason that makes it necessary to use a text input i would suggest making it possible to pass a locale or a custom decimal separator a list of custom decimal separators to the component also it would be great if it was possible to limit the decimal places to a fixed number so that the user can t enter decimals with longer decimal places this would be very useful for price inputs especially in that case it would also be great to add a zero to the decimal place e g would become expected api 期望的 api decimal separators string custom separators allowed for decimal numbers decimal places number the maximum length of decimal places for the number zeroes will be added automatically to fill the decimal places
| 0
|
36,651
| 8,048,825,374
|
IssuesEvent
|
2018-08-01 08:11:25
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
Dialog/OverlayPanel error when using appendTo
|
defect
|
**I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
```
**Plunkr Case (Bug Reports)**
https://github-df7qmz.stackblitz.io/
**Current behavior**
When the dialog is not visible, hiding its containing element via ngIf (causing dialog OnDestroy) causes an error when using the appendTo attribute
**Expected behavior**
I'm not sure exactly what is supposed to happen here, but not the error of course :)
Maybe old code that tries to put the dialog back in its "non-body appended" place but with the dialog getting removed from the DOM in 6.0.2 that isn't a thing anymore. Anyway, you guys get it.
**Minimal reproduction of the problem with instructions**
In stackblitz, show the dialog. Toggle the container back and forth, everything is fine. Hide the dialog, and as soon as you hide the container you get the error.
ERROR TypeError: Failed to execute 'appendChild' on 'Node': parameter 1 is not of type 'Node'.
at Dialog.restoreAppend (dialog.ts:555)
at Dialog.ngOnDestroy (dialog.ts:590)
|
1.0
|
Dialog/OverlayPanel error when using appendTo - **I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
```
**Plunkr Case (Bug Reports)**
https://github-df7qmz.stackblitz.io/
**Current behavior**
When the dialog is not visible, hiding its containing element via ngIf (causing dialog OnDestroy) causes an error when using the appendTo attribute
**Expected behavior**
I'm not sure exactly what is supposed to happen here, but not the error of course :)
Maybe old code that tries to put the dialog back in its "non-body appended" place but with the dialog getting removed from the DOM in 6.0.2 that isn't a thing anymore. Anyway, you guys get it.
**Minimal reproduction of the problem with instructions**
In stackblitz, show the dialog. Toggle the container back and forth, everything is fine. Hide the dialog, and as soon as you hide the container you get the error.
ERROR TypeError: Failed to execute 'appendChild' on 'Node': parameter 1 is not of type 'Node'.
at Dialog.restoreAppend (dialog.ts:555)
at Dialog.ngOnDestroy (dialog.ts:590)
|
defect
|
dialog overlaypanel error when using appendto i m submitting a check one with x bug report search github for a similar issue or pr before submitting plunkr case bug reports current behavior when the dialog is not visible hiding its containing element via ngif causing dialog ondestroy causes an error when using the appendto attribute expected behavior i m not sure exactly what is supposed to happen here but not the error of course maybe old code that tries to put the dialog back in its non body appended place but with the dialog getting removed from the dom in that isn t a thing anymore anyway you guys get it minimal reproduction of the problem with instructions in stackblitz show the dialog toggle the container back and forth everything is fine hide the dialog and as soon as you hide the container you get the error error typeerror failed to execute appendchild on node parameter is not of type node at dialog restoreappend dialog ts at dialog ngondestroy dialog ts
| 1
|
269,012
| 28,959,965,143
|
IssuesEvent
|
2023-05-10 01:04:08
|
dpteam/RK3188_TABLET
|
https://api.github.com/repos/dpteam/RK3188_TABLET
|
reopened
|
CVE-2013-7264 (Medium) detected in linuxv3.0
|
Mend: dependency security vulnerability
|
## CVE-2013-7264 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (6)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/phonet/datagram.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ieee802154/dgram.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/phonet/datagram.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv4/udp.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/phonet/datagram.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ieee802154/dgram.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The l2tp_ip_recvmsg function in net/l2tp/l2tp_ip.c in the Linux kernel before 3.12.4 updates a certain length value before ensuring that an associated data structure has been initialized, which allows local users to obtain sensitive information from kernel stack memory via a (1) recvfrom, (2) recvmmsg, or (3) recvmsg system call.
<p>Publish Date: 2014-01-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2013-7264>CVE-2013-7264</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2013-7264">https://nvd.nist.gov/vuln/detail/CVE-2013-7264</a></p>
<p>Release Date: 2014-01-06</p>
<p>Fix Resolution: 3.12.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2013-7264 (Medium) detected in linuxv3.0 - ## CVE-2013-7264 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (6)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/phonet/datagram.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ieee802154/dgram.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/phonet/datagram.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv4/udp.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/phonet/datagram.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ieee802154/dgram.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The l2tp_ip_recvmsg function in net/l2tp/l2tp_ip.c in the Linux kernel before 3.12.4 updates a certain length value before ensuring that an associated data structure has been initialized, which allows local users to obtain sensitive information from kernel stack memory via a (1) recvfrom, (2) recvmmsg, or (3) recvmsg system call.
<p>Publish Date: 2014-01-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2013-7264>CVE-2013-7264</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2013-7264">https://nvd.nist.gov/vuln/detail/CVE-2013-7264</a></p>
<p>Release Date: 2014-01-06</p>
<p>Fix Resolution: 3.12.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files net phonet datagram c net dgram c net phonet datagram c net udp c net phonet datagram c net dgram c vulnerability details the ip recvmsg function in net ip c in the linux kernel before updates a certain length value before ensuring that an associated data structure has been initialized which allows local users to obtain sensitive information from kernel stack memory via a recvfrom recvmmsg or recvmsg system call publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
42,161
| 12,879,018,176
|
IssuesEvent
|
2020-07-11 19:41:35
|
loftwah/loftwah-dev-2020
|
https://api.github.com/repos/loftwah/loftwah-dev-2020
|
opened
|
CVE-2018-20821 (Medium) detected in node-sass-4.14.1.tgz, node-sass-v4.13.1
|
security vulnerability
|
## CVE-2018-20821 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/loftwah-dev-2020/wordpress/app/wp-content/themes/twentynineteen/package.json</p>
<p>Path to vulnerable library: /loftwah-dev-2020/wordpress/app/wp-content/themes/twentynineteen/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/loftwah/loftwah-dev-2020/commit/89fd212689151e9f7ffbaa73a2adacf222367b48">89fd212689151e9f7ffbaa73a2adacf222367b48</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-20821 (Medium) detected in node-sass-4.14.1.tgz, node-sass-v4.13.1 - ## CVE-2018-20821 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/loftwah-dev-2020/wordpress/app/wp-content/themes/twentynineteen/package.json</p>
<p>Path to vulnerable library: /loftwah-dev-2020/wordpress/app/wp-content/themes/twentynineteen/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/loftwah/loftwah-dev-2020/commit/89fd212689151e9f7ffbaa73a2adacf222367b48">89fd212689151e9f7ffbaa73a2adacf222367b48</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in node sass tgz node sass cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm loftwah dev wordpress app wp content themes twentynineteen package json path to vulnerable library loftwah dev wordpress app wp content themes twentynineteen node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details the parsing component in libsass through allows attackers to cause a denial of service uncontrolled recursion in sass parser parse css variable value in parser cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
| 0
|
59,542
| 17,023,157,151
|
IssuesEvent
|
2021-07-03 00:37:51
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Slippy map doesn't work in IE6
|
Component: rails_port Priority: major Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 4.52pm, Monday, 7th May 2007]**
IE6 doesn't work with slippy map - there's no layer selector visible in
top right, and no tiles appear to load, leaving a plain white screen.
The map boundary box and the pan/zoom controls at top left do show up, however, so that's progress, as they didn't before.
|
1.0
|
Slippy map doesn't work in IE6 - **[Submitted to the original trac issue database at 4.52pm, Monday, 7th May 2007]**
IE6 doesn't work with slippy map - there's no layer selector visible in
top right, and no tiles appear to load, leaving a plain white screen.
The map boundary box and the pan/zoom controls at top left do show up, however, so that's progress, as they didn't before.
|
defect
|
slippy map doesn t work in doesn t work with slippy map there s no layer selector visible in top right and no tiles appear to load leaving a plain white screen the map boundary box and the pan zoom controls at top left do show up however so that s progress as they didn t before
| 1
|
433,804
| 12,510,915,458
|
IssuesEvent
|
2020-06-02 19:33:17
|
CDH-Studio/UpSkill
|
https://api.github.com/repos/CDH-Studio/UpSkill
|
closed
|
User name overlaps avatar in my profile
|
Low Priority UI bug
|
**Describe the bug**
User name in profile view's BasicInfoView overlaps avatar
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'My profile'
**Expected behavior**
The name should not overlap the avatar
**Screenshots**
As you can see, Ben is over the avatar

|
1.0
|
User name overlaps avatar in my profile - **Describe the bug**
User name in profile view's BasicInfoView overlaps avatar
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'My profile'
**Expected behavior**
The name should not overlap the avatar
**Screenshots**
As you can see, Ben is over the avatar

|
non_defect
|
user name overlaps avatar in my profile describe the bug user name in profile view s basicinfoview overlaps avatar to reproduce steps to reproduce the behavior go to my profile expected behavior the name should not overlap the avatar screenshots as you can see ben is over the avatar
| 0
|
52,793
| 13,225,069,690
|
IssuesEvent
|
2020-08-17 20:25:40
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'> (Trac #348)
|
Migrated from Trac combo simulation defect
|
I can't see the revision log for romeo trunk
==== How to Reproduce ====
While doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.
''(please provide additional details here)''
Request parameters:
```text
{'path': u'/projects/romeo/trunk'}
```
User Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`
==== System Information ====
|| '''Trac''' || `0.11.1` ||
|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||
|| '''setuptools''' || `0.6` ||
|| '''psycopg2''' || `2.2.1` ||
|| '''Genshi''' || `0.6` ||
|| '''mod_python''' || `3.3.1` ||
|| '''Pygments''' || `0.10` ||
|| '''Subversion''' || `1.5.1 (r32289)` ||
|| '''jQuery:''' || `1.2.6` ||
==== Python Traceback ====
```text
Traceback (most recent call last):
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py", line 423, in _dispatch_request
dispatcher.dispatch(req)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py", line 219, in dispatch
data, content_type)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 715, in render_template
return stream.render(method, doctype=doctype)
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 183, in render
return encode(generator, method=method, encoding=encoding, out=out)
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 57, in encode
return _encode(''.join(list(iterator)))
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 339, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 826, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 670, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 771, in __call__
for kind, data, pos in chain(stream, [(None, None, None)]):
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 586, in __call__
for ev in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 767, in _strip_accesskeys
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 756, in _generate
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 378, in _match
ctxt, start=idx + 1, **vars):
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 378, in _match
ctxt, start=idx + 1, **vars):
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 327, in _match
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 545, in _flatten
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/path.py", line 588, in _generate
subevent = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 316, in _strip
event = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 545, in _flatten
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/path.py", line 588, in _generate
subevent = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 316, in _strip
event = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 565, in _flatten
result = _eval_expr(data, ctxt, vars)
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 277, in _eval_expr
retval = expr.evaluate(ctxt)
File "/usr/lib/pymodules/python2.5/genshi/template/eval.py", line 178, in evaluate
return eval(self.code, _globals, {'__data__': data})
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html", line 134, in <Expression u'dateinfo(change.date)'>
<td class="date" py:content="dateinfo(change.date)" />
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py", line 230, in dateinfo
pretty_timedelta(date),
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py", line 71, in pretty_timedelta
time1 = to_datetime(time1)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py", line 51, in to_datetime
type(t))
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/348">https://code.icecube.wisc.edu/projects/icecube/ticket/348</a>, reported by maseand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T20:08:32",
"_ts": "1547237312321270",
"description": "I can't see the revision log for romeo trunk\n\n\n==== How to Reproduce ====\n\nWhile doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.\n\n''(please provide additional details here)''\n\n\nRequest parameters:\n{{{\n{'path': u'/projects/romeo/trunk'}\n}}}\n\n\nUser Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`\n\n==== System Information ====\n\n|| '''Trac''' || `0.11.1` ||\n|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||\n|| '''setuptools''' || `0.6` ||\n|| '''psycopg2''' || `2.2.1` ||\n|| '''Genshi''' || `0.6` ||\n|| '''mod_python''' || `3.3.1` ||\n|| '''Pygments''' || `0.10` ||\n|| '''Subversion''' || `1.5.1 (r32289)` ||\n|| '''jQuery:''' || `1.2.6` ||\n\n==== Python Traceback ====\n{{{\nTraceback (most recent call last):\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 423, in _dispatch_request\n dispatcher.dispatch(req)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 219, in dispatch\n data, content_type)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 715, in render_template\n return stream.render(method, doctype=doctype)\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 183, in render\n return encode(generator, method=method, encoding=encoding, out=out)\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 57, in encode\n return _encode(''.join(list(iterator)))\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 339, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 826, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 670, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 771, in __call__\n for kind, data, pos in chain(stream, [(None, None, None)]):\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 586, in __call__\n for ev in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 767, in _strip_accesskeys\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 756, in _generate\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 327, in _match\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 565, in _flatten\n result = _eval_expr(data, ctxt, vars)\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 277, in _eval_expr\n retval = expr.evaluate(ctxt)\n File \"/usr/lib/pymodules/python2.5/genshi/template/eval.py\", line 178, in evaluate\n return eval(self.code, _globals, {'__data__': data})\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html\", line 134, in <Expression u'dateinfo(change.date)'>\n <td class=\"date\" py:content=\"dateinfo(change.date)\" />\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py\", line 230, in dateinfo\n pretty_timedelta(date),\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 71, in pretty_timedelta\n time1 = to_datetime(time1)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 51, in to_datetime\n type(t))\nTypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>\n\n}}}\n ",
"reporter": "mase",
"cc": "",
"resolution": "fixed",
"time": "2012-01-30T03:41:32",
"component": "combo simulation",
"summary": "TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'> (Trac #348) - I can't see the revision log for romeo trunk
==== How to Reproduce ====
While doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.
''(please provide additional details here)''
Request parameters:
```text
{'path': u'/projects/romeo/trunk'}
```
User Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`
==== System Information ====
|| '''Trac''' || `0.11.1` ||
|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||
|| '''setuptools''' || `0.6` ||
|| '''psycopg2''' || `2.2.1` ||
|| '''Genshi''' || `0.6` ||
|| '''mod_python''' || `3.3.1` ||
|| '''Pygments''' || `0.10` ||
|| '''Subversion''' || `1.5.1 (r32289)` ||
|| '''jQuery:''' || `1.2.6` ||
==== Python Traceback ====
```text
Traceback (most recent call last):
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py", line 423, in _dispatch_request
dispatcher.dispatch(req)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py", line 219, in dispatch
data, content_type)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 715, in render_template
return stream.render(method, doctype=doctype)
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 183, in render
return encode(generator, method=method, encoding=encoding, out=out)
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 57, in encode
return _encode(''.join(list(iterator)))
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 339, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 826, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 670, in __call__
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 771, in __call__
for kind, data, pos in chain(stream, [(None, None, None)]):
File "/usr/lib/pymodules/python2.5/genshi/output.py", line 586, in __call__
for ev in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 767, in _strip_accesskeys
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py", line 756, in _generate
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 378, in _match
ctxt, start=idx + 1, **vars):
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 378, in _match
ctxt, start=idx + 1, **vars):
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 327, in _match
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 545, in _flatten
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/path.py", line 588, in _generate
subevent = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 316, in _strip
event = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 545, in _flatten
for kind, data, pos in stream:
File "/usr/lib/pymodules/python2.5/genshi/core.py", line 288, in _ensure
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/path.py", line 588, in _generate
subevent = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 605, in _include
for event in stream:
File "/usr/lib/pymodules/python2.5/genshi/template/markup.py", line 316, in _strip
event = next()
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 565, in _flatten
result = _eval_expr(data, ctxt, vars)
File "/usr/lib/pymodules/python2.5/genshi/template/base.py", line 277, in _eval_expr
retval = expr.evaluate(ctxt)
File "/usr/lib/pymodules/python2.5/genshi/template/eval.py", line 178, in evaluate
return eval(self.code, _globals, {'__data__': data})
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html", line 134, in <Expression u'dateinfo(change.date)'>
<td class="date" py:content="dateinfo(change.date)" />
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py", line 230, in dateinfo
pretty_timedelta(date),
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py", line 71, in pretty_timedelta
time1 = to_datetime(time1)
File "/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py", line 51, in to_datetime
type(t))
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/348">https://code.icecube.wisc.edu/projects/icecube/ticket/348</a>, reported by maseand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-11T20:08:32",
"_ts": "1547237312321270",
"description": "I can't see the revision log for romeo trunk\n\n\n==== How to Reproduce ====\n\nWhile doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.\n\n''(please provide additional details here)''\n\n\nRequest parameters:\n{{{\n{'path': u'/projects/romeo/trunk'}\n}}}\n\n\nUser Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`\n\n==== System Information ====\n\n|| '''Trac''' || `0.11.1` ||\n|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||\n|| '''setuptools''' || `0.6` ||\n|| '''psycopg2''' || `2.2.1` ||\n|| '''Genshi''' || `0.6` ||\n|| '''mod_python''' || `3.3.1` ||\n|| '''Pygments''' || `0.10` ||\n|| '''Subversion''' || `1.5.1 (r32289)` ||\n|| '''jQuery:''' || `1.2.6` ||\n\n==== Python Traceback ====\n{{{\nTraceback (most recent call last):\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 423, in _dispatch_request\n dispatcher.dispatch(req)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 219, in dispatch\n data, content_type)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 715, in render_template\n return stream.render(method, doctype=doctype)\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 183, in render\n return encode(generator, method=method, encoding=encoding, out=out)\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 57, in encode\n return _encode(''.join(list(iterator)))\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 339, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 826, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 670, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 771, in __call__\n for kind, data, pos in chain(stream, [(None, None, None)]):\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 586, in __call__\n for ev in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 767, in _strip_accesskeys\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 756, in _generate\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 327, in _match\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 565, in _flatten\n result = _eval_expr(data, ctxt, vars)\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 277, in _eval_expr\n retval = expr.evaluate(ctxt)\n File \"/usr/lib/pymodules/python2.5/genshi/template/eval.py\", line 178, in evaluate\n return eval(self.code, _globals, {'__data__': data})\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html\", line 134, in <Expression u'dateinfo(change.date)'>\n <td class=\"date\" py:content=\"dateinfo(change.date)\" />\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py\", line 230, in dateinfo\n pretty_timedelta(date),\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 71, in pretty_timedelta\n time1 = to_datetime(time1)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 51, in to_datetime\n type(t))\nTypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>\n\n}}}\n ",
"reporter": "mase",
"cc": "",
"resolution": "fixed",
"time": "2012-01-30T03:41:32",
"component": "combo simulation",
"summary": "TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
typeerror expecting datetime int long float or none got trac i can t see the revision log for romeo trunk how to reproduce while doing a get operation on log projects romeo trunk trac issued an internal error please provide additional details here request parameters text path u projects romeo trunk user agent was mozilla u linux ja rv gecko ubuntu maverick firefox system information trac python nov setuptools genshi mod python pygments subversion jquery python traceback text traceback most recent call last file opt trac local lib site packages trac egg trac web main py line in dispatch request dispatcher dispatch req file opt trac local lib site packages trac egg trac web main py line in dispatch data content type file opt trac local lib site packages trac egg trac web chrome py line in render template return stream render method doctype doctype file usr lib pymodules genshi core py line in render return encode generator method method encoding encoding out out file usr lib pymodules genshi output py line in encode return encode join list iterator file usr lib pymodules genshi output py line in call for kind data pos in stream file usr lib pymodules genshi output py line in call for kind data pos in stream file usr lib pymodules genshi output py line in call for kind data pos in stream file usr lib pymodules genshi output py line in call for kind data pos in chain stream file usr lib pymodules genshi output py line in call for ev in stream file usr lib pymodules genshi core py line in ensure for event in stream file usr lib pymodules genshi core py line in ensure for event in stream file opt trac local lib site packages trac egg trac web chrome py line in strip accesskeys for kind data pos in stream file usr lib pymodules genshi core py line in ensure for event in stream file opt trac local lib site packages trac egg trac web chrome py line in generate for kind data pos in stream file usr lib pymodules genshi template base py line in include for event in stream file usr lib pymodules genshi template markup py line in match ctxt start idx vars file usr lib pymodules genshi template markup py line in match ctxt start idx vars file usr lib pymodules genshi template markup py line in match for event in stream file usr lib pymodules genshi template base py line in flatten for kind data pos in stream file usr lib pymodules genshi core py line in ensure for event in stream file usr lib pymodules genshi path py line in generate subevent next file usr lib pymodules genshi template base py line in include for event in stream file usr lib pymodules genshi template markup py line in strip event next file usr lib pymodules genshi template base py line in flatten for kind data pos in stream file usr lib pymodules genshi core py line in ensure for event in stream file usr lib pymodules genshi path py line in generate subevent next file usr lib pymodules genshi template base py line in include for event in stream file usr lib pymodules genshi template markup py line in strip event next file usr lib pymodules genshi template base py line in flatten result eval expr data ctxt vars file usr lib pymodules genshi template base py line in eval expr retval expr evaluate ctxt file usr lib pymodules genshi template eval py line in evaluate return eval self code globals data data file opt trac local lib site packages trac egg trac versioncontrol templates revisionlog html line in file opt trac local lib site packages trac egg trac timeline web ui py line in dateinfo pretty timedelta date file opt trac local lib site packages trac egg trac util datefmt py line in pretty timedelta to datetime file opt trac local lib site packages trac egg trac util datefmt py line in to datetime type t typeerror expecting datetime int long float or none got migrated from json status closed changetime ts description i can t see the revision log for romeo trunk n n n how to reproduce n nwhile doing a get operation on log projects romeo trunk trac issued an internal error n n please provide additional details here n n nrequest parameters n n path u projects romeo trunk n n n nuser agent was mozilla u linux ja rv gecko ubuntu maverick firefox n n system information n n trac n python nov n setuptools n n genshi n mod python n pygments n subversion n jquery n n python traceback n ntraceback most recent call last n file opt trac local lib site packages trac egg trac web main py line in dispatch request n dispatcher dispatch req n file opt trac local lib site packages trac egg trac web main py line in dispatch n data content type n file opt trac local lib site packages trac egg trac web chrome py line in render template n return stream render method doctype doctype n file usr lib pymodules genshi core py line in render n return encode generator method method encoding encoding out out n file usr lib pymodules genshi output py line in encode n return encode join list iterator n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in chain stream n file usr lib pymodules genshi output py line in call n for ev in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file opt trac local lib site packages trac egg trac web chrome py line in strip accesskeys n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file opt trac local lib site packages trac egg trac web chrome py line in generate n for kind data pos in stream n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in match n ctxt start idx vars n file usr lib pymodules genshi template markup py line in match n ctxt start idx vars n file usr lib pymodules genshi template markup py line in match n for event in stream n file usr lib pymodules genshi template base py line in flatten n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi path py line in generate n subevent next n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in strip n event next n file usr lib pymodules genshi template base py line in flatten n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi path py line in generate n subevent next n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in strip n event next n file usr lib pymodules genshi template base py line in flatten n result eval expr data ctxt vars n file usr lib pymodules genshi template base py line in eval expr n retval expr evaluate ctxt n file usr lib pymodules genshi template eval py line in evaluate n return eval self code globals data data n file opt trac local lib site packages trac egg trac versioncontrol templates revisionlog html line in n n file opt trac local lib site packages trac egg trac timeline web ui py line in dateinfo n pretty timedelta date n file opt trac local lib site packages trac egg trac util datefmt py line in pretty timedelta n to datetime n file opt trac local lib site packages trac egg trac util datefmt py line in to datetime n type t ntypeerror expecting datetime int long float or none got n n n reporter mase cc resolution fixed time component combo simulation summary typeerror expecting datetime int long float or none got priority normal keywords milestone owner nega type defect
| 1
|
320,505
| 27,438,765,756
|
IssuesEvent
|
2023-03-02 09:31:35
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
[Flaky Test] can be converted when nested to paragraphs
|
[Type] Flaky Test
|
<!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
can be converted when nested to paragraphs
## Test path
`/test/e2e/specs/editor/blocks/list.spec.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2023-02-27T18:34:39.875Z"><code>[2023-02-27T18:34:39.875Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4285127412"><code>try/something</code></a>.
</summary>
```
locator.click: Timeout 10000ms exceeded.
=========================== logs ===========================
waiting for locator('role=toolbar[name="Block tools"i]').locator('role=button[name="Indent"]')
============================================================
at Editor.clickBlockToolbarButton (/home/runner/work/gutenberg/gutenberg/packages/e2e-test-utils-playwright/src/editor/click-block-toolbar-button.ts:20:15)
at /home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/blocks/list.spec.js:337:3
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
|
1.0
|
[Flaky Test] can be converted when nested to paragraphs - <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
can be converted when nested to paragraphs
## Test path
`/test/e2e/specs/editor/blocks/list.spec.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2023-02-27T18:34:39.875Z"><code>[2023-02-27T18:34:39.875Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4285127412"><code>try/something</code></a>.
</summary>
```
locator.click: Timeout 10000ms exceeded.
=========================== logs ===========================
waiting for locator('role=toolbar[name="Block tools"i]').locator('role=button[name="Indent"]')
============================================================
at Editor.clickBlockToolbarButton (/home/runner/work/gutenberg/gutenberg/packages/e2e-test-utils-playwright/src/editor/click-block-toolbar-button.ts:20:15)
at /home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/blocks/list.spec.js:337:3
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
|
non_defect
|
can be converted when nested to paragraphs flaky test detected this is an auto generated issue by github actions please do not edit this manually test title can be converted when nested to paragraphs test path test specs editor blocks list spec js errors test passed after failed attempt on a href locator click timeout exceeded logs waiting for locator role toolbar locator role button at editor clickblocktoolbarbutton home runner work gutenberg gutenberg packages test utils playwright src editor click block toolbar button ts at home runner work gutenberg gutenberg test specs editor blocks list spec js
| 0
|
16,000
| 2,870,250,565
|
IssuesEvent
|
2015-06-07 00:34:38
|
pdelia/away3d
|
https://api.github.com/repos/pdelia/away3d
|
closed
|
TextField3D render problems, text overwrite problem, and crashes
|
auto-migrated Priority-Medium Type-Defect
|
#52 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:25Z
```
What steps will reproduce the problem?
1. Create TextField3D object and apply ColorMaterial
2. Create Trident
3. Add both to scene
4. Every frame, change the text property of TextField3D object
What is the expected output? What do you see instead?
a) Severe rendering problems, similar to Issue 49. (Extraneous shapes
rendered to the right of the TextField3D object, depending on view angle.)
b) Run-time crash: TypeError: Error #1009: Cannot access a property or
method of a null object reference. At away3d.core.base::Geometry/removeFace()
What version of the product are you using? On what operating system?
Away3D trunk as of 9 Aug 2009 (rev 1594).
Flash CS3 IDE.
FP9.
XP SR3.
Please provide any additional information below.
a) See attached renderBug1a.fla and renderBug1a.swf.
b) Run time error seems to be related to changing the text property. Does
not crash on every text change - seems to depend on the new string somehow.
c) Rendering errors seem to be related to use of ColorMaterial and FP9.
Renders OK with FP10.
d) Script fragments follow:
var yellowTextMaterial:ColorMaterial = new ColorMaterial(0xFFFF00);
var textfield:TextField3D = new TextField3D("Arial",
{material:yellowTextMaterial, text:"Hello!", size:100, leading:500,
kerning:0, textWidth:100000, align:"TL"});
function onEnterFrame(event:Event):void {
frameCount++;
var str = "frameCount = "+frameCount;
textfield.text = str;
}
e) Crash output details follow:
TypeError: Error #1009: Cannot access a property or method of a null object
reference.
at away3d.core.base::Geometry/removeFace()
at away3d.core.base::Mesh/removeFace()
at away3d.primitives::AbstractPrimitive/buildPrimitive()
at away3d.primitives::TextField3D/buildPrimitive()
at away3d.primitives::AbstractPrimitive/updatePrimitive()
at away3d.primitives::AbstractPrimitive/get geometry()
at away3d.core.traverse::ProjectionTraverser/apply()
at away3d.core.base::Object3D/traverse()
at away3d.containers::ObjectContainer3D/traverse()
at away3d.containers::Scene3D/update()
at away3d.containers::Scene3D/onUpdate()
at flash.events::EventDispatcher/dispatchEventFunction()
at flash.events::EventDispatcher/dispatchEvent()
at away3d.containers::View3D/notifySceneUpdate()
at away3d.containers::View3D/render()
at renderBug1a_fla::MainTimeline/onEnterFrame()
```
Original issue reported on code.google.com by `ralph%sa...@gtempaccount.com` on 9 Aug 2009 at 9:51
Attachments:
* [renderBug1a.fla](https://storage.googleapis.com/google-code-attachments/away3d/issue-52/comment-0/renderBug1a.fla)
* [renderBug1a.swf](https://storage.googleapis.com/google-code-attachments/away3d/issue-52/comment-0/renderBug1a.swf)
|
1.0
|
TextField3D render problems, text overwrite problem, and crashes - #52 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:25Z
```
What steps will reproduce the problem?
1. Create TextField3D object and apply ColorMaterial
2. Create Trident
3. Add both to scene
4. Every frame, change the text property of TextField3D object
What is the expected output? What do you see instead?
a) Severe rendering problems, similar to Issue 49. (Extraneous shapes
rendered to the right of the TextField3D object, depending on view angle.)
b) Run-time crash: TypeError: Error #1009: Cannot access a property or
method of a null object reference. At away3d.core.base::Geometry/removeFace()
What version of the product are you using? On what operating system?
Away3D trunk as of 9 Aug 2009 (rev 1594).
Flash CS3 IDE.
FP9.
XP SR3.
Please provide any additional information below.
a) See attached renderBug1a.fla and renderBug1a.swf.
b) Run time error seems to be related to changing the text property. Does
not crash on every text change - seems to depend on the new string somehow.
c) Rendering errors seem to be related to use of ColorMaterial and FP9.
Renders OK with FP10.
d) Script fragments follow:
var yellowTextMaterial:ColorMaterial = new ColorMaterial(0xFFFF00);
var textfield:TextField3D = new TextField3D("Arial",
{material:yellowTextMaterial, text:"Hello!", size:100, leading:500,
kerning:0, textWidth:100000, align:"TL"});
function onEnterFrame(event:Event):void {
frameCount++;
var str = "frameCount = "+frameCount;
textfield.text = str;
}
e) Crash output details follow:
TypeError: Error #1009: Cannot access a property or method of a null object
reference.
at away3d.core.base::Geometry/removeFace()
at away3d.core.base::Mesh/removeFace()
at away3d.primitives::AbstractPrimitive/buildPrimitive()
at away3d.primitives::TextField3D/buildPrimitive()
at away3d.primitives::AbstractPrimitive/updatePrimitive()
at away3d.primitives::AbstractPrimitive/get geometry()
at away3d.core.traverse::ProjectionTraverser/apply()
at away3d.core.base::Object3D/traverse()
at away3d.containers::ObjectContainer3D/traverse()
at away3d.containers::Scene3D/update()
at away3d.containers::Scene3D/onUpdate()
at flash.events::EventDispatcher/dispatchEventFunction()
at flash.events::EventDispatcher/dispatchEvent()
at away3d.containers::View3D/notifySceneUpdate()
at away3d.containers::View3D/render()
at renderBug1a_fla::MainTimeline/onEnterFrame()
```
Original issue reported on code.google.com by `ralph%sa...@gtempaccount.com` on 9 Aug 2009 at 9:51
Attachments:
* [renderBug1a.fla](https://storage.googleapis.com/google-code-attachments/away3d/issue-52/comment-0/renderBug1a.fla)
* [renderBug1a.swf](https://storage.googleapis.com/google-code-attachments/away3d/issue-52/comment-0/renderBug1a.swf)
|
defect
|
render problems text overwrite problem and crashes issue by googlecodeexporter created on what steps will reproduce the problem create object and apply colormaterial create trident add both to scene every frame change the text property of object what is the expected output what do you see instead a severe rendering problems similar to issue extraneous shapes rendered to the right of the object depending on view angle b run time crash typeerror error cannot access a property or method of a null object reference at core base geometry removeface what version of the product are you using on what operating system trunk as of aug rev flash ide xp please provide any additional information below a see attached fla and swf b run time error seems to be related to changing the text property does not crash on every text change seems to depend on the new string somehow c rendering errors seem to be related to use of colormaterial and renders ok with d script fragments follow var yellowtextmaterial colormaterial new colormaterial var textfield new arial material yellowtextmaterial text hello size leading kerning textwidth align tl function onenterframe event event void framecount var str framecount framecount textfield text str e crash output details follow typeerror error cannot access a property or method of a null object reference at core base geometry removeface at core base mesh removeface at primitives abstractprimitive buildprimitive at primitives buildprimitive at primitives abstractprimitive updateprimitive at primitives abstractprimitive get geometry at core traverse projectiontraverser apply at core base traverse at containers traverse at containers update at containers onupdate at flash events eventdispatcher dispatcheventfunction at flash events eventdispatcher dispatchevent at containers notifysceneupdate at containers render at fla maintimeline onenterframe original issue reported on code google com by ralph sa gtempaccount com on aug at attachments
| 1
|
54,692
| 13,884,978,541
|
IssuesEvent
|
2020-10-18 18:12:07
|
MDAnalysis/mdanalysis
|
https://api.github.com/repos/MDAnalysis/mdanalysis
|
closed
|
MultiPDBWriter is returned even if multiframe is False
|
Component-Writers Format-PDB defect
|
### Expected behaviour
When asking for a not multiframe PDB writer, the PDB writer should be a single frame writer.
### Actual behaviour
The writer is `MultiPDBWriter`.
The issue was uncovered by @abhinavgupta94 in #764.
### Code to reproduce the behaviour
``` python
import MDAnalysis as mda
writer = mda.Writer('test.pdb', multiframe=False)
```
### Currently version of MDAnalysis:
0.14.0-dev0
|
1.0
|
MultiPDBWriter is returned even if multiframe is False - ### Expected behaviour
When asking for a not multiframe PDB writer, the PDB writer should be a single frame writer.
### Actual behaviour
The writer is `MultiPDBWriter`.
The issue was uncovered by @abhinavgupta94 in #764.
### Code to reproduce the behaviour
``` python
import MDAnalysis as mda
writer = mda.Writer('test.pdb', multiframe=False)
```
### Currently version of MDAnalysis:
0.14.0-dev0
|
defect
|
multipdbwriter is returned even if multiframe is false expected behaviour when asking for a not multiframe pdb writer the pdb writer should be a single frame writer actual behaviour the writer is multipdbwriter the issue was uncovered by in code to reproduce the behaviour python import mdanalysis as mda writer mda writer test pdb multiframe false currently version of mdanalysis
| 1
|
321,285
| 27,520,241,665
|
IssuesEvent
|
2023-03-06 14:36:33
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix tensor.test_torch_instance_sort
|
PyTorch Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|
1.0
|
Fix tensor.test_torch_instance_sort - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|
non_defect
|
fix tensor test torch instance sort tensorflow img src torch img src numpy img src jax img src
| 0
|
71,188
| 30,825,531,940
|
IssuesEvent
|
2023-08-01 19:44:35
|
hashicorp/terraform-provider-azurerm
|
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
|
closed
|
azurerm_mysql_server if autogrowth happened a redeploy triggers unsupported shrink
|
bug service/mysql
|
Hi Community,
we are using azurerm_mysql_server to create MySQL managed instances in Azure.
We configured them with a starting size of 250GB and the nice feature "auto_grow_enabled = true".
When our PROD environment exceeded this size and Azure increased the max available storage, then every following plan will say that this property will be changed and the following applies fails.
While we could change our parameters for this database higher or equal to the current size, we would need to change settings we hoped to not care about. Also we don't want to increase the storage up-front because it is cheaper and also that's why auto_grow feature should be used for.
I propose the following:
If "auto_grow_enabled = true" and "storage_mb" set in code is smaller than existing instance, then ignore "storage_mb" or just set "storage_mb" to size of instance. Therefore no change will be applied when auto_grow happened on the instance.
`Error: waiting for MySQL Server "servername" (Resource Group "resourcegroupname") to finish updating: Code="UnsupportedUpdate" Message="Cannot update StorageMB Downscaling for server"`
|
1.0
|
azurerm_mysql_server if autogrowth happened a redeploy triggers unsupported shrink - Hi Community,
we are using azurerm_mysql_server to create MySQL managed instances in Azure.
We configured them with a starting size of 250GB and the nice feature "auto_grow_enabled = true".
When our PROD environment exceeded this size and Azure increased the max available storage, then every following plan will say that this property will be changed and the following applies fails.
While we could change our parameters for this database higher or equal to the current size, we would need to change settings we hoped to not care about. Also we don't want to increase the storage up-front because it is cheaper and also that's why auto_grow feature should be used for.
I propose the following:
If "auto_grow_enabled = true" and "storage_mb" set in code is smaller than existing instance, then ignore "storage_mb" or just set "storage_mb" to size of instance. Therefore no change will be applied when auto_grow happened on the instance.
`Error: waiting for MySQL Server "servername" (Resource Group "resourcegroupname") to finish updating: Code="UnsupportedUpdate" Message="Cannot update StorageMB Downscaling for server"`
|
non_defect
|
azurerm mysql server if autogrowth happened a redeploy triggers unsupported shrink hi community we are using azurerm mysql server to create mysql managed instances in azure we configured them with a starting size of and the nice feature auto grow enabled true when our prod environment exceeded this size and azure increased the max available storage then every following plan will say that this property will be changed and the following applies fails while we could change our parameters for this database higher or equal to the current size we would need to change settings we hoped to not care about also we don t want to increase the storage up front because it is cheaper and also that s why auto grow feature should be used for i propose the following if auto grow enabled true and storage mb set in code is smaller than existing instance then ignore storage mb or just set storage mb to size of instance therefore no change will be applied when auto grow happened on the instance error waiting for mysql server servername resource group resourcegroupname to finish updating code unsupportedupdate message cannot update storagemb downscaling for server
| 0
|
23,289
| 3,786,355,438
|
IssuesEvent
|
2016-03-21 01:48:12
|
DemNTor/varma_android_samples
|
https://api.github.com/repos/DemNTor/varma_android_samples
|
closed
|
WebServer cannot provide busy traffic(such as 1080p Video )
|
auto-migrated Priority-Medium Type-Defect
|
```
I take aws project as a example, build my WebServer. It works good for most
videos ,but cannot afford HD videos.
How can I speed up my webserver?
and what's more, my H.263 video cannot be played, can you give me some suggest?
My Video is 1920x1080,framerate 24, and another 1080x720 with framerate 24
works fine, and apache webserver can afford both 1920x1080 and H.263 videos.
```
Original issue reported on code.google.com by `d63...@gmail.com` on 20 May 2013 at 7:08
|
1.0
|
WebServer cannot provide busy traffic(such as 1080p Video ) - ```
I take aws project as a example, build my WebServer. It works good for most
videos ,but cannot afford HD videos.
How can I speed up my webserver?
and what's more, my H.263 video cannot be played, can you give me some suggest?
My Video is 1920x1080,framerate 24, and another 1080x720 with framerate 24
works fine, and apache webserver can afford both 1920x1080 and H.263 videos.
```
Original issue reported on code.google.com by `d63...@gmail.com` on 20 May 2013 at 7:08
|
defect
|
webserver cannot provide busy traffic such as video i take aws project as a example build my webserver it works good for most videos but cannot afford hd videos how can i speed up my webserver and what s more my h video cannot be played can you give me some suggest my video is framerate and another with framerate works fine and apache webserver can afford both and h videos original issue reported on code google com by gmail com on may at
| 1
|
562,473
| 16,661,686,697
|
IssuesEvent
|
2021-06-06 12:46:40
|
bounswe/2021SpringGroup4
|
https://api.github.com/repos/bounswe/2021SpringGroup4
|
opened
|
Implementation Assignment: Create an API for creating an event , Create a frontend for API and write unit tests
|
Frontend Priority: High Status: In Progress Type: Development individual
|
Create an API for creating event and and integrate it into the our practice-app.
Write Unit Tests. Use pydate, ensure that the event created can only be created for the future time.
|
1.0
|
Implementation Assignment: Create an API for creating an event , Create a frontend for API and write unit tests - Create an API for creating event and and integrate it into the our practice-app.
Write Unit Tests. Use pydate, ensure that the event created can only be created for the future time.
|
non_defect
|
implementation assignment create an api for creating an event create a frontend for api and write unit tests create an api for creating event and and integrate it into the our practice app write unit tests use pydate ensure that the event created can only be created for the future time
| 0
|
18,255
| 4,241,350,753
|
IssuesEvent
|
2016-07-06 16:03:32
|
projectcalico/calico-containers
|
https://api.github.com/repos/projectcalico/calico-containers
|
opened
|
Dockerless Calico is Outdated
|
documentation enhancement
|
Our [Dockerless Calico](https://github.com/projectcalico/calico-containers/blob/master/docs/DockerlessCalicoManual.md) guide has fallen out of date, namely with the introduction of `setup.py`.
|
1.0
|
Dockerless Calico is Outdated - Our [Dockerless Calico](https://github.com/projectcalico/calico-containers/blob/master/docs/DockerlessCalicoManual.md) guide has fallen out of date, namely with the introduction of `setup.py`.
|
non_defect
|
dockerless calico is outdated our guide has fallen out of date namely with the introduction of setup py
| 0
|
36,568
| 7,992,862,527
|
IssuesEvent
|
2018-07-20 04:20:19
|
lagom/lagom
|
https://api.github.com/repos/lagom/lagom
|
closed
|
Creating a LagomClientApplication fails if no application.conf found
|
help wanted topic:service-client type:defect
|
A Lagom Client application will fail to run if an `application.conf` file is missing. Adding an empty `application.conf` file in `src/main/resources` solved the issue.
To reproduce, clone the [repo](https://github.com/ignasi35/lagom-client-demo), rename `application.conf` and start the `Main`.
|
1.0
|
Creating a LagomClientApplication fails if no application.conf found - A Lagom Client application will fail to run if an `application.conf` file is missing. Adding an empty `application.conf` file in `src/main/resources` solved the issue.
To reproduce, clone the [repo](https://github.com/ignasi35/lagom-client-demo), rename `application.conf` and start the `Main`.
|
defect
|
creating a lagomclientapplication fails if no application conf found a lagom client application will fail to run if an application conf file is missing adding an empty application conf file in src main resources solved the issue to reproduce clone the rename application conf and start the main
| 1
|
113,570
| 4,562,003,909
|
IssuesEvent
|
2016-09-14 13:40:13
|
cul-2016/quiz
|
https://api.github.com/repos/cul-2016/quiz
|
closed
|
Create `StrengthsWeaknesses` component
|
priority-2
|
#121
## Messages to display
### If total number quizzes < 3
As you take more quizzes we will identify those that you are doing particularly well on relative to other people, and those that you are doing less well on.
Otherwise display:
* Looking across all the quizzes you’ve taken, your top quiz was `[Name of quiz with the highest difference score]`.
The quiz where you could improve most, relative to other people, is `[Name of quiz with the lowest difference score]`. You can use this information to help guide your revision, and work out which areas you might want to spend more time getting to grips with.
|
1.0
|
Create `StrengthsWeaknesses` component - #121
## Messages to display
### If total number quizzes < 3
As you take more quizzes we will identify those that you are doing particularly well on relative to other people, and those that you are doing less well on.
Otherwise display:
* Looking across all the quizzes you’ve taken, your top quiz was `[Name of quiz with the highest difference score]`.
The quiz where you could improve most, relative to other people, is `[Name of quiz with the lowest difference score]`. You can use this information to help guide your revision, and work out which areas you might want to spend more time getting to grips with.
|
non_defect
|
create strengthsweaknesses component messages to display if total number quizzes as you take more quizzes we will identify those that you are doing particularly well on relative to other people and those that you are doing less well on otherwise display looking across all the quizzes you’ve taken your top quiz was the quiz where you could improve most relative to other people is you can use this information to help guide your revision and work out which areas you might want to spend more time getting to grips with
| 0
|
46,149
| 5,790,695,838
|
IssuesEvent
|
2017-05-02 01:48:41
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed tests on master: test (proposer evaluated kv)/TestPushTxnHeartbeatTimeout
|
Robot test-failure
|
The following tests appear to have failed:
[#193266](https://teamcity.cockroachdb.com/viewLog.html?buildId=193266):
```
--- FAIL: test (proposer evaluated kv)/TestPushTxnHeartbeatTimeout (0.010s)
replica_test.go:4387: 3: expSuccess=false; got pErr <nil>, reply header:<num_keys:0 > pushee_txn:<meta:<id:<f6bd9663-cd89-4cb8-8cf7-e87834b6195f> isolation:SERIALIZABLE key:"key-3" epoch:0 timestamp:<wall_time:5000000123 logical:26 > priority:0 sequence:1 batch_index:0 > name:"test-3" status:PENDING orig_timestamp:<wall_time:123 logical:53 > max_timestamp:<wall_time:124 logical:53 > writing:true write_too_old:false retry_on_push:false >
------- Stdout: -------
I170324 15:14:03.833121 5121 storage/store.go:1318 [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I170324 15:14:03.836126 5121 util/stop/stopper.go:493 quiescing; tasks left:
2 gossip/infostore.go:301
I170324 15:14:03.836160 5121 util/stop/stopper.go:493 quiescing; tasks left:
1 gossip/infostore.go:301
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed tests on master: test (proposer evaluated kv)/TestPushTxnHeartbeatTimeout - The following tests appear to have failed:
[#193266](https://teamcity.cockroachdb.com/viewLog.html?buildId=193266):
```
--- FAIL: test (proposer evaluated kv)/TestPushTxnHeartbeatTimeout (0.010s)
replica_test.go:4387: 3: expSuccess=false; got pErr <nil>, reply header:<num_keys:0 > pushee_txn:<meta:<id:<f6bd9663-cd89-4cb8-8cf7-e87834b6195f> isolation:SERIALIZABLE key:"key-3" epoch:0 timestamp:<wall_time:5000000123 logical:26 > priority:0 sequence:1 batch_index:0 > name:"test-3" status:PENDING orig_timestamp:<wall_time:123 logical:53 > max_timestamp:<wall_time:124 logical:53 > writing:true write_too_old:false retry_on_push:false >
------- Stdout: -------
I170324 15:14:03.833121 5121 storage/store.go:1318 [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I170324 15:14:03.836126 5121 util/stop/stopper.go:493 quiescing; tasks left:
2 gossip/infostore.go:301
I170324 15:14:03.836160 5121 util/stop/stopper.go:493 quiescing; tasks left:
1 gossip/infostore.go:301
```
Please assign, take a look and update the issue accordingly.
|
non_defect
|
teamcity failed tests on master test proposer evaluated kv testpushtxnheartbeattimeout the following tests appear to have failed fail test proposer evaluated kv testpushtxnheartbeattimeout replica test go expsuccess false got perr reply header pushee txn isolation serializable key key epoch timestamp priority sequence batch index name test status pending orig timestamp max timestamp writing true write too old false retry on push false stdout storage store go failed initial metrics computation system config not yet available util stop stopper go quiescing tasks left gossip infostore go util stop stopper go quiescing tasks left gossip infostore go please assign take a look and update the issue accordingly
| 0
|
278,803
| 24,177,552,630
|
IssuesEvent
|
2022-09-23 04:45:23
|
TauCetiStation/TauCetiClassic
|
https://api.github.com/repos/TauCetiStation/TauCetiClassic
|
closed
|
Блоб не ломает канистры с газом
|
Bug Test Feedback
|
<!--
ВАЖНО: Если ваш ишью является не репортом о баге, а предложением для чего-либо, то ОБЯЗАТЕЛЬНО добавьте в название тег [Proposal]
1. ОТВЕТЫ ОСТАВЛЯТЬ ПОД СООТВЕТСТВУЮЩИЕ ЗАГОЛОВКИ
(они в самом низу, после всех правил)
2. В ОДНОМ РЕПОРТЕ ДОЛЖНО БЫТЬ ОПИСАНИЕ ТОЛЬКО ОДНОЙ ПРОБЛЕМЫ
3. КОРРЕКТНОЕ НАЗВАНИЕ РЕПОРТА НЕ МЕНЕЕ ВАЖНО ЧЕМ ОПИСАНИЕ
-. Ниже описание каждого пункта.
1. Весь данный текст что уже написан до вас -
НЕ УДАЛЯТЬ И НЕ РЕДАКТИРОВАТЬ.
Если нечего написать в тот или иной пункт -
просто оставить пустым.
2. Не надо описывать пачку багов в одном репорте,
(!даже если там все описать можно парой слов!)
шанс что их исправят за раз, крайне мал.
А вот использовать на гите удобную функцию -
автозакрытия репорта при мерже пулл реквеста -
исправляющего данный репорт, будет невозможно.
3. Корректное и в меру подробное название репорта -
тоже очень важно! Чтобы даже не заходя в сам репорт -
было понятно что за проблема.
Плохой пример: "Ковер." - что мы должны понять из такого названия?
Хороший пример: "Некорректное отображение спрайтов ковра." -
а вот так уже будет понятно о чем репорт.
Это надо как минимум для того, чтобы вам же самим -
было видно, что репорта_нейм еще нет или наоборот,
уже есть, и это можно было понять не углубляясь в -
чтение каждого репорта внутри. Когда название не имеет конкретики, из -
которого нельзя понять о чем репорт, это также затрудняет функцию поиска.
-->
#### Подробное описание проблемы
Блоб теперь не ломает канистры с газом, вместо этого канистры просто исчезают, без выпускания газа в атмосферу
#### Что должно было произойти
Канистра ломается, газ выходит.
#### Что произошло на самом деле
Канистра с газом просто исчезает
#### Как повторить
Сломать блобом канистру
#### Дополнительная информация:
Возможно из-за полной разрушаемости.
|
1.0
|
Блоб не ломает канистры с газом - <!--
ВАЖНО: Если ваш ишью является не репортом о баге, а предложением для чего-либо, то ОБЯЗАТЕЛЬНО добавьте в название тег [Proposal]
1. ОТВЕТЫ ОСТАВЛЯТЬ ПОД СООТВЕТСТВУЮЩИЕ ЗАГОЛОВКИ
(они в самом низу, после всех правил)
2. В ОДНОМ РЕПОРТЕ ДОЛЖНО БЫТЬ ОПИСАНИЕ ТОЛЬКО ОДНОЙ ПРОБЛЕМЫ
3. КОРРЕКТНОЕ НАЗВАНИЕ РЕПОРТА НЕ МЕНЕЕ ВАЖНО ЧЕМ ОПИСАНИЕ
-. Ниже описание каждого пункта.
1. Весь данный текст что уже написан до вас -
НЕ УДАЛЯТЬ И НЕ РЕДАКТИРОВАТЬ.
Если нечего написать в тот или иной пункт -
просто оставить пустым.
2. Не надо описывать пачку багов в одном репорте,
(!даже если там все описать можно парой слов!)
шанс что их исправят за раз, крайне мал.
А вот использовать на гите удобную функцию -
автозакрытия репорта при мерже пулл реквеста -
исправляющего данный репорт, будет невозможно.
3. Корректное и в меру подробное название репорта -
тоже очень важно! Чтобы даже не заходя в сам репорт -
было понятно что за проблема.
Плохой пример: "Ковер." - что мы должны понять из такого названия?
Хороший пример: "Некорректное отображение спрайтов ковра." -
а вот так уже будет понятно о чем репорт.
Это надо как минимум для того, чтобы вам же самим -
было видно, что репорта_нейм еще нет или наоборот,
уже есть, и это можно было понять не углубляясь в -
чтение каждого репорта внутри. Когда название не имеет конкретики, из -
которого нельзя понять о чем репорт, это также затрудняет функцию поиска.
-->
#### Подробное описание проблемы
Блоб теперь не ломает канистры с газом, вместо этого канистры просто исчезают, без выпускания газа в атмосферу
#### Что должно было произойти
Канистра ломается, газ выходит.
#### Что произошло на самом деле
Канистра с газом просто исчезает
#### Как повторить
Сломать блобом канистру
#### Дополнительная информация:
Возможно из-за полной разрушаемости.
|
non_defect
|
блоб не ломает канистры с газом важно если ваш ишью является не репортом о баге а предложением для чего либо то обязательно добавьте в название тег ответы оставлять под соответствующие заголовки они в самом низу после всех правил в одном репорте должно быть описание только одной проблемы корректное название репорта не менее важно чем описание ниже описание каждого пункта весь данный текст что уже написан до вас не удалять и не редактировать если нечего написать в тот или иной пункт просто оставить пустым не надо описывать пачку багов в одном репорте даже если там все описать можно парой слов шанс что их исправят за раз крайне мал а вот использовать на гите удобную функцию автозакрытия репорта при мерже пулл реквеста исправляющего данный репорт будет невозможно корректное и в меру подробное название репорта тоже очень важно чтобы даже не заходя в сам репорт было понятно что за проблема плохой пример ковер что мы должны понять из такого названия хороший пример некорректное отображение спрайтов ковра а вот так уже будет понятно о чем репорт это надо как минимум для того чтобы вам же самим было видно что репорта нейм еще нет или наоборот уже есть и это можно было понять не углубляясь в чтение каждого репорта внутри когда название не имеет конкретики из которого нельзя понять о чем репорт это также затрудняет функцию поиска подробное описание проблемы блоб теперь не ломает канистры с газом вместо этого канистры просто исчезают без выпускания газа в атмосферу что должно было произойти канистра ломается газ выходит что произошло на самом деле канистра с газом просто исчезает как повторить сломать блобом канистру дополнительная информация возможно из за полной разрушаемости
| 0
|
26,090
| 4,581,177,083
|
IssuesEvent
|
2016-09-19 03:01:44
|
zealdocs/zeal
|
https://api.github.com/repos/zealdocs/zeal
|
closed
|
Meteor Js Docs
|
Component: Docset Registry Platform: Linux Resolution: Unable To Reproduce Type: Defect
|
Meteor Js docs always shutdown my pc when i open it. please help we have tried it on several systems but result.
|
1.0
|
Meteor Js Docs - Meteor Js docs always shutdown my pc when i open it. please help we have tried it on several systems but result.
|
defect
|
meteor js docs meteor js docs always shutdown my pc when i open it please help we have tried it on several systems but result
| 1
|
55,193
| 6,445,306,003
|
IssuesEvent
|
2017-08-13 02:08:35
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed tests on master: acceptance/TestDockerC/Success, acceptance/TestDockerC
|
Robot test-failure
|
The following tests appear to have failed:
[#318821](https://teamcity.cockroachdb.com/viewLog.html?buildId=318821):
```
--- FAIL: acceptance/TestDockerC/Success (0.000s)
Test ended in panic.
------- Stdout: -------
xenial-20170214: Pulling from library/ubuntu
Digest: sha256:dd7808d8792c9841d0b460122f1acf0a2dd1f56404f8d1e56298048885e45535
Status: Image is up to date for ubuntu:xenial-20170214
Pulling repository docker.io/library/ubuntu
panic: Tag xenial-20170214 not found in repository docker.io/library/ubuntu [recovered]
panic: Tag xenial-20170214 not found in repository docker.io/library/ubuntu [recovered]
panic: Tag xenial-20170214 not found in repository docker.io/library/ubuntu
goroutine 55 [running]:
testing.tRunner.func1(0xc42027d5f0)
/usr/local/go/src/testing/testing.go:622 +0x29d
panic(0x186ea80, 0xc42039eac0)
/usr/local/go/src/runtime/panic.go:489 +0x2cf
github.com/cockroachdb/cockroach/pkg/acceptance/cluster.(*LocalCluster).stopOnPanic(0xc42082f0e0, 0x7fabf6583118, 0xc420018de8)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/cluster/localcluster.go:273 +0xb1
panic(0x186ea80, 0xc42039eac0)
/usr/local/go/src/runtime/panic.go:489 +0x2cf
github.com/cockroachdb/cockroach/pkg/acceptance/cluster.maybePanic(0x266ca40, 0xc42039eac0)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/cluster/docker.go:153 +0x4a
github.com/cockroachdb/cockroach/pkg/acceptance/cluster.(*LocalCluster).RunInitCommand(0xc42082f0e0, 0x7fabf6583118, 0xc420018de8, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/cluster/localcluster.go:571 +0x2a3
github.com/cockroachdb/cockroach/pkg/acceptance/cluster.(*LocalCluster).Start(0xc42082f0e0, 0x7fabf6583118, 0xc420018de8)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/cluster/localcluster.go:695 +0x542
github.com/cockroachdb/cockroach/pkg/acceptance.StartCluster(0x7fabf6583118, 0xc420018de8, 0xc42027d5f0, 0x1a4d088, 0x1, 0xc4203c5c80, 0x1, 0x1, 0x12a05f200, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/util.go:414 +0x191
github.com/cockroachdb/cockroach/pkg/acceptance.testDocker(0x7fabf6583118, 0xc420018de8, 0xc42027d5f0, 0xc400000001, 0x1a4d088, 0x1, 0x0, 0x0, 0x0, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/util.go:555 +0x1bb
github.com/cockroachdb/cockroach/pkg/acceptance.testDockerSingleNode(0x7fabf6583118, 0xc420018de8, 0xc42027d5f0, 0x1a4d088, 0x1, 0x0, 0x0, 0x0, 0x0, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/util.go:575 +0x9f
github.com/cockroachdb/cockroach/pkg/acceptance.testDockerSuccess(0x7fabf6583118, 0xc420018de8, 0xc42027d5f0, 0x1a4d088, 0x1, 0xc4208302d0, 0x3, 0x3)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/util.go:535 +0xea
github.com/cockroachdb/cockroach/pkg/acceptance.TestDockerC.func1(0xc42027d5f0)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/c_test.go:32 +0x14b
testing.tRunner(0xc42027d5f0, 0xc4203c5c20)
/usr/local/go/src/testing/testing.go:657 +0x96
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:697 +0x2ca
--- FAIL: acceptance/TestDockerC (0.000s)
Test ended in panic.
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed tests on master: acceptance/TestDockerC/Success, acceptance/TestDockerC - The following tests appear to have failed:
[#318821](https://teamcity.cockroachdb.com/viewLog.html?buildId=318821):
```
--- FAIL: acceptance/TestDockerC/Success (0.000s)
Test ended in panic.
------- Stdout: -------
xenial-20170214: Pulling from library/ubuntu
Digest: sha256:dd7808d8792c9841d0b460122f1acf0a2dd1f56404f8d1e56298048885e45535
Status: Image is up to date for ubuntu:xenial-20170214
Pulling repository docker.io/library/ubuntu
panic: Tag xenial-20170214 not found in repository docker.io/library/ubuntu [recovered]
panic: Tag xenial-20170214 not found in repository docker.io/library/ubuntu [recovered]
panic: Tag xenial-20170214 not found in repository docker.io/library/ubuntu
goroutine 55 [running]:
testing.tRunner.func1(0xc42027d5f0)
/usr/local/go/src/testing/testing.go:622 +0x29d
panic(0x186ea80, 0xc42039eac0)
/usr/local/go/src/runtime/panic.go:489 +0x2cf
github.com/cockroachdb/cockroach/pkg/acceptance/cluster.(*LocalCluster).stopOnPanic(0xc42082f0e0, 0x7fabf6583118, 0xc420018de8)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/cluster/localcluster.go:273 +0xb1
panic(0x186ea80, 0xc42039eac0)
/usr/local/go/src/runtime/panic.go:489 +0x2cf
github.com/cockroachdb/cockroach/pkg/acceptance/cluster.maybePanic(0x266ca40, 0xc42039eac0)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/cluster/docker.go:153 +0x4a
github.com/cockroachdb/cockroach/pkg/acceptance/cluster.(*LocalCluster).RunInitCommand(0xc42082f0e0, 0x7fabf6583118, 0xc420018de8, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/cluster/localcluster.go:571 +0x2a3
github.com/cockroachdb/cockroach/pkg/acceptance/cluster.(*LocalCluster).Start(0xc42082f0e0, 0x7fabf6583118, 0xc420018de8)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/cluster/localcluster.go:695 +0x542
github.com/cockroachdb/cockroach/pkg/acceptance.StartCluster(0x7fabf6583118, 0xc420018de8, 0xc42027d5f0, 0x1a4d088, 0x1, 0xc4203c5c80, 0x1, 0x1, 0x12a05f200, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/util.go:414 +0x191
github.com/cockroachdb/cockroach/pkg/acceptance.testDocker(0x7fabf6583118, 0xc420018de8, 0xc42027d5f0, 0xc400000001, 0x1a4d088, 0x1, 0x0, 0x0, 0x0, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/util.go:555 +0x1bb
github.com/cockroachdb/cockroach/pkg/acceptance.testDockerSingleNode(0x7fabf6583118, 0xc420018de8, 0xc42027d5f0, 0x1a4d088, 0x1, 0x0, 0x0, 0x0, 0x0, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/util.go:575 +0x9f
github.com/cockroachdb/cockroach/pkg/acceptance.testDockerSuccess(0x7fabf6583118, 0xc420018de8, 0xc42027d5f0, 0x1a4d088, 0x1, 0xc4208302d0, 0x3, 0x3)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/util.go:535 +0xea
github.com/cockroachdb/cockroach/pkg/acceptance.TestDockerC.func1(0xc42027d5f0)
/go/src/github.com/cockroachdb/cockroach/pkg/acceptance/c_test.go:32 +0x14b
testing.tRunner(0xc42027d5f0, 0xc4203c5c20)
/usr/local/go/src/testing/testing.go:657 +0x96
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:697 +0x2ca
--- FAIL: acceptance/TestDockerC (0.000s)
Test ended in panic.
```
Please assign, take a look and update the issue accordingly.
|
non_defect
|
teamcity failed tests on master acceptance testdockerc success acceptance testdockerc the following tests appear to have failed fail acceptance testdockerc success test ended in panic stdout xenial pulling from library ubuntu digest status image is up to date for ubuntu xenial pulling repository docker io library ubuntu panic tag xenial not found in repository docker io library ubuntu panic tag xenial not found in repository docker io library ubuntu panic tag xenial not found in repository docker io library ubuntu goroutine testing trunner usr local go src testing testing go panic usr local go src runtime panic go github com cockroachdb cockroach pkg acceptance cluster localcluster stoponpanic go src github com cockroachdb cockroach pkg acceptance cluster localcluster go panic usr local go src runtime panic go github com cockroachdb cockroach pkg acceptance cluster maybepanic go src github com cockroachdb cockroach pkg acceptance cluster docker go github com cockroachdb cockroach pkg acceptance cluster localcluster runinitcommand go src github com cockroachdb cockroach pkg acceptance cluster localcluster go github com cockroachdb cockroach pkg acceptance cluster localcluster start go src github com cockroachdb cockroach pkg acceptance cluster localcluster go github com cockroachdb cockroach pkg acceptance startcluster go src github com cockroachdb cockroach pkg acceptance util go github com cockroachdb cockroach pkg acceptance testdocker go src github com cockroachdb cockroach pkg acceptance util go github com cockroachdb cockroach pkg acceptance testdockersinglenode go src github com cockroachdb cockroach pkg acceptance util go github com cockroachdb cockroach pkg acceptance testdockersuccess go src github com cockroachdb cockroach pkg acceptance util go github com cockroachdb cockroach pkg acceptance testdockerc go src github com cockroachdb cockroach pkg acceptance c test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go fail acceptance testdockerc test ended in panic please assign take a look and update the issue accordingly
| 0
|
646,105
| 21,038,024,032
|
IssuesEvent
|
2022-03-31 09:39:11
|
googleapis/nodejs-video-intelligence
|
https://api.github.com/repos/googleapis/nodejs-video-intelligence
|
opened
|
analyze v1p2beta1 samples: should track objects in a GCS file failed
|
type: bug priority: p1 flakybot: issue
|
Note: #642 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 02e37d247131b2892c1183a34b1d1e4b6d8ea270
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/ca632117-27c5-4979-8940-f0d691ce4f29), [Sponge](http://sponge2/ca632117-27c5-4979-8940-f0d691ce4f29)
status: failed
<details><summary>Test output</summary><br><pre>expected '' to match /cat/
AssertionError: expected '' to match /cat/
at Context.<anonymous> (system-test/analyze.v1p2beta1.test.js:40:12)
at processImmediate (internal/timers.js:461:21)</pre></details>
|
1.0
|
analyze v1p2beta1 samples: should track objects in a GCS file failed - Note: #642 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 02e37d247131b2892c1183a34b1d1e4b6d8ea270
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/ca632117-27c5-4979-8940-f0d691ce4f29), [Sponge](http://sponge2/ca632117-27c5-4979-8940-f0d691ce4f29)
status: failed
<details><summary>Test output</summary><br><pre>expected '' to match /cat/
AssertionError: expected '' to match /cat/
at Context.<anonymous> (system-test/analyze.v1p2beta1.test.js:40:12)
at processImmediate (internal/timers.js:461:21)</pre></details>
|
non_defect
|
analyze samples should track objects in a gcs file failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output expected to match cat assertionerror expected to match cat at context system test analyze test js at processimmediate internal timers js
| 0
|
236,934
| 7,753,700,625
|
IssuesEvent
|
2018-05-31 02:15:07
|
Monika-After-Story/MonikaModDev
|
https://api.github.com/repos/Monika-After-Story/MonikaModDev
|
closed
|
[Suggestion] LOOP tags in custom BGM
|
enhancement low priority
|
As you know, some of the songs in DDLC/MAS loop properly thanks to loop points.
For example, Sayo-nara ("Surprise!" in MAS) loops back to the 36.782 position when it reaches the end, and the 80s-style cover of DDLC's main theme loops between the 17.451 and 119.999 positions.
Some programs/games that support Ogg/Vorbis files also support looping by using Vorbis Comment tags named `LOOPSTART` and `LOOPLENGTH` (RPG Maker VX Ace is an example that uses those tags).
However, while DDLC/MAS use a time in `seconds.milliseconds`, `LOOPSTART` and `LOOPLENGTH` have their time in number of samples.
It would be nice if MAS's custom BGM system could parse those tags.
You'd need to convert samples to a time in `seconds.milliseconds` (which means getting the sample rate from the file's metadata), then determining the loop's end point by adding the length's time to the start point's position. Or something similar.
|
1.0
|
[Suggestion] LOOP tags in custom BGM - As you know, some of the songs in DDLC/MAS loop properly thanks to loop points.
For example, Sayo-nara ("Surprise!" in MAS) loops back to the 36.782 position when it reaches the end, and the 80s-style cover of DDLC's main theme loops between the 17.451 and 119.999 positions.
Some programs/games that support Ogg/Vorbis files also support looping by using Vorbis Comment tags named `LOOPSTART` and `LOOPLENGTH` (RPG Maker VX Ace is an example that uses those tags).
However, while DDLC/MAS use a time in `seconds.milliseconds`, `LOOPSTART` and `LOOPLENGTH` have their time in number of samples.
It would be nice if MAS's custom BGM system could parse those tags.
You'd need to convert samples to a time in `seconds.milliseconds` (which means getting the sample rate from the file's metadata), then determining the loop's end point by adding the length's time to the start point's position. Or something similar.
|
non_defect
|
loop tags in custom bgm as you know some of the songs in ddlc mas loop properly thanks to loop points for example sayo nara surprise in mas loops back to the position when it reaches the end and the style cover of ddlc s main theme loops between the and positions some programs games that support ogg vorbis files also support looping by using vorbis comment tags named loopstart and looplength rpg maker vx ace is an example that uses those tags however while ddlc mas use a time in seconds milliseconds loopstart and looplength have their time in number of samples it would be nice if mas s custom bgm system could parse those tags you d need to convert samples to a time in seconds milliseconds which means getting the sample rate from the file s metadata then determining the loop s end point by adding the length s time to the start point s position or something similar
| 0
|
212,507
| 16,455,652,085
|
IssuesEvent
|
2021-05-21 12:12:28
|
Reiningecho90/S.A.N.E.-AI
|
https://api.github.com/repos/Reiningecho90/S.A.N.E.-AI
|
closed
|
Settings not Changing Properly
|
bug in-work testing
|
JSON file data not changing when function runs.
1. Run setup program
2. Click both buttons sequentially
3. Look at JSON file data
4. No changes made
The JSON file data should change when one of the buttons are clicked.
Windows 10 VS Code
|
1.0
|
Settings not Changing Properly - JSON file data not changing when function runs.
1. Run setup program
2. Click both buttons sequentially
3. Look at JSON file data
4. No changes made
The JSON file data should change when one of the buttons are clicked.
Windows 10 VS Code
|
non_defect
|
settings not changing properly json file data not changing when function runs run setup program click both buttons sequentially look at json file data no changes made the json file data should change when one of the buttons are clicked windows vs code
| 0
|
74,281
| 25,036,339,618
|
IssuesEvent
|
2022-11-04 16:19:55
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
"Error decrypting image" in the timeline, but I can download and view the image fine
|
T-Defect
|
### Steps to reproduce

I saw this in a timeline. Switching back and forth between rooms did not change anything. Clicking "download" on the image allowed me to view it correctly.
My assumption is that the keys for this message did not initially arrive, causing this error, but once they did arrive the error did not disappear?
### Outcome
#### What did you expect?
To not be able to download and decrypt the image, because it wasn't able to be displayed in the timeline.
#### What happened instead?
I couldn't decrypt/display the image in the timeline, but I could see the image when clicking "download". Why are these different?
### Operating system
Arch Linux
### Application version
Element version: 2022102901
Olm version: 3.2.12
### How did you install the app?
_No response_
### Homeserver
element.io
### Will you send logs?
Yes
|
1.0
|
"Error decrypting image" in the timeline, but I can download and view the image fine - ### Steps to reproduce

I saw this in a timeline. Switching back and forth between rooms did not change anything. Clicking "download" on the image allowed me to view it correctly.
My assumption is that the keys for this message did not initially arrive, causing this error, but once they did arrive the error did not disappear?
### Outcome
#### What did you expect?
To not be able to download and decrypt the image, because it wasn't able to be displayed in the timeline.
#### What happened instead?
I couldn't decrypt/display the image in the timeline, but I could see the image when clicking "download". Why are these different?
### Operating system
Arch Linux
### Application version
Element version: 2022102901
Olm version: 3.2.12
### How did you install the app?
_No response_
### Homeserver
element.io
### Will you send logs?
Yes
|
defect
|
error decrypting image in the timeline but i can download and view the image fine steps to reproduce i saw this in a timeline switching back and forth between rooms did not change anything clicking download on the image allowed me to view it correctly my assumption is that the keys for this message did not initially arrive causing this error but once they did arrive the error did not disappear outcome what did you expect to not be able to download and decrypt the image because it wasn t able to be displayed in the timeline what happened instead i couldn t decrypt display the image in the timeline but i could see the image when clicking download why are these different operating system arch linux application version element version olm version how did you install the app no response homeserver element io will you send logs yes
| 1
|
706,223
| 24,262,498,450
|
IssuesEvent
|
2022-09-28 01:03:02
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Bazel C/C++ Opt MacOS: //test/cpp/end2end:client_callback_end2end_test failed due to SIGSEGV
|
kind/bug lang/c++ priority/P2 untriaged
|
https://source.cloud.google.com/results/invocations/6804105c-50dc-418f-bb90-873eb8d7d48b/log
*******************************
Caught signal SIGSEGV
no backtrace
external/bazel_tools/tools/test/test-setup.sh: line 310: 98643 Segmentation fault: 11 "${TEST_PATH}" "$@" 2>&1
================================================================================
[10,546 / 10,554] 1656 / 2536 tests, 1 failed; Testing //test/core/end2end:h2_census_test@high_initial_seqno; 20s remote-cache, darwin-sandbox ... (16 actions running)
INFO: From Compiling test/core/channel/channel_args_test.cc:
test/core/channel/channel_args_test.cc:141:13: warning: unused function 'test_channel_create_with_global_mutator' [-Wunused-function]
static void test_channel_create_with_global_mutator(void) {
^
1 warning generated.
[12,094 / 12,111] 2129 / 2536 tests, 1 failed; Testing //test/core/end2end:h2_http_proxy_test@cancel_after_invoke; 35s remote-cache, darwin-sandbox ... (16 actions running)
[12,893 / 12,901] 2380 / 2536 tests, 1 failed; Testing //test/core/end2end:h2_tls_test@cancel_after_invoke; 11s remote-cache, darwin-sandbox ... (16 actions running)
INFO: Elapsed time: 1861.292s, Critical Path: 270.17s
INFO: 8049 processes: 8049 darwin-sandbox.
INFO: Build completed, 1 test FAILED, 13354 total actions
|
1.0
|
Bazel C/C++ Opt MacOS: //test/cpp/end2end:client_callback_end2end_test failed due to SIGSEGV - https://source.cloud.google.com/results/invocations/6804105c-50dc-418f-bb90-873eb8d7d48b/log
*******************************
Caught signal SIGSEGV
no backtrace
external/bazel_tools/tools/test/test-setup.sh: line 310: 98643 Segmentation fault: 11 "${TEST_PATH}" "$@" 2>&1
================================================================================
[10,546 / 10,554] 1656 / 2536 tests, 1 failed; Testing //test/core/end2end:h2_census_test@high_initial_seqno; 20s remote-cache, darwin-sandbox ... (16 actions running)
INFO: From Compiling test/core/channel/channel_args_test.cc:
test/core/channel/channel_args_test.cc:141:13: warning: unused function 'test_channel_create_with_global_mutator' [-Wunused-function]
static void test_channel_create_with_global_mutator(void) {
^
1 warning generated.
[12,094 / 12,111] 2129 / 2536 tests, 1 failed; Testing //test/core/end2end:h2_http_proxy_test@cancel_after_invoke; 35s remote-cache, darwin-sandbox ... (16 actions running)
[12,893 / 12,901] 2380 / 2536 tests, 1 failed; Testing //test/core/end2end:h2_tls_test@cancel_after_invoke; 11s remote-cache, darwin-sandbox ... (16 actions running)
INFO: Elapsed time: 1861.292s, Critical Path: 270.17s
INFO: 8049 processes: 8049 darwin-sandbox.
INFO: Build completed, 1 test FAILED, 13354 total actions
|
non_defect
|
bazel c c opt macos test cpp client callback test failed due to sigsegv caught signal sigsegv no backtrace external bazel tools tools test test setup sh line segmentation fault test path tests failed testing test core census test high initial seqno remote cache darwin sandbox actions running info from compiling test core channel channel args test cc test core channel channel args test cc warning unused function test channel create with global mutator static void test channel create with global mutator void warning generated tests failed testing test core http proxy test cancel after invoke remote cache darwin sandbox actions running tests failed testing test core tls test cancel after invoke remote cache darwin sandbox actions running info elapsed time critical path info processes darwin sandbox info build completed test failed total actions
| 0
|
53,844
| 13,262,370,195
|
IssuesEvent
|
2020-08-20 21:41:18
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
MuonGun produces inconsistent weights between different versions (Trac #2179)
|
Migrated from Trac analysis defect
|
The MuonGun weighting module produces wrong weights for events above 10^6^ GeV. This behaviour starts at rev 159352 and is probably at least related to tickets #2131 and #2149.
This can be tested with running the script
```text
/home/mmeier/test_muongun_weighting.py --out_file=/path/to/file.hd5
```
once with a combo metaproject with MuonGun rev 159351 and with rev 159352 or later.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2179">https://code.icecube.wisc.edu/projects/icecube/ticket/2179</a>, reported by mmeierand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2018-08-03T13:43:03",
"_ts": "1533303783524855",
"description": "The MuonGun weighting module produces wrong weights for events above 10^6^ GeV. This behaviour starts at rev 159352 and is probably at least related to tickets #2139 and #2157.\nThis can be tested with running the script\n{{{\n/home/mmeier/test_muongun_weighting.py --out_file=/path/to/file.hd5\n}}}\nonce with a combo metaproject with MuonGun rev 159351 and with rev 159352 or later.",
"reporter": "mmeier",
"cc": "",
"resolution": "fixed",
"time": "2018-08-02T15:15:04",
"component": "analysis",
"summary": "MuonGun produces inconsistent weights between different versions",
"priority": "normal",
"keywords": "MuonGun",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
MuonGun produces inconsistent weights between different versions (Trac #2179) - The MuonGun weighting module produces wrong weights for events above 10^6^ GeV. This behaviour starts at rev 159352 and is probably at least related to tickets #2131 and #2149.
This can be tested with running the script
```text
/home/mmeier/test_muongun_weighting.py --out_file=/path/to/file.hd5
```
once with a combo metaproject with MuonGun rev 159351 and with rev 159352 or later.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2179">https://code.icecube.wisc.edu/projects/icecube/ticket/2179</a>, reported by mmeierand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2018-08-03T13:43:03",
"_ts": "1533303783524855",
"description": "The MuonGun weighting module produces wrong weights for events above 10^6^ GeV. This behaviour starts at rev 159352 and is probably at least related to tickets #2139 and #2157.\nThis can be tested with running the script\n{{{\n/home/mmeier/test_muongun_weighting.py --out_file=/path/to/file.hd5\n}}}\nonce with a combo metaproject with MuonGun rev 159351 and with rev 159352 or later.",
"reporter": "mmeier",
"cc": "",
"resolution": "fixed",
"time": "2018-08-02T15:15:04",
"component": "analysis",
"summary": "MuonGun produces inconsistent weights between different versions",
"priority": "normal",
"keywords": "MuonGun",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
defect
|
muongun produces inconsistent weights between different versions trac the muongun weighting module produces wrong weights for events above gev this behaviour starts at rev and is probably at least related to tickets and this can be tested with running the script text home mmeier test muongun weighting py out file path to file once with a combo metaproject with muongun rev and with rev or later migrated from json status closed changetime ts description the muongun weighting module produces wrong weights for events above gev this behaviour starts at rev and is probably at least related to tickets and nthis can be tested with running the script n n home mmeier test muongun weighting py out file path to file n nonce with a combo metaproject with muongun rev and with rev or later reporter mmeier cc resolution fixed time component analysis summary muongun produces inconsistent weights between different versions priority normal keywords muongun milestone owner jvansanten type defect
| 1
|
58,922
| 16,919,173,629
|
IssuesEvent
|
2021-06-25 01:04:00
|
SAP/fundamental-ngx
|
https://api.github.com/repos/SAP/fundamental-ngx
|
opened
|
Action Sheet issues
|
Defect Hunting bug core denoland
|
#### Is this a bug, enhancement, or feature request?
bug
#### Briefly describe your proposal.
- [ ] this example is very tall

- [ ]
|
1.0
|
Action Sheet issues - #### Is this a bug, enhancement, or feature request?
bug
#### Briefly describe your proposal.
- [ ] this example is very tall

- [ ]
|
defect
|
action sheet issues is this a bug enhancement or feature request bug briefly describe your proposal this example is very tall
| 1
|
7,483
| 2,610,389,337
|
IssuesEvent
|
2015-02-26 20:05:58
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
Text in "Misc" Section of Configuration/Options/Preferences Screen is Illegible
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Open Options screen by clicking on wrench icon in lower-right corner of main
menu screen
2. Text for "Net nick", "Password", "Locale", and other options is mostly
illegible, tops and bottoms of all text (and text entry fields) in "Misc"
section have been cut off.
3.
What is the expected output? What do you see instead?
Normally, I would expect to see complete words, with characters showing at full
height.
Instead, I see characters and text entry fields that are missing their tops and
bottoms.
What version of the product are you using? On what operating system?
Hedgewars 0.9.17
Mac OS X 10.6.8
2.4 GHz Core2Duo
4GB 667 RAM
500GB HD, 28GB Free
Please provide any additional information below.
Resolution is set to default 1280x800, windowed mode (not fullscreen). Window
takes approximately half of vertical and horizontal screen real estate on a
1440x900 display.
Changing resolution to any other resolution, while leaving set to windowed
(non-fullscreen) mode, then restarting the app, results in similar behaviour.
Main menu "Hedgewars" title graphic is also cut off on bottom, and main menu
button graphics are breaking their yellow bounding boxes.
```
-----
Original issue reported on code.google.com by `Dash.M...@gmail.com` on 16 Feb 2012 at 4:08
|
1.0
|
Text in "Misc" Section of Configuration/Options/Preferences Screen is Illegible - ```
What steps will reproduce the problem?
1. Open Options screen by clicking on wrench icon in lower-right corner of main
menu screen
2. Text for "Net nick", "Password", "Locale", and other options is mostly
illegible, tops and bottoms of all text (and text entry fields) in "Misc"
section have been cut off.
3.
What is the expected output? What do you see instead?
Normally, I would expect to see complete words, with characters showing at full
height.
Instead, I see characters and text entry fields that are missing their tops and
bottoms.
What version of the product are you using? On what operating system?
Hedgewars 0.9.17
Mac OS X 10.6.8
2.4 GHz Core2Duo
4GB 667 RAM
500GB HD, 28GB Free
Please provide any additional information below.
Resolution is set to default 1280x800, windowed mode (not fullscreen). Window
takes approximately half of vertical and horizontal screen real estate on a
1440x900 display.
Changing resolution to any other resolution, while leaving set to windowed
(non-fullscreen) mode, then restarting the app, results in similar behaviour.
Main menu "Hedgewars" title graphic is also cut off on bottom, and main menu
button graphics are breaking their yellow bounding boxes.
```
-----
Original issue reported on code.google.com by `Dash.M...@gmail.com` on 16 Feb 2012 at 4:08
|
defect
|
text in misc section of configuration options preferences screen is illegible what steps will reproduce the problem open options screen by clicking on wrench icon in lower right corner of main menu screen text for net nick password locale and other options is mostly illegible tops and bottoms of all text and text entry fields in misc section have been cut off what is the expected output what do you see instead normally i would expect to see complete words with characters showing at full height instead i see characters and text entry fields that are missing their tops and bottoms what version of the product are you using on what operating system hedgewars mac os x ghz ram hd free please provide any additional information below resolution is set to default windowed mode not fullscreen window takes approximately half of vertical and horizontal screen real estate on a display changing resolution to any other resolution while leaving set to windowed non fullscreen mode then restarting the app results in similar behaviour main menu hedgewars title graphic is also cut off on bottom and main menu button graphics are breaking their yellow bounding boxes original issue reported on code google com by dash m gmail com on feb at
| 1
|
382,435
| 11,306,243,475
|
IssuesEvent
|
2020-01-18 12:38:42
|
googleapis/nodejs-datastore
|
https://api.github.com/repos/googleapis/nodejs-datastore
|
closed
|
Synthesis failed for nodejs-datastore
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate nodejs-datastore. :broken_heart:
Here's the output from running `synth.py`:
```
Traceback (most recent call last):
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connection.py", line 157, in _new_conn
(self._dns_host, self.port), self.timeout, **extra_kw
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/util/connection.py", line 84, in create_connection
raise err
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/util/connection.py", line 74, in create_connection
sock.connect(sa)
TimeoutError: [Errno 110] Connection timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connectionpool.py", line 672, in urlopen
chunked=chunked,
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connectionpool.py", line 376, in _make_request
self._validate_conn(conn)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connectionpool.py", line 994, in _validate_conn
conn.connect()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connection.py", line 334, in connect
conn = self._new_conn()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connection.py", line 169, in _new_conn
self, "Failed to establish a new connection: %s" % e
urllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x7f806f087e48>: Failed to establish a new connection: [Errno 110] Connection timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/adapters.py", line 449, in send
timeout=timeout
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connectionpool.py", line 720, in urlopen
method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/util/retry.py", line 436, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='api.github.com', port=443): Max retries exceeded with url: /repos/googleapis/nodejs-datastore/pulls?state=open&head=googleapis%3Aautosynth (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7f806f087e48>: Failed to establish a new connection: [Errno 110] Connection timed out',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 180, in main
pr = check_if_pr_already_exists(gh, args.repository, branch)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 35, in check_if_pr_already_exists
prs = gh.list_pull_requests(repo, state="open", head=f"{owner}:{branch}")
File "/tmpfs/src/git/autosynth/autosynth/github.py", line 47, in list_pull_requests
response = self.session.get(url, params=kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/sessions.py", line 546, in get
return self.request('GET', url, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/sessions.py", line 533, in request
resp = self.send(prep, **send_kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/sessions.py", line 646, in send
r = adapter.send(request, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/adapters.py", line 516, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='api.github.com', port=443): Max retries exceeded with url: /repos/googleapis/nodejs-datastore/pulls?state=open&head=googleapis%3Aautosynth (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7f806f087e48>: Failed to establish a new connection: [Errno 110] Connection timed out',))
```
Google internal developers can see the full log [here](https://sponge/e8e03059-7cc5-44b8-80c6-bb156b741670).
|
1.0
|
Synthesis failed for nodejs-datastore - Hello! Autosynth couldn't regenerate nodejs-datastore. :broken_heart:
Here's the output from running `synth.py`:
```
Traceback (most recent call last):
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connection.py", line 157, in _new_conn
(self._dns_host, self.port), self.timeout, **extra_kw
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/util/connection.py", line 84, in create_connection
raise err
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/util/connection.py", line 74, in create_connection
sock.connect(sa)
TimeoutError: [Errno 110] Connection timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connectionpool.py", line 672, in urlopen
chunked=chunked,
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connectionpool.py", line 376, in _make_request
self._validate_conn(conn)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connectionpool.py", line 994, in _validate_conn
conn.connect()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connection.py", line 334, in connect
conn = self._new_conn()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connection.py", line 169, in _new_conn
self, "Failed to establish a new connection: %s" % e
urllib3.exceptions.NewConnectionError: <urllib3.connection.VerifiedHTTPSConnection object at 0x7f806f087e48>: Failed to establish a new connection: [Errno 110] Connection timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/adapters.py", line 449, in send
timeout=timeout
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/connectionpool.py", line 720, in urlopen
method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/urllib3/util/retry.py", line 436, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='api.github.com', port=443): Max retries exceeded with url: /repos/googleapis/nodejs-datastore/pulls?state=open&head=googleapis%3Aautosynth (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7f806f087e48>: Failed to establish a new connection: [Errno 110] Connection timed out',))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 180, in main
pr = check_if_pr_already_exists(gh, args.repository, branch)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 35, in check_if_pr_already_exists
prs = gh.list_pull_requests(repo, state="open", head=f"{owner}:{branch}")
File "/tmpfs/src/git/autosynth/autosynth/github.py", line 47, in list_pull_requests
response = self.session.get(url, params=kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/sessions.py", line 546, in get
return self.request('GET', url, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/sessions.py", line 533, in request
resp = self.send(prep, **send_kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/sessions.py", line 646, in send
r = adapter.send(request, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/adapters.py", line 516, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='api.github.com', port=443): Max retries exceeded with url: /repos/googleapis/nodejs-datastore/pulls?state=open&head=googleapis%3Aautosynth (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7f806f087e48>: Failed to establish a new connection: [Errno 110] Connection timed out',))
```
Google internal developers can see the full log [here](https://sponge/e8e03059-7cc5-44b8-80c6-bb156b741670).
|
non_defect
|
synthesis failed for nodejs datastore hello autosynth couldn t regenerate nodejs datastore broken heart here s the output from running synth py traceback most recent call last file tmpfs src git autosynth env lib site packages connection py line in new conn self dns host self port self timeout extra kw file tmpfs src git autosynth env lib site packages util connection py line in create connection raise err file tmpfs src git autosynth env lib site packages util connection py line in create connection sock connect sa timeouterror connection timed out during handling of the above exception another exception occurred traceback most recent call last file tmpfs src git autosynth env lib site packages connectionpool py line in urlopen chunked chunked file tmpfs src git autosynth env lib site packages connectionpool py line in make request self validate conn conn file tmpfs src git autosynth env lib site packages connectionpool py line in validate conn conn connect file tmpfs src git autosynth env lib site packages connection py line in connect conn self new conn file tmpfs src git autosynth env lib site packages connection py line in new conn self failed to establish a new connection s e exceptions newconnectionerror failed to establish a new connection connection timed out during handling of the above exception another exception occurred traceback most recent call last file tmpfs src git autosynth env lib site packages requests adapters py line in send timeout timeout file tmpfs src git autosynth env lib site packages connectionpool py line in urlopen method url error e pool self stacktrace sys exc info file tmpfs src git autosynth env lib site packages util retry py line in increment raise maxretryerror pool url error or responseerror cause exceptions maxretryerror httpsconnectionpool host api github com port max retries exceeded with url repos googleapis nodejs datastore pulls state open head googleapis caused by newconnectionerror failed to establish a new connection connection timed out during handling of the above exception another exception occurred traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main pr check if pr already exists gh args repository branch file tmpfs src git autosynth autosynth synth py line in check if pr already exists prs gh list pull requests repo state open head f owner branch file tmpfs src git autosynth autosynth github py line in list pull requests response self session get url params kwargs file tmpfs src git autosynth env lib site packages requests sessions py line in get return self request get url kwargs file tmpfs src git autosynth env lib site packages requests sessions py line in request resp self send prep send kwargs file tmpfs src git autosynth env lib site packages requests sessions py line in send r adapter send request kwargs file tmpfs src git autosynth env lib site packages requests adapters py line in send raise connectionerror e request request requests exceptions connectionerror httpsconnectionpool host api github com port max retries exceeded with url repos googleapis nodejs datastore pulls state open head googleapis caused by newconnectionerror failed to establish a new connection connection timed out google internal developers can see the full log
| 0
|
417,534
| 12,167,067,921
|
IssuesEvent
|
2020-04-27 10:16:20
|
hazelcast/hazelcast-aws
|
https://api.github.com/repos/hazelcast/hazelcast-aws
|
closed
|
Fargate support
|
Estimation: XL Priority: High Type: Enhancement
|
Hi,
I'm trying to deploy Hazelcast cluster on AWS Fargate but I couldn't find any documentation about the process...
Does Hazelcast suport Fargate or it only supports ECS?
All the best!
|
1.0
|
Fargate support - Hi,
I'm trying to deploy Hazelcast cluster on AWS Fargate but I couldn't find any documentation about the process...
Does Hazelcast suport Fargate or it only supports ECS?
All the best!
|
non_defect
|
fargate support hi i m trying to deploy hazelcast cluster on aws fargate but i couldn t find any documentation about the process does hazelcast suport fargate or it only supports ecs all the best
| 0
|
229,186
| 18,286,656,109
|
IssuesEvent
|
2021-10-05 11:03:59
|
DILCISBoard/eark-ip-test-corpus
|
https://api.github.com/repos/DILCISBoard/eark-ip-test-corpus
|
closed
|
CSIP32 Test Case Description
|
test case ready
|
**Specification:**
- **Name:** E-ARK CSIP
- **Version:** 2.0-DRAFT
- **URL:** http://earkcsip.dilcis.eu/
**Requirement:**
- **Id:** CSIP32
- **Link:** http://earkcsip.dilcis.eu/#CSIP32
**Error Level:** WARNING
**Description:**
CSIP32 | Digital provenance metadata amdSec/digiprovMD | For recording information about preservation events the standard PREMIS is used. The PREMIS metadata must be either embedded or linked in a digital provenance metadata (digiprovMD) element. It is mandatory to include one digiprovMD element for each external PREMIS file placed in the “metadata/preservation” section, or for each embedded set of PREMIS metadata. | 0..n SHOULD
-- | -- | -- | --
|
1.0
|
CSIP32 Test Case Description - **Specification:**
- **Name:** E-ARK CSIP
- **Version:** 2.0-DRAFT
- **URL:** http://earkcsip.dilcis.eu/
**Requirement:**
- **Id:** CSIP32
- **Link:** http://earkcsip.dilcis.eu/#CSIP32
**Error Level:** WARNING
**Description:**
CSIP32 | Digital provenance metadata amdSec/digiprovMD | For recording information about preservation events the standard PREMIS is used. The PREMIS metadata must be either embedded or linked in a digital provenance metadata (digiprovMD) element. It is mandatory to include one digiprovMD element for each external PREMIS file placed in the “metadata/preservation” section, or for each embedded set of PREMIS metadata. | 0..n SHOULD
-- | -- | -- | --
|
non_defect
|
test case description specification name e ark csip version draft url requirement id link error level warning description digital provenance metadata amdsec digiprovmd for recording information about preservation events the standard premis is used the premis metadata must be either embedded or linked in a digital provenance metadata digiprovmd element it is mandatory to include one digiprovmd element for each external premis file placed in the “metadata preservation” section or for each embedded set of premis metadata n should
| 0
|
314,374
| 9,596,053,355
|
IssuesEvent
|
2019-05-09 17:36:16
|
EUCweb/BIS-F
|
https://api.github.com/repos/EUCweb/BIS-F
|
opened
|
Skip execution of PVS Target OS Optimization
|
Priority: Medium
|
recommend to use CTXO, vmOSOT OR GPP to control optimization instead of Target OS Optimizer.
|
1.0
|
Skip execution of PVS Target OS Optimization - recommend to use CTXO, vmOSOT OR GPP to control optimization instead of Target OS Optimizer.
|
non_defect
|
skip execution of pvs target os optimization recommend to use ctxo vmosot or gpp to control optimization instead of target os optimizer
| 0
|
27,547
| 5,047,521,381
|
IssuesEvent
|
2016-12-20 09:42:42
|
JohnAustinDev/osis-converters
|
https://api.github.com/repos/JohnAustinDev/osis-converters
|
closed
|
Glossary links within footnotes
|
auto-migrated Priority-High Type-Defect
|
```
Apparently the glossary link parser skips footnotes. This should be fixed.
```
Original issue reported on code.google.com by `gpl.prog...@gmail.com` on 21 Nov 2013 at 2:38
|
1.0
|
Glossary links within footnotes - ```
Apparently the glossary link parser skips footnotes. This should be fixed.
```
Original issue reported on code.google.com by `gpl.prog...@gmail.com` on 21 Nov 2013 at 2:38
|
defect
|
glossary links within footnotes apparently the glossary link parser skips footnotes this should be fixed original issue reported on code google com by gpl prog gmail com on nov at
| 1
|
48,256
| 5,950,563,881
|
IssuesEvent
|
2017-05-26 17:04:51
|
coreos/etcd
|
https://api.github.com/repos/coreos/etcd
|
opened
|
TestCtlV2GetRoleUserWithProxy: User username does not exist
|
area/testing
|
via semaphore (https://semaphoreci.com/coreos/etcd/branches/release-3-2/builds/4)
```
--- FAIL: TestCtlV2GetRoleUserWithProxy (1.18s)
ctl_v2_test.go:183: failed to get user (read /dev/ptmx: input/output error (expected "User: username", got ["auth: User username does not exist.\r\n"]))
```
|
1.0
|
TestCtlV2GetRoleUserWithProxy: User username does not exist - via semaphore (https://semaphoreci.com/coreos/etcd/branches/release-3-2/builds/4)
```
--- FAIL: TestCtlV2GetRoleUserWithProxy (1.18s)
ctl_v2_test.go:183: failed to get user (read /dev/ptmx: input/output error (expected "User: username", got ["auth: User username does not exist.\r\n"]))
```
|
non_defect
|
user username does not exist via semaphore fail ctl test go failed to get user read dev ptmx input output error expected user username got
| 0
|
70,864
| 23,348,065,039
|
IssuesEvent
|
2022-08-09 20:03:49
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
opened
|
Orphaned form label within Editorial Workflow component of the CMS
|
Needs refining ⭐️ Sitewide CMS 508/Accessibility 508-defect-2
|
## Description
Within the read only sections in the CMS, some text items in the section are displayed using the `<label>` element. However, there is no associated field causing this to be an orphaned form label.
## Additional Context
It is not clear if there is a pattern here, what is a label and what is not. Perhaps this is determined from where the content comes from originally? I do believe that this is occurring on the Edit screens only not the View screens
## Screenshot

## Accessibility Standard
WCAG version 2.0 A, [Criterion 1.3.1](https://www.w3.org/WAI/WCAG21/Understanding/info-and-relationships.html)
## Acceptance Criteria
- [ ] Determine is there is a pattern for where/when this label appears (is it a change once update or individual updates?)
- [ ] Determine if this text could be better displayed using a different element tag, perhaps `<p>` or even a heading
- [ ] Technical feasibility review
- [ ] Change management consulted
- [ ] Implementation ticket created
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
|
1.0
|
Orphaned form label within Editorial Workflow component of the CMS - ## Description
Within the read only sections in the CMS, some text items in the section are displayed using the `<label>` element. However, there is no associated field causing this to be an orphaned form label.
## Additional Context
It is not clear if there is a pattern here, what is a label and what is not. Perhaps this is determined from where the content comes from originally? I do believe that this is occurring on the Edit screens only not the View screens
## Screenshot

## Accessibility Standard
WCAG version 2.0 A, [Criterion 1.3.1](https://www.w3.org/WAI/WCAG21/Understanding/info-and-relationships.html)
## Acceptance Criteria
- [ ] Determine is there is a pattern for where/when this label appears (is it a change once update or individual updates?)
- [ ] Determine if this text could be better displayed using a different element tag, perhaps `<p>` or even a heading
- [ ] Technical feasibility review
- [ ] Change management consulted
- [ ] Implementation ticket created
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
|
defect
|
orphaned form label within editorial workflow component of the cms description within the read only sections in the cms some text items in the section are displayed using the element however there is no associated field causing this to be an orphaned form label additional context it is not clear if there is a pattern here what is a label and what is not perhaps this is determined from where the content comes from originally i do believe that this is occurring on the edit screens only not the view screens screenshot accessibility standard wcag version a acceptance criteria determine is there is a pattern for where when this label appears is it a change once update or individual updates determine if this text could be better displayed using a different element tag perhaps or even a heading technical feasibility review change management consulted implementation ticket created cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
| 1
|
370,012
| 10,924,144,857
|
IssuesEvent
|
2019-11-22 09:30:36
|
bounswe/bounswe2019group4
|
https://api.github.com/repos/bounswe/bounswe2019group4
|
closed
|
Add request body validation to the endpoints accepting POST requests
|
Back-End Priority: Medium Type: Enhancement
|
Our API can be enhanced with more usage of request body validation middleware wherever possible. We can check our endpoints and do necessary changes in `backend-enhancement-body-validation` branch.
- [x] article endpoints
- [x] comments endpoints
- [x] events endpoints
- [x] authentication endpoints
- [x] portfolio endpoints
- [x] profile endpoints
- [x] trading equipments endpoints
|
1.0
|
Add request body validation to the endpoints accepting POST requests - Our API can be enhanced with more usage of request body validation middleware wherever possible. We can check our endpoints and do necessary changes in `backend-enhancement-body-validation` branch.
- [x] article endpoints
- [x] comments endpoints
- [x] events endpoints
- [x] authentication endpoints
- [x] portfolio endpoints
- [x] profile endpoints
- [x] trading equipments endpoints
|
non_defect
|
add request body validation to the endpoints accepting post requests our api can be enhanced with more usage of request body validation middleware wherever possible we can check our endpoints and do necessary changes in backend enhancement body validation branch article endpoints comments endpoints events endpoints authentication endpoints portfolio endpoints profile endpoints trading equipments endpoints
| 0
|
68,956
| 22,033,506,043
|
IssuesEvent
|
2022-05-28 07:45:40
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Sometimes main timeline messages appear on a thread as well
|
T-Defect
|
### Steps to reproduce
1. Open https://matrix.to/#/!YTvKGNlinIzlkMTVRl:matrix.org/$o_u7bf4h1gbFReQlE17NzfSL180iC_uoA-fCS42HFHA?via=matrix.org&via=privacytools.io&via=mozilla.org
### Outcome
#### What did you expect?
Main timeline messages should not appear on the thread.
#### What happened instead?
Sometimes they appear on the thread as well. I think it is not expected. The issue is reproduced intermittently.


### Operating system
Debian
### Browser information
Firefox ESR 99
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Sometimes main timeline messages appear on a thread as well - ### Steps to reproduce
1. Open https://matrix.to/#/!YTvKGNlinIzlkMTVRl:matrix.org/$o_u7bf4h1gbFReQlE17NzfSL180iC_uoA-fCS42HFHA?via=matrix.org&via=privacytools.io&via=mozilla.org
### Outcome
#### What did you expect?
Main timeline messages should not appear on the thread.
#### What happened instead?
Sometimes they appear on the thread as well. I think it is not expected. The issue is reproduced intermittently.


### Operating system
Debian
### Browser information
Firefox ESR 99
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
sometimes main timeline messages appear on a thread as well steps to reproduce open outcome what did you expect main timeline messages should not appear on the thread what happened instead sometimes they appear on the thread as well i think it is not expected the issue is reproduced intermittently operating system debian browser information firefox esr url for webapp localhost application version develop branch homeserver no response will you send logs no
| 1
|
55,842
| 14,706,327,281
|
IssuesEvent
|
2021-01-04 19:39:29
|
mozilla-extensions/normandy-devtools
|
https://api.github.com/repos/mozilla-extensions/normandy-devtools
|
closed
|
Search box on listing page has some unexpected whitespace
|
defect
|
<img width="234" alt="image" src="https://user-images.githubusercontent.com/987136/98157235-0e1bb280-1ea7-11eb-80d0-c5579a33bfa1.png">
There's a little unexpected whitespace between the icon on the left and the actual input field.
|
1.0
|
Search box on listing page has some unexpected whitespace - <img width="234" alt="image" src="https://user-images.githubusercontent.com/987136/98157235-0e1bb280-1ea7-11eb-80d0-c5579a33bfa1.png">
There's a little unexpected whitespace between the icon on the left and the actual input field.
|
defect
|
search box on listing page has some unexpected whitespace img width alt image src there s a little unexpected whitespace between the icon on the left and the actual input field
| 1
|
41,945
| 10,723,093,770
|
IssuesEvent
|
2019-10-27 16:25:29
|
MDAnalysis/mdanalysis
|
https://api.github.com/repos/MDAnalysis/mdanalysis
|
closed
|
Non-terminal wildcards in select_atoms should raise an error
|
Component-Selections defect
|
**Expected behavior**
Currently, only terminal/stemming wildcards (*) are supported by `u.select_atoms`. Putting the asterisk anywhere else should raise an error or warning of some kind.
**Actual behavior**
Instead, the code ignores anything after the * and fails silently.
**Code to reproduce the behavior**
```python
>>> import MDAnalysis as mda
>>> mda.__version__
'0.20.1'
>>> from MDAnalysis.tests.datafiles import *
>>> psf = mda.Universe(PSF)
>>> psf.select_atoms("resname *E").residues.resnames[:10]
array(['MET', 'ARG', 'ILE', 'ILE', 'LEU', 'LEU', 'GLY', 'ALA', 'PRO',
'GLY'], dtype=object)
>>> psf.select_atoms("resname G*Y").residues.resnames[:10]
array(['GLY', 'GLY', 'GLY', 'GLY', 'GLN', 'GLN', 'GLU', 'GLY', 'GLN',
'GLY'], dtype=object)
```
**Currently version of MDAnalysis**
- Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.20.1
- Which version of Python (`python -V`)? 3.6 or 3.7
- Which operating system? MacOS Mojave
|
1.0
|
Non-terminal wildcards in select_atoms should raise an error - **Expected behavior**
Currently, only terminal/stemming wildcards (*) are supported by `u.select_atoms`. Putting the asterisk anywhere else should raise an error or warning of some kind.
**Actual behavior**
Instead, the code ignores anything after the * and fails silently.
**Code to reproduce the behavior**
```python
>>> import MDAnalysis as mda
>>> mda.__version__
'0.20.1'
>>> from MDAnalysis.tests.datafiles import *
>>> psf = mda.Universe(PSF)
>>> psf.select_atoms("resname *E").residues.resnames[:10]
array(['MET', 'ARG', 'ILE', 'ILE', 'LEU', 'LEU', 'GLY', 'ALA', 'PRO',
'GLY'], dtype=object)
>>> psf.select_atoms("resname G*Y").residues.resnames[:10]
array(['GLY', 'GLY', 'GLY', 'GLY', 'GLN', 'GLN', 'GLU', 'GLY', 'GLN',
'GLY'], dtype=object)
```
**Currently version of MDAnalysis**
- Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.20.1
- Which version of Python (`python -V`)? 3.6 or 3.7
- Which operating system? MacOS Mojave
|
defect
|
non terminal wildcards in select atoms should raise an error expected behavior currently only terminal stemming wildcards are supported by u select atoms putting the asterisk anywhere else should raise an error or warning of some kind actual behavior instead the code ignores anything after the and fails silently code to reproduce the behavior python import mdanalysis as mda mda version from mdanalysis tests datafiles import psf mda universe psf psf select atoms resname e residues resnames array met arg ile ile leu leu gly ala pro gly dtype object psf select atoms resname g y residues resnames array gly gly gly gly gln gln glu gly gln gly dtype object currently version of mdanalysis which version are you using run python c import mdanalysis as mda print mda version which version of python python v or which operating system macos mojave
| 1
|
53,990
| 13,283,942,764
|
IssuesEvent
|
2020-08-24 04:57:29
|
angular-hispano/angular-hispano
|
https://api.github.com/repos/angular-hispano/angular-hispano
|
closed
|
Actualice el servidor de Mattermost
|
defecto mattermost
|
**¿Está su solicitud de función relacionada con un problema? Por favor describa.**
La aplicación Mattermost para Android me dio un error el otro día diciendo que el servidor necesita ser actualizado para ser compatible con la aplicación.
|
1.0
|
Actualice el servidor de Mattermost - **¿Está su solicitud de función relacionada con un problema? Por favor describa.**
La aplicación Mattermost para Android me dio un error el otro día diciendo que el servidor necesita ser actualizado para ser compatible con la aplicación.
|
defect
|
actualice el servidor de mattermost ¿está su solicitud de función relacionada con un problema por favor describa la aplicación mattermost para android me dio un error el otro día diciendo que el servidor necesita ser actualizado para ser compatible con la aplicación
| 1
|
330,287
| 24,254,764,978
|
IssuesEvent
|
2022-09-27 16:47:48
|
yearn/yearn-exporter
|
https://api.github.com/repos/yearn/yearn-exporter
|
closed
|
fix: update readme with all currently used variables
|
documentation enhancement good first issue p3
|
Not all env variables are currently shown.
Miss the multichain node providers eg FTM_EXPLORER
|
1.0
|
fix: update readme with all currently used variables - Not all env variables are currently shown.
Miss the multichain node providers eg FTM_EXPLORER
|
non_defect
|
fix update readme with all currently used variables not all env variables are currently shown miss the multichain node providers eg ftm explorer
| 0
|
72,599
| 24,197,058,390
|
IssuesEvent
|
2022-09-24 02:59:21
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
BLD: issues while making macosx_arm64 natively on cirrus-ci
|
defect
|
### Describe your issue.
This is a catch-all issue for problems experienced while building `macosx_arm64` natively on cirrus-ci. I'm trying this in anticipation of #17029 being merged (native builds of `linux_aarch64`). The work is being done in https://github.com/andyfaff/scipy/pull/33, which is a branch off of the first PR.
I'm running into a few problems that span the cibuildwheel/meson/CI/pip ecosystem.
@rgommers, @eli-schwartz, the first problem is an architecture problem. Two matrix entries are for cp38 and cp39, see the build log [here](https://api.cirrus-ci.com/v1/task/4544934911934464/logs/cibuildwheel.log). For each of the builds meson thinks its doing a native build for `x86_64`, even though I'm on an `arm64` machine:
```
Build type: native build
Project name: SciPy
Project version: 1.10.0.dev0
C compiler for the host machine: cc (clang 13.1.6 "Apple clang version 13.1.6 (clang-1316.0.21.2.3)")
C linker for the host machine: cc ld64 762
C++ compiler for the host machine: c++ (clang 13.1.6 "Apple clang version 13.1.6 (clang-1316.0.21.2.3)")
C++ linker for the host machine: c++ ld64 762
Host machine cpu family: x86_64
Host machine cpu: x86_64
```
The cibuildwheel infrastructure thinks the build machine is arm64:
```
platform: 'macos'
architectures: {<Architecture.arm64: 'arm64'>}
```
Prior to the cibuildwheel step I also query the machine (with a different Python interpreter installed using brew):
```
> python -c "import platform;print(platform.python_version());print(platform.system());print(platform.machine())"
3.10.4
Darwin
arm64
> uname -m
arm64
> clang --version
Apple clang version 13.1.6 (clang-1316.0.21.2.3)
Target: arm64-apple-darwin21.4.0
Thread model: posix
InstalledDir: /Applications/Xcode-13.3.1.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin
```
How does meson pick up the host machine architecture?
I also note that the build process is trying to install `numpy-1.19.5-cp38-cp38-macosx_10_9_x86_64.whl`:
```
Ignoring numpy: markers 'python_version == "3.8" and platform_machine == "arm64" and platform_system == "Darwin"' don't match your environment
```
How can pip try to install an `x86_64` wheel onto an `arm64` machine? Why doesn't the pattern match; the system is `Darwin`, the machine is `arm64`, the python version is `3.8.10`? Does the `python_version` string have to match `3.8` exactly?
I then need to point out that things appear to work when using [Python 3.10 and 3.11](https://api.cirrus-ci.com/v1/task/4521218773614592/logs/cibuildwheel.log):
```
Build type: native build
Project name: SciPy
Project version: 1.10.0.dev0
C compiler for the host machine: cc (clang 13.1.6 "Apple clang version 13.1.6 (clang-1316.0.21.2.3)")
C linker for the host machine: cc ld64 762
C++ compiler for the host machine: c++ (clang 13.1.6 "Apple clang version 13.1.6 (clang-1316.0.21.2.3)")
C++ linker for the host machine: c++ ld64 762
Host machine cpu family: aarch64
Host machine cpu: arm64
```
### Reproducing Code Example
```python
N/A
```
### Error message
```shell
N/A
```
### SciPy/NumPy/Python version information
N/A
|
1.0
|
BLD: issues while making macosx_arm64 natively on cirrus-ci - ### Describe your issue.
This is a catch-all issue for problems experienced while building `macosx_arm64` natively on cirrus-ci. I'm trying this in anticipation of #17029 being merged (native builds of `linux_aarch64`). The work is being done in https://github.com/andyfaff/scipy/pull/33, which is a branch off of the first PR.
I'm running into a few problems that span the cibuildwheel/meson/CI/pip ecosystem.
@rgommers, @eli-schwartz, the first problem is an architecture problem. Two matrix entries are for cp38 and cp39, see the build log [here](https://api.cirrus-ci.com/v1/task/4544934911934464/logs/cibuildwheel.log). For each of the builds meson thinks its doing a native build for `x86_64`, even though I'm on an `arm64` machine:
```
Build type: native build
Project name: SciPy
Project version: 1.10.0.dev0
C compiler for the host machine: cc (clang 13.1.6 "Apple clang version 13.1.6 (clang-1316.0.21.2.3)")
C linker for the host machine: cc ld64 762
C++ compiler for the host machine: c++ (clang 13.1.6 "Apple clang version 13.1.6 (clang-1316.0.21.2.3)")
C++ linker for the host machine: c++ ld64 762
Host machine cpu family: x86_64
Host machine cpu: x86_64
```
The cibuildwheel infrastructure thinks the build machine is arm64:
```
platform: 'macos'
architectures: {<Architecture.arm64: 'arm64'>}
```
Prior to the cibuildwheel step I also query the machine (with a different Python interpreter installed using brew):
```
> python -c "import platform;print(platform.python_version());print(platform.system());print(platform.machine())"
3.10.4
Darwin
arm64
> uname -m
arm64
> clang --version
Apple clang version 13.1.6 (clang-1316.0.21.2.3)
Target: arm64-apple-darwin21.4.0
Thread model: posix
InstalledDir: /Applications/Xcode-13.3.1.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin
```
How does meson pick up the host machine architecture?
I also note that the build process is trying to install `numpy-1.19.5-cp38-cp38-macosx_10_9_x86_64.whl`:
```
Ignoring numpy: markers 'python_version == "3.8" and platform_machine == "arm64" and platform_system == "Darwin"' don't match your environment
```
How can pip try to install an `x86_64` wheel onto an `arm64` machine? Why doesn't the pattern match; the system is `Darwin`, the machine is `arm64`, the python version is `3.8.10`? Does the `python_version` string have to match `3.8` exactly?
I then need to point out that things appear to work when using [Python 3.10 and 3.11](https://api.cirrus-ci.com/v1/task/4521218773614592/logs/cibuildwheel.log):
```
Build type: native build
Project name: SciPy
Project version: 1.10.0.dev0
C compiler for the host machine: cc (clang 13.1.6 "Apple clang version 13.1.6 (clang-1316.0.21.2.3)")
C linker for the host machine: cc ld64 762
C++ compiler for the host machine: c++ (clang 13.1.6 "Apple clang version 13.1.6 (clang-1316.0.21.2.3)")
C++ linker for the host machine: c++ ld64 762
Host machine cpu family: aarch64
Host machine cpu: arm64
```
### Reproducing Code Example
```python
N/A
```
### Error message
```shell
N/A
```
### SciPy/NumPy/Python version information
N/A
|
defect
|
bld issues while making macosx natively on cirrus ci describe your issue this is a catch all issue for problems experienced while building macosx natively on cirrus ci i m trying this in anticipation of being merged native builds of linux the work is being done in which is a branch off of the first pr i m running into a few problems that span the cibuildwheel meson ci pip ecosystem rgommers eli schwartz the first problem is an architecture problem two matrix entries are for and see the build log for each of the builds meson thinks its doing a native build for even though i m on an machine build type native build project name scipy project version c compiler for the host machine cc clang apple clang version clang c linker for the host machine cc c compiler for the host machine c clang apple clang version clang c linker for the host machine c host machine cpu family host machine cpu the cibuildwheel infrastructure thinks the build machine is platform macos architectures prior to the cibuildwheel step i also query the machine with a different python interpreter installed using brew python c import platform print platform python version print platform system print platform machine darwin uname m clang version apple clang version clang target apple thread model posix installeddir applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin how does meson pick up the host machine architecture i also note that the build process is trying to install numpy macosx whl ignoring numpy markers python version and platform machine and platform system darwin don t match your environment how can pip try to install an wheel onto an machine why doesn t the pattern match the system is darwin the machine is the python version is does the python version string have to match exactly i then need to point out that things appear to work when using build type native build project name scipy project version c compiler for the host machine cc clang apple clang version clang c linker for the host machine cc c compiler for the host machine c clang apple clang version clang c linker for the host machine c host machine cpu family host machine cpu reproducing code example python n a error message shell n a scipy numpy python version information n a
| 1
|
64,552
| 18,737,669,353
|
IssuesEvent
|
2021-11-04 09:47:12
|
primefaces/primereact
|
https://api.github.com/repos/primefaces/primereact
|
closed
|
InputNumber doesn't submit a form
|
defect
|
**I'm submitting a ...**
```
[ x ] bug report
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57
```
**Codesandbox Case (Bug Reports)**
https://codesandbox.io/s/primereact-issue-inputnumber-submit-5rgpz?file=/src/index.tsx
**Current behavior**
InputNumber pressing Enter doesn't submit a form it is in. InputText does submit it. So does native inputs.
**Expected behavior**
Pressing Enter within focused InputNumber component should submit a form.
* **React version:** 16
* **PrimeReact version:** 6.6.0
* **Browser:** [Chrome XX]
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
|
1.0
|
InputNumber doesn't submit a form - **I'm submitting a ...**
```
[ x ] bug report
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57
```
**Codesandbox Case (Bug Reports)**
https://codesandbox.io/s/primereact-issue-inputnumber-submit-5rgpz?file=/src/index.tsx
**Current behavior**
InputNumber pressing Enter doesn't submit a form it is in. InputText does submit it. So does native inputs.
**Expected behavior**
Pressing Enter within focused InputNumber component should submit a form.
* **React version:** 16
* **PrimeReact version:** 6.6.0
* **Browser:** [Chrome XX]
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
|
defect
|
inputnumber doesn t submit a form i m submitting a bug report feature request support request please do not submit support request here instead see codesandbox case bug reports current behavior inputnumber pressing enter doesn t submit a form it is in inputtext does submit it so does native inputs expected behavior pressing enter within focused inputnumber component should submit a form react version primereact version browser language
| 1
|
214,657
| 16,602,898,159
|
IssuesEvent
|
2021-06-01 22:13:57
|
MDAnalysis/mdacli
|
https://api.github.com/repos/MDAnalysis/mdacli
|
closed
|
Review tests cancellation after py38 Successfull in MAC and WIN
|
testing
|
The implementations here look interesting.
With a quick look I can't understand why tests were cancelled on MAC and Windows after py38 was success. Need to investigate this in another PR.
_Originally posted by @joaomcteixeira in https://github.com/MDAnalysis/mdacli/issues/28#issuecomment-849531195_
|
1.0
|
Review tests cancellation after py38 Successfull in MAC and WIN - The implementations here look interesting.
With a quick look I can't understand why tests were cancelled on MAC and Windows after py38 was success. Need to investigate this in another PR.
_Originally posted by @joaomcteixeira in https://github.com/MDAnalysis/mdacli/issues/28#issuecomment-849531195_
|
non_defect
|
review tests cancellation after successfull in mac and win the implementations here look interesting with a quick look i can t understand why tests were cancelled on mac and windows after was success need to investigate this in another pr originally posted by joaomcteixeira in
| 0
|
77,445
| 26,992,957,267
|
IssuesEvent
|
2023-02-09 21:34:44
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
closed
|
Create admin tool bar focus state on staging environment
|
CMS Team 508/Accessibility 508-defect-1
|
## Background
On staging, the admin toolbar has no visible focus indicator on the elements within the toolbar.
Image of defective area:

## Accessibility Standard
WCAG version 2.0 A, [Criterion 2.4.3](https://www.w3.org/WAI/WCAG21/Understanding/focus-order.html#:~:text=3%3A%20Focus%20Order-,Success%20Criterion%202.4.,that%20preserves%20meaning%20and%20operability.)
## Acceptance Criteria
- [x] Determine if this is an upstream issue
- [x] When elements in the toolbar receive focus, a visible focus indicator should be shown
- [x] tech feasibility review
- [x] If design is needed, separate ticket should be created for the UX (no ticket required)
- [ ] change management is consulted
- [ ] Use the current tool bar hover state and add a standard green outline to create a focus state for the tool bar.
|
1.0
|
Create admin tool bar focus state on staging environment - ## Background
On staging, the admin toolbar has no visible focus indicator on the elements within the toolbar.
Image of defective area:

## Accessibility Standard
WCAG version 2.0 A, [Criterion 2.4.3](https://www.w3.org/WAI/WCAG21/Understanding/focus-order.html#:~:text=3%3A%20Focus%20Order-,Success%20Criterion%202.4.,that%20preserves%20meaning%20and%20operability.)
## Acceptance Criteria
- [x] Determine if this is an upstream issue
- [x] When elements in the toolbar receive focus, a visible focus indicator should be shown
- [x] tech feasibility review
- [x] If design is needed, separate ticket should be created for the UX (no ticket required)
- [ ] change management is consulted
- [ ] Use the current tool bar hover state and add a standard green outline to create a focus state for the tool bar.
|
defect
|
create admin tool bar focus state on staging environment background on staging the admin toolbar has no visible focus indicator on the elements within the toolbar image of defective area accessibility standard wcag version a acceptance criteria determine if this is an upstream issue when elements in the toolbar receive focus a visible focus indicator should be shown tech feasibility review if design is needed separate ticket should be created for the ux no ticket required change management is consulted use the current tool bar hover state and add a standard green outline to create a focus state for the tool bar
| 1
|
551,898
| 16,190,874,555
|
IssuesEvent
|
2021-05-04 08:18:39
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.youtube.com - see bug description
|
browser-fixme ml-needsdiagnosis-false ml-probability-high priority-critical
|
<!-- @browser: firefox app -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/72553 -->
**URL**: https://www.youtube.com/
**Browser / Version**: firefox app
**Operating System**: ipad
**Tested Another Browser**: Yes Safari
**Problem type**: Something else
**Description**: Fullscreen not available
**Steps to Reproduce**:
fullcreen is not possible!
............
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.youtube.com - see bug description - <!-- @browser: firefox app -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/72553 -->
**URL**: https://www.youtube.com/
**Browser / Version**: firefox app
**Operating System**: ipad
**Tested Another Browser**: Yes Safari
**Problem type**: Something else
**Description**: Fullscreen not available
**Steps to Reproduce**:
fullcreen is not possible!
............
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
see bug description url browser version firefox app operating system ipad tested another browser yes safari problem type something else description fullscreen not available steps to reproduce fullcreen is not possible browser configuration none from with ❤️
| 0
|
13,736
| 2,780,096,849
|
IssuesEvent
|
2015-05-06 00:27:57
|
larcenists/larceny
|
https://api.github.com/repos/larcenists/larceny
|
closed
|
-fno-stack-protector flag breaks Sparc Petit build
|
C: RUNTIME P: major R: wontfix T: defect
|
**Reported by: pnkfelix on Mon Apr 27 10:41:19 2009**
Our Sparc Petit Larceny builds have been failing for a few nights. Unfortunately the current CDash configuration does not allow one to easily access the build log files, but here's the relevant error message
```
Creating final image in "petit.heap"...
gcc -c -ISys -I../../include -I../../include/Sys -I../../include/Shared -O3 -DN
DEBUG2 -c -fno-stack-protector -falign-functions=4 -I../../include/Standard-C
-o Sys/larceny.o Sys/larceny.c
cc1: unrecognized option `-fno-stack-protector'
make: *** [Sys/larceny.o] Error 1
```
Clearly changeset:6190 had a wider effect than I had intended.
(I had some nonsense written down about how this was not a gcc build, but it obviously is.)
The problem is probably that the version of gcc on the CCIS Solaris boxes is version 3.2.3, from 2002... so perhaps the makefile building system needs to be aware of the gcc version. Or the makefile itself needs to query what version of gcc we are running. Or we need to start using ```./configure``` as our users are clamoring for (but I am not eager to make such a change)
|
1.0
|
-fno-stack-protector flag breaks Sparc Petit build - **Reported by: pnkfelix on Mon Apr 27 10:41:19 2009**
Our Sparc Petit Larceny builds have been failing for a few nights. Unfortunately the current CDash configuration does not allow one to easily access the build log files, but here's the relevant error message
```
Creating final image in "petit.heap"...
gcc -c -ISys -I../../include -I../../include/Sys -I../../include/Shared -O3 -DN
DEBUG2 -c -fno-stack-protector -falign-functions=4 -I../../include/Standard-C
-o Sys/larceny.o Sys/larceny.c
cc1: unrecognized option `-fno-stack-protector'
make: *** [Sys/larceny.o] Error 1
```
Clearly changeset:6190 had a wider effect than I had intended.
(I had some nonsense written down about how this was not a gcc build, but it obviously is.)
The problem is probably that the version of gcc on the CCIS Solaris boxes is version 3.2.3, from 2002... so perhaps the makefile building system needs to be aware of the gcc version. Or the makefile itself needs to query what version of gcc we are running. Or we need to start using ```./configure``` as our users are clamoring for (but I am not eager to make such a change)
|
defect
|
fno stack protector flag breaks sparc petit build reported by pnkfelix on mon apr our sparc petit larceny builds have been failing for a few nights unfortunately the current cdash configuration does not allow one to easily access the build log files but here s the relevant error message creating final image in petit heap gcc c isys i include i include sys i include shared dn c fno stack protector falign functions i include standard c o sys larceny o sys larceny c unrecognized option fno stack protector make error clearly changeset had a wider effect than i had intended i had some nonsense written down about how this was not a gcc build but it obviously is the problem is probably that the version of gcc on the ccis solaris boxes is version from so perhaps the makefile building system needs to be aware of the gcc version or the makefile itself needs to query what version of gcc we are running or we need to start using configure as our users are clamoring for but i am not eager to make such a change
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.