Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
76,151 | 21,191,138,242 | IssuesEvent | 2022-04-08 17:33:12 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | build: strip the produced cockroach-sql binary | C-enhancement A-build-system T-dev-inf | Followup to #75800; the `cockroach-sql` binary should be stripped to ensure it remains small.
Epic: CC-5993 | 1.0 | build: strip the produced cockroach-sql binary - Followup to #75800; the `cockroach-sql` binary should be stripped to ensure it remains small.
Epic: CC-5993 | non_defect | build strip the produced cockroach sql binary followup to the cockroach sql binary should be stripped to ensure it remains small epic cc | 0 |
4,247 | 7,006,045,109 | IssuesEvent | 2017-12-19 06:25:15 | GiselleSerate/myaliases | https://api.github.com/repos/GiselleSerate/myaliases | opened | check for setup/uninstall script compatibility on osx | compatibility | Please do the following when you get access to a lab mac:
- [ ] setup works!
- [ ] uninstall works!
- [ ] documentation for this process written! | True | check for setup/uninstall script compatibility on osx - Please do the following when you get access to a lab mac:
- [ ] setup works!
- [ ] uninstall works!
- [ ] documentation for this process written! | non_defect | check for setup uninstall script compatibility on osx please do the following when you get access to a lab mac setup works uninstall works documentation for this process written | 0 |
49,991 | 13,187,304,059 | IssuesEvent | 2020-08-13 02:59:22 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | Coverage: no gcda files are found (Trac #2414) | Incomplete Migration Migrated from Trac combo core defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2414">https://code.icecube.wisc.edu/ticket/2414</a>, reported by fiedl and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2020-06-24T12:28:40",
"description": "I\u2019ve tried to generate a coverage report (https://github.com/IceCube-SPNO/IceTrayCombo/blob/master/docs/general/test_guides/testing_coverage.rst). But \u201cno gcda files\u201d are found.\n\nSee: https://icecube-spno.slack.com/archives/C02KQL9KN/p1582820079064000\n\n{{{\n[2020-02-27 14:38:33] fiedl@fiedl-mbp ~/icecube/combo fiedl/monopole-generator\n\u25b6 brew install lcov\n\n[2020-02-27 15:26:10] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator\n\u25b6 mkdir coverage\n\n[2020-02-27 15:26:16] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 cmake -D CMAKE_BUILD_TYPE=Coverage -D SYSTEM_PACKAGES=true -D Python_ROOT=/usr/local/opt/python/ ~/icecube/combo\n\n[2020-02-27 15:28:30] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 ./env-shell.sh make monopole-generator-test -j 4\n\n[2020-02-27 15:40:24] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 lcov -d . -z\nDeleting all .da files in . and subdirectories\nDone.\n\n[2020-02-27 15:40:33] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 lcov -d . -c -i -o test_base.info\nCapturing coverage data from .\nFound LLVM gcov version 11.0.0, which emulates gcov version 4.2.0\nScanning . for .gcno files ...\ngeninfo: WARNING: no .gcno files found in . - skipping!\nFinished .info-file creation\n\n[2020-02-27 15:44:06] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 ./env-shell.sh ctest -R 'monopole-generator::' --verbose\n\n[2020-02-27 16:42:40] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 lcov -d . -c -o test_run.info\nCapturing coverage data from .\nFound LLVM gcov version 11.0.0, which emulates gcov version 4.2.0\nScanning . for .gcda files ...\ngeninfo: WARNING: no .gcda files found in . - skipping!\nFinished .info-file creation\n\n[2020-02-28 00:56:06] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 cmake --version\ncmake version 3.16.4\n}}}\n\nI've also tried the full `make coverage`. Same result.\n\nEntire cmake output:\n\n{{{\n[2020-02-27 16:01:56] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 cmake -D CMAKE_BUILD_TYPE=Coverage -D SYSTEM_PACKAGES=true -D Python_ROOT=/usr/local/opt/python/ ~/icecube/combo\n-- The C compiler identification is AppleClang 11.0.0.11000033\n-- The CXX compiler identification is AppleClang 11.0.0.11000033\n-- Check for working C compiler: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/cc\n-- Check for working C compiler: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/cc -- works\n-- Detecting C compiler ABI info\n-- Detecting C compiler ABI info - done\n-- Detecting C compile features\n-- Detecting C compile features - done\n-- Check for working CXX compiler: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/c++\n-- Check for working CXX compiler: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/c++ -- works\n-- Detecting CXX compiler ABI info\n-- Detecting CXX compiler ABI info - done\n-- Detecting CXX compile features\n-- Detecting CXX compile features - done\nCMake Deprecation Warning at CMakeLists.txt:39 (cmake_policy):\n The OLD behavior for policy CMP0046 will be removed from a future version\n of CMake.\n\n The cmake-policies(7) manual explains that the OLD behaviors of all\n policies are deprecated and that a policy should be set to OLD only under\n specific short-term circumstances. Projects should be ported to the NEW\n behavior and not rely on setting a policy to OLD.\n\n\n--\n-- IceCube Configuration starting\n--\nUnable to determine upstream SVN information from working tree history\n\n-- OSTYPE = Darwin\n-- OSVERSION = 19.3.0\n-- ARCH = x86_64\n-- BUILDNAME = Darwin-19.3.0/x86_64/clang-11.0.0\n-- TOOLSET = clang-11.0.0/x86_64/Coverage\n-- HOSTNAME = fiedl-mbp.local\n-- CMake path = /usr/local/Cellar/cmake/3.16.4/bin/cmake\n-- CMake version = 3.16.4\n-- SVN_REVISION = 0\n-- SVN_URL = Unknown\n-- META_PROJECT = Unknown\n--\n-- Setting compiler, compile drivers, and linker\n--\n-- distcc found at /usr/local/bin/distcc\n-- ccache found at /usr/local/bin/ccache\n-- Performing Test CXX_HAS_Wdeprecated\n-- Performing Test CXX_HAS_Wdeprecated - Success\n-- Performing Test CXX_HAS_Wunused_variable\n-- Performing Test CXX_HAS_Wunused_variable - Success\n-- Performing Test CXX_HAS_Wunused_local_typedef\n-- Performing Test CXX_HAS_Wunused_local_typedef - Success\n-- Performing Test CXX_HAS_Wunused_local_typedefs\n-- Performing Test CXX_HAS_Wunused_local_typedefs - Success\n-- Performing Test CXX_HAS_Wpotentially_evaluated_expression\n-- Performing Test CXX_HAS_Wpotentially_evaluated_expression - Success\n-- Setting default compiler flags and build type.\n-- *** I3_TESTDATA is not set. Using the default value of /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage/test-data\n--\n-- Configuring tools...\n--\n-- Using system packages.\n-- Using default site cmake dir of /usr/share/fizzicks/cmake\n--\n-- root\n-- + ROOT not found or disabled: building without ROOT support\n--\n-- python\n-- Found Python: /usr/local/opt/python/bin/python3.7 (found version \"3.7.6\") found components: Interpreter Development NumPy\n-- + version: 3.7.6\n-- + binary: /usr/local/opt/python/bin/python3.7\n-- + includes: /usr/local/Cellar/python/3.7.6_1/Frameworks/Python.framework/Versions/3.7/include/python3.7m\n-- + libs: /usr/local/Cellar/python/3.7.6_1/Frameworks/Python.framework/Versions/3.7/lib/libpython3.7m.dylib\n-- + numpy: /usr/local/lib/python3.7/site-packages/numpy/core/include\n-- + scipy: FOUND\n--\n-- Boost\n-- Looking for pthread.h\n-- Looking for pthread.h - found\n-- Performing Test CMAKE_HAVE_LIBC_PTHREAD\n-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success\n-- Found Threads: TRUE\n-- Found Boost: /usr/local/include (found version \"1.71.0\") found components: system thread date_time filesystem program_options regex iostreams python37 chrono atomic\n--\n-- boostnumpy\n-- - boost/numpy.hpp not found in include\n-- - boost_numpy\n--\n-- blas and lapack\n-- Looking for dgemm_\n-- Looking for dgemm_ - found\n-- Found BLAS: /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/System/Library/Frameworks/Accelerate.framework\n-- Looking for cheev_\n-- Looking for cheev_ - found\n-- A library with LAPACK API found.\n--\n-- gsl\n-- + gsl/gsl_rng.h found at /usr/local/include\n-- + /usr/local/lib/libgsl.dylib\n--\n-- sprng\n-- - sprng/sprng.h not found in include/sprng-2.0a\n-- - sprng\n--\n-- sprng\n-- - sprng/sprng.h not found in include\n-- - sprng\n--\n-- erfa\n-- - erfa.h not found in include\n-- - erfa\n--\n-- pal\n-- - star/pal.h not found in include\n-- - pal\n--\n-- pal\n-- - star/pal.h not found in include\n-- - starlink_pal\n--\n-- libarchive\n-- + /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libarchive.tbd\n-- Found MySQL: /usr/local/lib/libmysqlclient.dylib\n--\n-- suitesparse\n-- - cholmod.h not found in include\n-- - camd\n-- - ccolamd\n-- - spqr\n-- - cholmod\n-- - amd\n-- - colamd\n-- + /usr/local/lib/libtbb.dylib\n-- - suitesparseconfig\n--\n-- suitesparse\n-- - suitesparse/cholmod.h not found in include\n-- - camd\n-- - ccolamd\n-- - spqr\n-- - cholmod\n-- - amd\n-- - colamd\n-- - suitesparseconfig\n--\n-- suitesparse\n-- - suitesparse/cholmod.h not found in include\n-- - camd\n-- - ccolamd\n-- - spqr\n-- - cholmod\n-- - amd\n-- - colamd\n--\n-- ncurses\n-- Looking for wsyncup in /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libcurses.tbd\n-- Looking for wsyncup in /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libcurses.tbd - found\n-- Looking for cbreak in /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libncurses.tbd\n-- Looking for cbreak in /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libncurses.tbd - found\n-- Found Curses: /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libncurses.tbd\n-- + ncurses.h found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include\n-- + libncurses found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libncurses.tbd\n--\n-- cdk\n-- - cdk/cdk.h not found in include\n-- - cdk\n--\n-- cdk\n-- - cdk.h not found in include\n-- - cdk\n--\n-- cfitsio\n-- + fitsio.h found at /usr/local/include\n-- + /usr/local/lib/libcfitsio.dylib\n--\n-- healpix-cxx\n-- - healpix_cxx/healpix_map.h not found in include\n-- - healpix_cxx\n--\n-- hdf5\n-- - hdf5.h not found in include\n-- - hdf5\n-- - hdf5_hl\n--\n-- hdf5\n-- - hdf5.h not found in /usr/include/hdf5/serial\n-- - hdf5_serial\n-- - hdf5_serial_hl\n-- - Minuit2: USE_ROOT set to FALSE\n--\n-- clhep\n-- + CLHEP/ClhepVersion.h found at /usr/local/include\n-- + /usr/local/lib/libCLHEP.dylib\n-- Looking for Geant4 geant4-config program\n-- Looking for Geant4 geant4-config program -- not found\n-- Looking for Geant4 liblist program\n-- Looking for Geant4 liblist program -- not found\n--\n-- zlib\n-- + zlib.h found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include\n-- + /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libz.tbd\n--\n-- OpenCL\n-- + Using the OpenCL Framework because we're on Apple\n-- + cl.h found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/System/Library/Frameworks/OpenCL.framework/Headers\n-- + OpenCL framework found at -framework OpenCL\n-- Looking for CL_VERSION_2_0\n-- Looking for CL_VERSION_2_0 - not found\n-- Looking for CL_VERSION_1_2\n-- Looking for CL_VERSION_1_2 - found\n--\n-- gmp\n-- + gmp.h found at /usr/local\n-- + /usr/local/lib/libgmp.dylib\n--\n-- log4cpp\n-- + log4cpp/Category.hh found at /usr/local/include\n-- + /usr/local/lib/liblog4cpp.dylib\n--\n-- xml2\n-- + libxml/parser.h found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/libxml2\n-- + /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libxml2.tbd\n--\n-- genie\n-- Looking for Genie genie-config program -- not found\n--\n-- zmq\n-- - zmq.hpp not found in ZMQ_INCLUDE_DIR-NOTFOUND\n-- + /usr/local/lib/libzmq.dylib\n--\n-- multinest\n-- - multinest.h not found in include\n-- - multinest\n--\n-- nlopt\n-- - nlopt.hpp not found in include\n-- - nlopt\n--\n-- zstd\n-- - zstd.h not found in ZSTD_INCLUDE_DIR-NOTFOUND\n-- - zstd\n--\n-- fftw3\n-- - fftw3.h not found in include\n-- - fftw3\n--\n-- Configuring projects:\n--\n-- + BadDomList\n-- +-- python [symlinks]\n-- + CascadeVariables\n-- +-- python [symlinks]\n-- + CoincSuite\n-- +-- python [symlinks]\n-- +-- CoincSuite-pybindings\n-- + CommonVariables\n-- +-- python [symlinks]\n-- +-- Include \"direct_hits\" sub project.\n-- +-- Include \"hit_multiplicity\" sub project.\n-- +-- Include \"hit_statistics\" sub project.\n-- +-- Include \"track_characteristics\" sub project.\n-- +-- Include \"time_characteristics\" sub project.\n-- +-- common_variables-pybindings\n-- +-- common_variables__hit_multiplicity-pybindings\n-- +-- common_variables__track_characteristics-pybindings\n-- +-- common_variables__hit_statistics-pybindings\n-- +-- common_variables__direct_hits-pybindings\n-- +-- common_variables__time_characteristics-pybindings\n-- + DOMLauncher\n-- +-- python [symlinks]\n-- +-- DOMLauncher-pybindings\n-- + DeepCore_Filter\n-- +-- python [symlinks]\n-- + DomTools\n-- +-- python [symlinks]\n-- + HiveSplitter\n-- +-- python [symlinks]\n-- + IceHive\n-- +-- python [symlinks]\n-- +-- IceHive-pybindings\n-- + KalmanFilter\n-- +-- python [symlinks]\n-- + MuonGun\n-- +-- python [symlinks]\n-- +-- Muonitron table-maker\n-- +-- MuonGun-pybindings\n-- + NoiseEngine\n-- Skipping NoiseEngine (healpix-cxx not found)\n-- + PROPOSAL\n-- +-- PROPOSAL-pybindings\n-- + SLOPtools\n-- +-- python [symlinks]\n-- + STTools\n-- +-- python [symlinks]\n-- +-- STTools-pybindings\n-- + StartingTrackVeto\n-- +-- python [symlinks]\n-- + TopologicalSplitter\n-- +-- python [symlinks]\n-- + VHESelfVeto\n-- +-- python [symlinks]\n-- +-- VHESelfVeto-pybindings\n-- + WaveCalibrator\n-- +-- python [symlinks]\n-- + astro\n-- - PAL not found, skipping build of icetray\n Ask your admin to install PAL. Instructions for PAL can be found at:\n http://software.icecube.wisc.edu/documentation/projects/cmake/tools/pal.html\n-- + bayesian-priors\n-- +-- python [symlinks]\n-- + clast\n-- +-- python [symlinks]\n-- + clsim\n-- +-- python [symlinks]\n-- --- ZMQ not found. I3CLSimMakePhotons will not function.\n-- +-- tabulator (have OpenCL 1.2)\n-- +-- safeprimes_base32.gz already downloaded\n-- +-- Geant4 or OpenCL is not installed on your system. clsim will fail if it is not used with parameterizations.\n-- +-- gmp support (make_safeprimes utility)\n-- +-- clsim-pybindings\n-- + cmake\n-- + cmc\n-- +-- python [symlinks]\n-- +-- cmc-pybindings\n-- + coinc-twc\n-- +-- python [symlinks]\n-- + core-removal\n-- +-- python [symlinks]\n-- + corsika-reader\n-- +-- python [symlinks]\n-- + Boost >= 6.4 : Adding CORSIKA Client/Server\n-- +-- corsika_reader-pybindings\n-- + cramer-rao\n-- +-- python [symlinks]\n-- + cscd-llh\n-- +-- python [symlinks]\n-- + daq-decode\n-- +-- python [symlinks]\n-- +-- daq_decode-pybindings\n-- + dataclasses\n-- +-- python [symlinks]\n-- +-- dataclasses-pybindings\n-- + dataio\n-- +-- python [symlinks]\n-- +-- dataio-pybindings\n-- +-- dataio-shovel *not* included (missing CDK or ncurses libraries)\n-- +-- test_unregistered-pybindings\n-- + ddddr\n-- +-- python [symlinks]\n-- +-- ddddr-pybindings\n-- + dipolefit\n-- +-- python [symlinks]\n-- + docs\n-- + double-muon\n-- +-- python [symlinks]\n-- + dst\n-- +-- Skipping dst (healpix-cxx not found)\n-- + earthmodel-service\n-- +-- earthmodel_service-pybindings\n-- + fill-ratio\n-- +-- python [symlinks]\n-- + filter-tools\n-- +-- python [symlinks]\n-- + filterscripts\n-- +-- python [symlinks]\n-- +-- astro required to build filterscripts-cxx\n-- + finiteReco\n-- +-- python [symlinks]\n-- + frame_object_diff\n-- +-- python [symlinks]\n-- +-- frame_object_diff-pybindings\n-- + full_event_followup\n-- +-- python [symlinks]\n-- + g4-tankresponse\n-- Looking for Geant4\n-- Looking for Geant4 -- not found\n-- Looking for CLHEP\n-- Looking for CLHEP -- found\n-- - Geant4 and/or CLHEP not found - Skipping build of g4-tankresponse\n-- +-- g4-tankresponse has not been thoroughly tested with Geant4 versions 4.10 and above\n-- + gcdserver\n-- +-- gcdserver does not work. Not building.\n-- + gulliver\n-- +-- python [symlinks]\n-- +-- gulliver-pybindings\n-- + gulliver-bootstrap\n-- +-- python [symlinks]\n-- +-- gulliver_bootstrap-pybindings\n-- + gulliver-modules\n-- +-- python [symlinks]\n-- + hdfwriter\n-- +-- python [symlinks]\n-- +-- hdf5 library not found, skipping build of hdfwriter\n-- *** Ask your admin to install the hdf5 package.\n-- + ice-models\n-- +-- python [symlinks]\n-- + icepick\n-- +-- python [symlinks]\n-- + icetop_Level3_scripts\n-- +-- python [symlinks]\n-- + icetray\n-- +-- libarchive *not* found, omitting optional tarfile support\n-- +-- libarchive *not* found, omitting optional tarfile support\n-- +-- python [symlinks]\n-- +-- memory tracking disabled\n-- +-- icetray-pybindings\n-- + interfaces\n-- +-- interfaces-pybindings\n-- + ipdf\n-- +-- python [symlinks]\n-- +-- ipdf-pybindings\n-- + level3-filter-cascade\n-- +-- python [symlinks]\n-- + level3-filter-muon\n-- +-- python [symlinks]\n-- +-- level3_filter_muon-pybindings\n-- + lilliput\n-- +-- python [symlinks]\n-- + ipdf support\n-- - Minuit2 support...not found\n-- - multinest support...not found\n-- - NLopt support...not found\n-- +-- lilliput-pybindings\n-- + linefit\n-- +-- python [symlinks]\n-- + millipede\n-- +-- SuiteSparse not found. Skipping millipede.\n If you need it you can install SuiteSparse through I3Ports.\n-- + monopole-generator\n-- +-- python [symlinks]\n-- + mue\n-- +-- python [symlinks]\n-- + neutrino-generator\n-- +-- python [symlinks]\n-- +-- /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage/neutrino-generator/cross_section_data does not exist. Downloading...\n-- +-- Downloading and unpacking the crosssection data files...\ntrying to download /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage/neutrino-generator/nugen-v3.1-tables.tgz...\n downloading http://code.icecube.wisc.edu/tools/neutrino-generator/nugen-v3.1-tables.tgz...\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 44.6M 100 44.6M 0 0 3572k 0 0:00:12 0:00:12 --:--:-- 8988k\n downloading http://code.icecube.wisc.edu/tools/neutrino-generator/nugen-v3.1-tables.md5sum...\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 56 100 56 0 0 265 0 --:--:-- --:--:-- --:--:-- 265\n check md5sum ...\n extracting http://code.icecube.wisc.edu/tools/neutrino-generator/nugen-v3.1-tables.tgz ...\n /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage/neutrino-generator/nugen-v3.1-tables.tgz is downloaded.\n-- +-- ...download complete.\n-- +-- genie-icetray NOT found. building without GENIE support.\n-- +-- neutrino_generator-pybindings\n-- + ophelia\n-- +-- python [symlinks]\n-- + paraboloid\n-- +-- python [symlinks]\n-- +-- paraboloid-pybindings\n-- + payload-parsing\n-- +-- python [symlinks]\n-- +-- payload_parsing-pybindings\n-- + photonics-service\n-- +-- python [symlinks]\n-- +-- photospline support\n-- +-- numpy support\n-- +-- photonics_service-pybindings\n-- + photospline\n-- +-- python [symlinks]\n-- +-- photospline-pybindings\n-- + phys-services\n-- +-- python [symlinks]\n-- +-- sprng *not* found, omitting optional SPRNGRandomService\n-- +-- phys_services-pybindings\n-- + polyplopia\n-- +-- python [symlinks]\n-- +-- polyplopia-pybindings\n-- + portia\n-- +-- python [symlinks]\n-- + ppc\n-- +-- python [symlinks]\n-- +-- OpenCL SDK found. Building ppc.\n-- + production-histograms\n-- +-- python [symlinks]\n-- + pybdt\n-- + recclasses\n-- +-- python [symlinks]\n-- +-- recclasses-pybindings\n-- + rootwriter\n-- +-- python [symlinks]\n-- +-- ROOT has been disabled in your configuration.\n-- +-- Skipping build of rootwriter\n-- + serialization\n-- + shield\n-- +-- python [symlinks]\n-- + sim-services\n-- +-- python [symlinks]\n-- +-- sim_services-pybindings\n-- + simclasses\n-- +-- python [symlinks]\n-- + tableio support\n-- +-- simclasses-pybindings\n-- + simprod-scripts\n-- +-- python [symlinks]\n-- + smallshower-filter\n-- + spline-reco\n-- +-- python [symlinks]\n-- +-- spline_reco-pybindings\n-- + static-twc\n-- +-- python [symlinks]\n-- + steamshovel\n-- +-- python [symlinks]\nCMake Deprecation Warning at steamshovel/CMakeLists.txt:13 (cmake_policy):\n The OLD behavior for policy CMP0053 will be removed from a future version\n of CMake.\n\n The cmake-policies(7) manual explains that the OLD behaviors of all\n policies are deprecated and that a policy should be set to OLD only under\n specific short-term circumstances. Projects should be ported to the NEW\n behavior and not rely on setting a policy to OLD.\n\n\n-- Found OpenGL: /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/System/Library/Frameworks/OpenGL.framework\n-- Found GLUT: /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/System/Library/Frameworks/GLUT.framework\n-- +-- shovelart-pybindings\n-- +-- shovelio-pybindings\n-- + stochastics\n-- +-- python [symlinks]\n-- +-- stochastics-pybindings\n-- + tableio\n-- +-- python [symlinks]\n-- +-- tableio-pybindings\n-- + tensor-of-inertia\n-- +-- python [symlinks]\n-- + topeventcleaning\n-- +-- python [symlinks]\n-- + toprec\n-- +-- python [symlinks]\n-- +-- toprec-pybindings\n-- + topsimulator\n-- +-- python [symlinks]\n-- +-- topsimulator-pybindings\n-- + tpx\n-- +-- python [symlinks]\n-- +-- tpx-pybindings\n-- +-- FETCH_TPX set to OFF. SLC calibration data may not be available.\n-- + trigger-sim\n-- +-- python [symlinks]\n-- +-- trigger_sim-pybindings\n-- + trigger-splitter\n-- +-- python [symlinks]\n-- + truncated_energy\n-- +-- python [symlinks]\n-- + vemcal\n-- +-- python [symlinks]\n-- +-- vemcal-pybindings\n-- + vuvuzela\n-- +-- python [symlinks]\n-- +-- vuvuzela-pybindings\n-- + wavedeform\n-- +-- SuiteSparse not found. Skipping wavedeform.\n If you need it you can install SuiteSparse through I3Ports.\n-- + wavereform\n-- +-- python [symlinks]\n-- +-- wavereform-pybindings\n-- + weighting\n-- +-- python [symlinks]\n-- + wimpsim-reader\n-- +-- python [symlinks]\n-- +-- astro not built. Skipping test 'test_sun_position'\n-- Generating env-shell.sh\n-- Generating icetray-config\n-- Configuring 'gfilt' STL decryptor\n-- Configuring done\n-- Generating done\n-- Build files have been written to: /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n}}}",
"reporter": "fiedl",
"cc": "",
"resolution": "fixed",
"_ts": "1593001720016649",
"component": "combo core",
"summary": "Coverage: no gcda files are found",
"priority": "normal",
"keywords": "coverage, cmake",
"time": "2020-02-28T00:01:26",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Coverage: no gcda files are found (Trac #2414) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2414">https://code.icecube.wisc.edu/ticket/2414</a>, reported by fiedl and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2020-06-24T12:28:40",
"description": "I\u2019ve tried to generate a coverage report (https://github.com/IceCube-SPNO/IceTrayCombo/blob/master/docs/general/test_guides/testing_coverage.rst). But \u201cno gcda files\u201d are found.\n\nSee: https://icecube-spno.slack.com/archives/C02KQL9KN/p1582820079064000\n\n{{{\n[2020-02-27 14:38:33] fiedl@fiedl-mbp ~/icecube/combo fiedl/monopole-generator\n\u25b6 brew install lcov\n\n[2020-02-27 15:26:10] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator\n\u25b6 mkdir coverage\n\n[2020-02-27 15:26:16] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 cmake -D CMAKE_BUILD_TYPE=Coverage -D SYSTEM_PACKAGES=true -D Python_ROOT=/usr/local/opt/python/ ~/icecube/combo\n\n[2020-02-27 15:28:30] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 ./env-shell.sh make monopole-generator-test -j 4\n\n[2020-02-27 15:40:24] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 lcov -d . -z\nDeleting all .da files in . and subdirectories\nDone.\n\n[2020-02-27 15:40:33] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 lcov -d . -c -i -o test_base.info\nCapturing coverage data from .\nFound LLVM gcov version 11.0.0, which emulates gcov version 4.2.0\nScanning . for .gcno files ...\ngeninfo: WARNING: no .gcno files found in . - skipping!\nFinished .info-file creation\n\n[2020-02-27 15:44:06] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 ./env-shell.sh ctest -R 'monopole-generator::' --verbose\n\n[2020-02-27 16:42:40] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 lcov -d . -c -o test_run.info\nCapturing coverage data from .\nFound LLVM gcov version 11.0.0, which emulates gcov version 4.2.0\nScanning . for .gcda files ...\ngeninfo: WARNING: no .gcda files found in . - skipping!\nFinished .info-file creation\n\n[2020-02-28 00:56:06] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 cmake --version\ncmake version 3.16.4\n}}}\n\nI've also tried the full `make coverage`. Same result.\n\nEntire cmake output:\n\n{{{\n[2020-02-27 16:01:56] fiedl@fiedl-mbp ~/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n\u25b6 cmake -D CMAKE_BUILD_TYPE=Coverage -D SYSTEM_PACKAGES=true -D Python_ROOT=/usr/local/opt/python/ ~/icecube/combo\n-- The C compiler identification is AppleClang 11.0.0.11000033\n-- The CXX compiler identification is AppleClang 11.0.0.11000033\n-- Check for working C compiler: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/cc\n-- Check for working C compiler: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/cc -- works\n-- Detecting C compiler ABI info\n-- Detecting C compiler ABI info - done\n-- Detecting C compile features\n-- Detecting C compile features - done\n-- Check for working CXX compiler: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/c++\n-- Check for working CXX compiler: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/c++ -- works\n-- Detecting CXX compiler ABI info\n-- Detecting CXX compiler ABI info - done\n-- Detecting CXX compile features\n-- Detecting CXX compile features - done\nCMake Deprecation Warning at CMakeLists.txt:39 (cmake_policy):\n The OLD behavior for policy CMP0046 will be removed from a future version\n of CMake.\n\n The cmake-policies(7) manual explains that the OLD behaviors of all\n policies are deprecated and that a policy should be set to OLD only under\n specific short-term circumstances. Projects should be ported to the NEW\n behavior and not rely on setting a policy to OLD.\n\n\n--\n-- IceCube Configuration starting\n--\nUnable to determine upstream SVN information from working tree history\n\n-- OSTYPE = Darwin\n-- OSVERSION = 19.3.0\n-- ARCH = x86_64\n-- BUILDNAME = Darwin-19.3.0/x86_64/clang-11.0.0\n-- TOOLSET = clang-11.0.0/x86_64/Coverage\n-- HOSTNAME = fiedl-mbp.local\n-- CMake path = /usr/local/Cellar/cmake/3.16.4/bin/cmake\n-- CMake version = 3.16.4\n-- SVN_REVISION = 0\n-- SVN_URL = Unknown\n-- META_PROJECT = Unknown\n--\n-- Setting compiler, compile drivers, and linker\n--\n-- distcc found at /usr/local/bin/distcc\n-- ccache found at /usr/local/bin/ccache\n-- Performing Test CXX_HAS_Wdeprecated\n-- Performing Test CXX_HAS_Wdeprecated - Success\n-- Performing Test CXX_HAS_Wunused_variable\n-- Performing Test CXX_HAS_Wunused_variable - Success\n-- Performing Test CXX_HAS_Wunused_local_typedef\n-- Performing Test CXX_HAS_Wunused_local_typedef - Success\n-- Performing Test CXX_HAS_Wunused_local_typedefs\n-- Performing Test CXX_HAS_Wunused_local_typedefs - Success\n-- Performing Test CXX_HAS_Wpotentially_evaluated_expression\n-- Performing Test CXX_HAS_Wpotentially_evaluated_expression - Success\n-- Setting default compiler flags and build type.\n-- *** I3_TESTDATA is not set. Using the default value of /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage/test-data\n--\n-- Configuring tools...\n--\n-- Using system packages.\n-- Using default site cmake dir of /usr/share/fizzicks/cmake\n--\n-- root\n-- + ROOT not found or disabled: building without ROOT support\n--\n-- python\n-- Found Python: /usr/local/opt/python/bin/python3.7 (found version \"3.7.6\") found components: Interpreter Development NumPy\n-- + version: 3.7.6\n-- + binary: /usr/local/opt/python/bin/python3.7\n-- + includes: /usr/local/Cellar/python/3.7.6_1/Frameworks/Python.framework/Versions/3.7/include/python3.7m\n-- + libs: /usr/local/Cellar/python/3.7.6_1/Frameworks/Python.framework/Versions/3.7/lib/libpython3.7m.dylib\n-- + numpy: /usr/local/lib/python3.7/site-packages/numpy/core/include\n-- + scipy: FOUND\n--\n-- Boost\n-- Looking for pthread.h\n-- Looking for pthread.h - found\n-- Performing Test CMAKE_HAVE_LIBC_PTHREAD\n-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success\n-- Found Threads: TRUE\n-- Found Boost: /usr/local/include (found version \"1.71.0\") found components: system thread date_time filesystem program_options regex iostreams python37 chrono atomic\n--\n-- boostnumpy\n-- - boost/numpy.hpp not found in include\n-- - boost_numpy\n--\n-- blas and lapack\n-- Looking for dgemm_\n-- Looking for dgemm_ - found\n-- Found BLAS: /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/System/Library/Frameworks/Accelerate.framework\n-- Looking for cheev_\n-- Looking for cheev_ - found\n-- A library with LAPACK API found.\n--\n-- gsl\n-- + gsl/gsl_rng.h found at /usr/local/include\n-- + /usr/local/lib/libgsl.dylib\n--\n-- sprng\n-- - sprng/sprng.h not found in include/sprng-2.0a\n-- - sprng\n--\n-- sprng\n-- - sprng/sprng.h not found in include\n-- - sprng\n--\n-- erfa\n-- - erfa.h not found in include\n-- - erfa\n--\n-- pal\n-- - star/pal.h not found in include\n-- - pal\n--\n-- pal\n-- - star/pal.h not found in include\n-- - starlink_pal\n--\n-- libarchive\n-- + /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libarchive.tbd\n-- Found MySQL: /usr/local/lib/libmysqlclient.dylib\n--\n-- suitesparse\n-- - cholmod.h not found in include\n-- - camd\n-- - ccolamd\n-- - spqr\n-- - cholmod\n-- - amd\n-- - colamd\n-- + /usr/local/lib/libtbb.dylib\n-- - suitesparseconfig\n--\n-- suitesparse\n-- - suitesparse/cholmod.h not found in include\n-- - camd\n-- - ccolamd\n-- - spqr\n-- - cholmod\n-- - amd\n-- - colamd\n-- - suitesparseconfig\n--\n-- suitesparse\n-- - suitesparse/cholmod.h not found in include\n-- - camd\n-- - ccolamd\n-- - spqr\n-- - cholmod\n-- - amd\n-- - colamd\n--\n-- ncurses\n-- Looking for wsyncup in /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libcurses.tbd\n-- Looking for wsyncup in /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libcurses.tbd - found\n-- Looking for cbreak in /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libncurses.tbd\n-- Looking for cbreak in /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libncurses.tbd - found\n-- Found Curses: /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libncurses.tbd\n-- + ncurses.h found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include\n-- + libncurses found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libncurses.tbd\n--\n-- cdk\n-- - cdk/cdk.h not found in include\n-- - cdk\n--\n-- cdk\n-- - cdk.h not found in include\n-- - cdk\n--\n-- cfitsio\n-- + fitsio.h found at /usr/local/include\n-- + /usr/local/lib/libcfitsio.dylib\n--\n-- healpix-cxx\n-- - healpix_cxx/healpix_map.h not found in include\n-- - healpix_cxx\n--\n-- hdf5\n-- - hdf5.h not found in include\n-- - hdf5\n-- - hdf5_hl\n--\n-- hdf5\n-- - hdf5.h not found in /usr/include/hdf5/serial\n-- - hdf5_serial\n-- - hdf5_serial_hl\n-- - Minuit2: USE_ROOT set to FALSE\n--\n-- clhep\n-- + CLHEP/ClhepVersion.h found at /usr/local/include\n-- + /usr/local/lib/libCLHEP.dylib\n-- Looking for Geant4 geant4-config program\n-- Looking for Geant4 geant4-config program -- not found\n-- Looking for Geant4 liblist program\n-- Looking for Geant4 liblist program -- not found\n--\n-- zlib\n-- + zlib.h found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include\n-- + /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libz.tbd\n--\n-- OpenCL\n-- + Using the OpenCL Framework because we're on Apple\n-- + cl.h found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/System/Library/Frameworks/OpenCL.framework/Headers\n-- + OpenCL framework found at -framework OpenCL\n-- Looking for CL_VERSION_2_0\n-- Looking for CL_VERSION_2_0 - not found\n-- Looking for CL_VERSION_1_2\n-- Looking for CL_VERSION_1_2 - found\n--\n-- gmp\n-- + gmp.h found at /usr/local\n-- + /usr/local/lib/libgmp.dylib\n--\n-- log4cpp\n-- + log4cpp/Category.hh found at /usr/local/include\n-- + /usr/local/lib/liblog4cpp.dylib\n--\n-- xml2\n-- + libxml/parser.h found at /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/include/libxml2\n-- + /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/usr/lib/libxml2.tbd\n--\n-- genie\n-- Looking for Genie genie-config program -- not found\n--\n-- zmq\n-- - zmq.hpp not found in ZMQ_INCLUDE_DIR-NOTFOUND\n-- + /usr/local/lib/libzmq.dylib\n--\n-- multinest\n-- - multinest.h not found in include\n-- - multinest\n--\n-- nlopt\n-- - nlopt.hpp not found in include\n-- - nlopt\n--\n-- zstd\n-- - zstd.h not found in ZSTD_INCLUDE_DIR-NOTFOUND\n-- - zstd\n--\n-- fftw3\n-- - fftw3.h not found in include\n-- - fftw3\n--\n-- Configuring projects:\n--\n-- + BadDomList\n-- +-- python [symlinks]\n-- + CascadeVariables\n-- +-- python [symlinks]\n-- + CoincSuite\n-- +-- python [symlinks]\n-- +-- CoincSuite-pybindings\n-- + CommonVariables\n-- +-- python [symlinks]\n-- +-- Include \"direct_hits\" sub project.\n-- +-- Include \"hit_multiplicity\" sub project.\n-- +-- Include \"hit_statistics\" sub project.\n-- +-- Include \"track_characteristics\" sub project.\n-- +-- Include \"time_characteristics\" sub project.\n-- +-- common_variables-pybindings\n-- +-- common_variables__hit_multiplicity-pybindings\n-- +-- common_variables__track_characteristics-pybindings\n-- +-- common_variables__hit_statistics-pybindings\n-- +-- common_variables__direct_hits-pybindings\n-- +-- common_variables__time_characteristics-pybindings\n-- + DOMLauncher\n-- +-- python [symlinks]\n-- +-- DOMLauncher-pybindings\n-- + DeepCore_Filter\n-- +-- python [symlinks]\n-- + DomTools\n-- +-- python [symlinks]\n-- + HiveSplitter\n-- +-- python [symlinks]\n-- + IceHive\n-- +-- python [symlinks]\n-- +-- IceHive-pybindings\n-- + KalmanFilter\n-- +-- python [symlinks]\n-- + MuonGun\n-- +-- python [symlinks]\n-- +-- Muonitron table-maker\n-- +-- MuonGun-pybindings\n-- + NoiseEngine\n-- Skipping NoiseEngine (healpix-cxx not found)\n-- + PROPOSAL\n-- +-- PROPOSAL-pybindings\n-- + SLOPtools\n-- +-- python [symlinks]\n-- + STTools\n-- +-- python [symlinks]\n-- +-- STTools-pybindings\n-- + StartingTrackVeto\n-- +-- python [symlinks]\n-- + TopologicalSplitter\n-- +-- python [symlinks]\n-- + VHESelfVeto\n-- +-- python [symlinks]\n-- +-- VHESelfVeto-pybindings\n-- + WaveCalibrator\n-- +-- python [symlinks]\n-- + astro\n-- - PAL not found, skipping build of icetray\n Ask your admin to install PAL. Instructions for PAL can be found at:\n http://software.icecube.wisc.edu/documentation/projects/cmake/tools/pal.html\n-- + bayesian-priors\n-- +-- python [symlinks]\n-- + clast\n-- +-- python [symlinks]\n-- + clsim\n-- +-- python [symlinks]\n-- --- ZMQ not found. I3CLSimMakePhotons will not function.\n-- +-- tabulator (have OpenCL 1.2)\n-- +-- safeprimes_base32.gz already downloaded\n-- +-- Geant4 or OpenCL is not installed on your system. clsim will fail if it is not used with parameterizations.\n-- +-- gmp support (make_safeprimes utility)\n-- +-- clsim-pybindings\n-- + cmake\n-- + cmc\n-- +-- python [symlinks]\n-- +-- cmc-pybindings\n-- + coinc-twc\n-- +-- python [symlinks]\n-- + core-removal\n-- +-- python [symlinks]\n-- + corsika-reader\n-- +-- python [symlinks]\n-- + Boost >= 6.4 : Adding CORSIKA Client/Server\n-- +-- corsika_reader-pybindings\n-- + cramer-rao\n-- +-- python [symlinks]\n-- + cscd-llh\n-- +-- python [symlinks]\n-- + daq-decode\n-- +-- python [symlinks]\n-- +-- daq_decode-pybindings\n-- + dataclasses\n-- +-- python [symlinks]\n-- +-- dataclasses-pybindings\n-- + dataio\n-- +-- python [symlinks]\n-- +-- dataio-pybindings\n-- +-- dataio-shovel *not* included (missing CDK or ncurses libraries)\n-- +-- test_unregistered-pybindings\n-- + ddddr\n-- +-- python [symlinks]\n-- +-- ddddr-pybindings\n-- + dipolefit\n-- +-- python [symlinks]\n-- + docs\n-- + double-muon\n-- +-- python [symlinks]\n-- + dst\n-- +-- Skipping dst (healpix-cxx not found)\n-- + earthmodel-service\n-- +-- earthmodel_service-pybindings\n-- + fill-ratio\n-- +-- python [symlinks]\n-- + filter-tools\n-- +-- python [symlinks]\n-- + filterscripts\n-- +-- python [symlinks]\n-- +-- astro required to build filterscripts-cxx\n-- + finiteReco\n-- +-- python [symlinks]\n-- + frame_object_diff\n-- +-- python [symlinks]\n-- +-- frame_object_diff-pybindings\n-- + full_event_followup\n-- +-- python [symlinks]\n-- + g4-tankresponse\n-- Looking for Geant4\n-- Looking for Geant4 -- not found\n-- Looking for CLHEP\n-- Looking for CLHEP -- found\n-- - Geant4 and/or CLHEP not found - Skipping build of g4-tankresponse\n-- +-- g4-tankresponse has not been thoroughly tested with Geant4 versions 4.10 and above\n-- + gcdserver\n-- +-- gcdserver does not work. Not building.\n-- + gulliver\n-- +-- python [symlinks]\n-- +-- gulliver-pybindings\n-- + gulliver-bootstrap\n-- +-- python [symlinks]\n-- +-- gulliver_bootstrap-pybindings\n-- + gulliver-modules\n-- +-- python [symlinks]\n-- + hdfwriter\n-- +-- python [symlinks]\n-- +-- hdf5 library not found, skipping build of hdfwriter\n-- *** Ask your admin to install the hdf5 package.\n-- + ice-models\n-- +-- python [symlinks]\n-- + icepick\n-- +-- python [symlinks]\n-- + icetop_Level3_scripts\n-- +-- python [symlinks]\n-- + icetray\n-- +-- libarchive *not* found, omitting optional tarfile support\n-- +-- libarchive *not* found, omitting optional tarfile support\n-- +-- python [symlinks]\n-- +-- memory tracking disabled\n-- +-- icetray-pybindings\n-- + interfaces\n-- +-- interfaces-pybindings\n-- + ipdf\n-- +-- python [symlinks]\n-- +-- ipdf-pybindings\n-- + level3-filter-cascade\n-- +-- python [symlinks]\n-- + level3-filter-muon\n-- +-- python [symlinks]\n-- +-- level3_filter_muon-pybindings\n-- + lilliput\n-- +-- python [symlinks]\n-- + ipdf support\n-- - Minuit2 support...not found\n-- - multinest support...not found\n-- - NLopt support...not found\n-- +-- lilliput-pybindings\n-- + linefit\n-- +-- python [symlinks]\n-- + millipede\n-- +-- SuiteSparse not found. Skipping millipede.\n If you need it you can install SuiteSparse through I3Ports.\n-- + monopole-generator\n-- +-- python [symlinks]\n-- + mue\n-- +-- python [symlinks]\n-- + neutrino-generator\n-- +-- python [symlinks]\n-- +-- /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage/neutrino-generator/cross_section_data does not exist. Downloading...\n-- +-- Downloading and unpacking the crosssection data files...\ntrying to download /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage/neutrino-generator/nugen-v3.1-tables.tgz...\n downloading http://code.icecube.wisc.edu/tools/neutrino-generator/nugen-v3.1-tables.tgz...\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 44.6M 100 44.6M 0 0 3572k 0 0:00:12 0:00:12 --:--:-- 8988k\n downloading http://code.icecube.wisc.edu/tools/neutrino-generator/nugen-v3.1-tables.md5sum...\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n100 56 100 56 0 0 265 0 --:--:-- --:--:-- --:--:-- 265\n check md5sum ...\n extracting http://code.icecube.wisc.edu/tools/neutrino-generator/nugen-v3.1-tables.tgz ...\n /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage/neutrino-generator/nugen-v3.1-tables.tgz is downloaded.\n-- +-- ...download complete.\n-- +-- genie-icetray NOT found. building without GENIE support.\n-- +-- neutrino_generator-pybindings\n-- + ophelia\n-- +-- python [symlinks]\n-- + paraboloid\n-- +-- python [symlinks]\n-- +-- paraboloid-pybindings\n-- + payload-parsing\n-- +-- python [symlinks]\n-- +-- payload_parsing-pybindings\n-- + photonics-service\n-- +-- python [symlinks]\n-- +-- photospline support\n-- +-- numpy support\n-- +-- photonics_service-pybindings\n-- + photospline\n-- +-- python [symlinks]\n-- +-- photospline-pybindings\n-- + phys-services\n-- +-- python [symlinks]\n-- +-- sprng *not* found, omitting optional SPRNGRandomService\n-- +-- phys_services-pybindings\n-- + polyplopia\n-- +-- python [symlinks]\n-- +-- polyplopia-pybindings\n-- + portia\n-- +-- python [symlinks]\n-- + ppc\n-- +-- python [symlinks]\n-- +-- OpenCL SDK found. Building ppc.\n-- + production-histograms\n-- +-- python [symlinks]\n-- + pybdt\n-- + recclasses\n-- +-- python [symlinks]\n-- +-- recclasses-pybindings\n-- + rootwriter\n-- +-- python [symlinks]\n-- +-- ROOT has been disabled in your configuration.\n-- +-- Skipping build of rootwriter\n-- + serialization\n-- + shield\n-- +-- python [symlinks]\n-- + sim-services\n-- +-- python [symlinks]\n-- +-- sim_services-pybindings\n-- + simclasses\n-- +-- python [symlinks]\n-- + tableio support\n-- +-- simclasses-pybindings\n-- + simprod-scripts\n-- +-- python [symlinks]\n-- + smallshower-filter\n-- + spline-reco\n-- +-- python [symlinks]\n-- +-- spline_reco-pybindings\n-- + static-twc\n-- +-- python [symlinks]\n-- + steamshovel\n-- +-- python [symlinks]\nCMake Deprecation Warning at steamshovel/CMakeLists.txt:13 (cmake_policy):\n The OLD behavior for policy CMP0053 will be removed from a future version\n of CMake.\n\n The cmake-policies(7) manual explains that the OLD behaviors of all\n policies are deprecated and that a policy should be set to OLD only under\n specific short-term circumstances. Projects should be ported to the NEW\n behavior and not rely on setting a policy to OLD.\n\n\n-- Found OpenGL: /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/System/Library/Frameworks/OpenGL.framework\n-- Found GLUT: /Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.15.sdk/System/Library/Frameworks/GLUT.framework\n-- +-- shovelart-pybindings\n-- +-- shovelio-pybindings\n-- + stochastics\n-- +-- python [symlinks]\n-- +-- stochastics-pybindings\n-- + tableio\n-- +-- python [symlinks]\n-- +-- tableio-pybindings\n-- + tensor-of-inertia\n-- +-- python [symlinks]\n-- + topeventcleaning\n-- +-- python [symlinks]\n-- + toprec\n-- +-- python [symlinks]\n-- +-- toprec-pybindings\n-- + topsimulator\n-- +-- python [symlinks]\n-- +-- topsimulator-pybindings\n-- + tpx\n-- +-- python [symlinks]\n-- +-- tpx-pybindings\n-- +-- FETCH_TPX set to OFF. SLC calibration data may not be available.\n-- + trigger-sim\n-- +-- python [symlinks]\n-- +-- trigger_sim-pybindings\n-- + trigger-splitter\n-- +-- python [symlinks]\n-- + truncated_energy\n-- +-- python [symlinks]\n-- + vemcal\n-- +-- python [symlinks]\n-- +-- vemcal-pybindings\n-- + vuvuzela\n-- +-- python [symlinks]\n-- +-- vuvuzela-pybindings\n-- + wavedeform\n-- +-- SuiteSparse not found. Skipping wavedeform.\n If you need it you can install SuiteSparse through I3Ports.\n-- + wavereform\n-- +-- python [symlinks]\n-- +-- wavereform-pybindings\n-- + weighting\n-- +-- python [symlinks]\n-- + wimpsim-reader\n-- +-- python [symlinks]\n-- +-- astro not built. Skipping test 'test_sun_position'\n-- Generating env-shell.sh\n-- Generating icetray-config\n-- Configuring 'gfilt' STL decryptor\n-- Configuring done\n-- Generating done\n-- Build files have been written to: /Users/fiedl/icecube/icecube-combo-build/fiedl/monopole-generator/coverage\n}}}",
"reporter": "fiedl",
"cc": "",
"resolution": "fixed",
"_ts": "1593001720016649",
"component": "combo core",
"summary": "Coverage: no gcda files are found",
"priority": "normal",
"keywords": "coverage, cmake",
"time": "2020-02-28T00:01:26",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| defect | coverage no gcda files are found trac migrated from json status closed changetime description i tried to generate a coverage report but gcda files are found n nsee fiedl fiedl mbp icecube combo fiedl monopole generator n brew install lcov n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator n mkdir coverage n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator coverage n cmake d cmake build type coverage d system packages true d python root usr local opt python icecube combo n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator coverage n env shell sh make monopole generator test j n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator coverage n lcov d z ndeleting all da files in and subdirectories ndone n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator coverage n lcov d c i o test base info ncapturing coverage data from nfound llvm gcov version which emulates gcov version nscanning for gcno files ngeninfo warning no gcno files found in skipping nfinished info file creation n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator coverage n env shell sh ctest r monopole generator verbose n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator coverage n lcov d c o test run info ncapturing coverage data from nfound llvm gcov version which emulates gcov version nscanning for gcda files ngeninfo warning no gcda files found in skipping nfinished info file creation n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator coverage n cmake version ncmake version n n ni ve also tried the full make coverage same result n nentire cmake output n n n fiedl fiedl mbp icecube icecube combo build fiedl monopole generator coverage n cmake d cmake build type coverage d system packages true d python root usr local opt python icecube combo n the c compiler identification is appleclang n the cxx compiler identification is appleclang n check for working c compiler applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin cc n check for working c compiler applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin cc works n detecting c compiler abi info n detecting c compiler abi info done n detecting c compile features n detecting c compile features done n check for working cxx compiler applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin c n check for working cxx compiler applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin c works n detecting cxx compiler abi info n detecting cxx compiler abi info done n detecting cxx compile features n detecting cxx compile features done ncmake deprecation warning at cmakelists txt cmake policy n the old behavior for policy will be removed from a future version n of cmake n n the cmake policies manual explains that the old behaviors of all n policies are deprecated and that a policy should be set to old only under n specific short term circumstances projects should be ported to the new n behavior and not rely on setting a policy to old n n n n icecube configuration starting n nunable to determine upstream svn information from working tree history n n ostype darwin n osversion n arch n buildname darwin clang n toolset clang coverage n hostname fiedl mbp local n cmake path usr local cellar cmake bin cmake n cmake version n svn revision n svn url unknown n meta project unknown n n setting compiler compile drivers and linker n n distcc found at usr local bin distcc n ccache found at usr local bin ccache n performing test cxx has wdeprecated n performing test cxx has wdeprecated success n performing test cxx has wunused variable n performing test cxx has wunused variable success n performing test cxx has wunused local typedef n performing test cxx has wunused local typedef success n performing test cxx has wunused local typedefs n performing test cxx has wunused local typedefs success n performing test cxx has wpotentially evaluated expression n performing test cxx has wpotentially evaluated expression success n setting default compiler flags and build type n testdata is not set using the default value of users fiedl icecube icecube combo build fiedl monopole generator coverage test data n n configuring tools n n using system packages n using default site cmake dir of usr share fizzicks cmake n n root n root not found or disabled building without root support n n python n found python usr local opt python bin found version found components interpreter development numpy n version n binary usr local opt python bin n includes usr local cellar python frameworks python framework versions include n libs usr local cellar python frameworks python framework versions lib dylib n numpy usr local lib site packages numpy core include n scipy found n n boost n looking for pthread h n looking for pthread h found n performing test cmake have libc pthread n performing test cmake have libc pthread success n found threads true n found boost usr local include found version found components system thread date time filesystem program options regex iostreams chrono atomic n n boostnumpy n boost numpy hpp not found in include n boost numpy n n blas and lapack n looking for dgemm n looking for dgemm found n found blas applications xcode app contents developer platforms macosx platform developer sdks sdk system library frameworks accelerate framework n looking for cheev n looking for cheev found n a library with lapack api found n n gsl n gsl gsl rng h found at usr local include n usr local lib libgsl dylib n n sprng n sprng sprng h not found in include sprng n sprng n n sprng n sprng sprng h not found in include n sprng n n erfa n erfa h not found in include n erfa n n pal n star pal h not found in include n pal n n pal n star pal h not found in include n starlink pal n n libarchive n applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib libarchive tbd n found mysql usr local lib libmysqlclient dylib n n suitesparse n cholmod h not found in include n camd n ccolamd n spqr n cholmod n amd n colamd n usr local lib libtbb dylib n suitesparseconfig n n suitesparse n suitesparse cholmod h not found in include n camd n ccolamd n spqr n cholmod n amd n colamd n suitesparseconfig n n suitesparse n suitesparse cholmod h not found in include n camd n ccolamd n spqr n cholmod n amd n colamd n n ncurses n looking for wsyncup in applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib libcurses tbd n looking for wsyncup in applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib libcurses tbd found n looking for cbreak in applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib libncurses tbd n looking for cbreak in applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib libncurses tbd found n found curses applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib libncurses tbd n ncurses h found at applications xcode app contents developer platforms macosx platform developer sdks sdk usr include n libncurses found at applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib libncurses tbd n n cdk n cdk cdk h not found in include n cdk n n cdk n cdk h not found in include n cdk n n cfitsio n fitsio h found at usr local include n usr local lib libcfitsio dylib n n healpix cxx n healpix cxx healpix map h not found in include n healpix cxx n n n h not found in include n n hl n n n h not found in usr include serial n serial n serial hl n use root set to false n n clhep n clhep clhepversion h found at usr local include n usr local lib libclhep dylib n looking for config program n looking for config program not found n looking for liblist program n looking for liblist program not found n n zlib n zlib h found at applications xcode app contents developer platforms macosx platform developer sdks sdk usr include n applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib libz tbd n n opencl n using the opencl framework because we re on apple n cl h found at applications xcode app contents developer platforms macosx platform developer sdks sdk system library frameworks opencl framework headers n opencl framework found at framework opencl n looking for cl version n looking for cl version not found n looking for cl version n looking for cl version found n n gmp n gmp h found at usr local n usr local lib libgmp dylib n n n category hh found at usr local include n usr local lib dylib n n n libxml parser h found at applications xcode app contents developer platforms macosx platform developer sdks sdk usr include n applications xcode app contents developer platforms macosx platform developer sdks sdk usr lib tbd n n genie n looking for genie genie config program not found n n zmq n zmq hpp not found in zmq include dir notfound n usr local lib libzmq dylib n n multinest n multinest h not found in include n multinest n n nlopt n nlopt hpp not found in include n nlopt n n zstd n zstd h not found in zstd include dir notfound n zstd n n n h not found in include n n n configuring projects n n baddomlist n python n cascadevariables n python n coincsuite n python n coincsuite pybindings n commonvariables n python n include direct hits sub project n include hit multiplicity sub project n include hit statistics sub project n include track characteristics sub project n include time characteristics sub project n common variables pybindings n common variables hit multiplicity pybindings n common variables track characteristics pybindings n common variables hit statistics pybindings n common variables direct hits pybindings n common variables time characteristics pybindings n domlauncher n python n domlauncher pybindings n deepcore filter n python n domtools n python n hivesplitter n python n icehive n python n icehive pybindings n kalmanfilter n python n muongun n python n muonitron table maker n muongun pybindings n noiseengine n skipping noiseengine healpix cxx not found n proposal n proposal pybindings n sloptools n python n sttools n python n sttools pybindings n startingtrackveto n python n topologicalsplitter n python n vheselfveto n python n vheselfveto pybindings n wavecalibrator n python n astro n pal not found skipping build of icetray n ask your admin to install pal instructions for pal can be found at n bayesian priors n python n clast n python n clsim n python n zmq not found will not function n tabulator have opencl n safeprimes gz already downloaded n or opencl is not installed on your system clsim will fail if it is not used with parameterizations n gmp support make safeprimes utility n clsim pybindings n cmake n cmc n python n cmc pybindings n coinc twc n python n core removal n python n corsika reader n python n boost adding corsika client server n corsika reader pybindings n cramer rao n python n cscd llh n python n daq decode n python n daq decode pybindings n dataclasses n python n dataclasses pybindings n dataio n python n dataio pybindings n dataio shovel not included missing cdk or ncurses libraries n test unregistered pybindings n ddddr n python n ddddr pybindings n dipolefit n python n docs n double muon n python n dst n skipping dst healpix cxx not found n earthmodel service n earthmodel service pybindings n fill ratio n python n filter tools n python n filterscripts n python n astro required to build filterscripts cxx n finitereco n python n frame object diff n python n frame object diff pybindings n full event followup n python n tankresponse n looking for n looking for not found n looking for clhep n looking for clhep found n and or clhep not found skipping build of tankresponse n tankresponse has not been thoroughly tested with versions and above n gcdserver n gcdserver does not work not building n gulliver n python n gulliver pybindings n gulliver bootstrap n python n gulliver bootstrap pybindings n gulliver modules n python n hdfwriter n python n library not found skipping build of hdfwriter n ask your admin to install the package n ice models n python n icepick n python n icetop scripts n python n icetray n libarchive not found omitting optional tarfile support n libarchive not found omitting optional tarfile support n python n memory tracking disabled n icetray pybindings n interfaces n interfaces pybindings n ipdf n python n ipdf pybindings n filter cascade n python n filter muon n python n filter muon pybindings n lilliput n python n ipdf support n support not found n multinest support not found n nlopt support not found n lilliput pybindings n linefit n python n millipede n suitesparse not found skipping millipede n if you need it you can install suitesparse through n monopole generator n python n mue n python n neutrino generator n python n users fiedl icecube icecube combo build fiedl monopole generator coverage neutrino generator cross section data does not exist downloading n downloading and unpacking the crosssection data files ntrying to download users fiedl icecube icecube combo build fiedl monopole generator coverage neutrino generator nugen tables tgz n downloading total received xferd average speed time time time current n dload upload total spent left speed n downloading total received xferd average speed time time time current n dload upload total spent left speed n check n extracting n users fiedl icecube icecube combo build fiedl monopole generator coverage neutrino generator nugen tables tgz is downloaded n download complete n genie icetray not found building without genie support n neutrino generator pybindings n ophelia n python n paraboloid n python n paraboloid pybindings n payload parsing n python n payload parsing pybindings n photonics service n python n photospline support n numpy support n photonics service pybindings n photospline n python n photospline pybindings n phys services n python n sprng not found omitting optional sprngrandomservice n phys services pybindings n polyplopia n python n polyplopia pybindings n portia n python n ppc n python n opencl sdk found building ppc n production histograms n python n pybdt n recclasses n python n recclasses pybindings n rootwriter n python n root has been disabled in your configuration n skipping build of rootwriter n serialization n shield n python n sim services n python n sim services pybindings n simclasses n python n tableio support n simclasses pybindings n simprod scripts n python n smallshower filter n spline reco n python n spline reco pybindings n static twc n python n steamshovel n python ncmake deprecation warning at steamshovel cmakelists txt cmake policy n the old behavior for policy will be removed from a future version n of cmake n n the cmake policies manual explains that the old behaviors of all n policies are deprecated and that a policy should be set to old only under n specific short term circumstances projects should be ported to the new n behavior and not rely on setting a policy to old n n n found opengl applications xcode app contents developer platforms macosx platform developer sdks sdk system library frameworks opengl framework n found glut applications xcode app contents developer platforms macosx platform developer sdks sdk system library frameworks glut framework n shovelart pybindings n shovelio pybindings n stochastics n python n stochastics pybindings n tableio n python n tableio pybindings n tensor of inertia n python n topeventcleaning n python n toprec n python n toprec pybindings n topsimulator n python n topsimulator pybindings n tpx n python n tpx pybindings n fetch tpx set to off slc calibration data may not be available n trigger sim n python n trigger sim pybindings n trigger splitter n python n truncated energy n python n vemcal n python n vemcal pybindings n vuvuzela n python n vuvuzela pybindings n wavedeform n suitesparse not found skipping wavedeform n if you need it you can install suitesparse through n wavereform n python n wavereform pybindings n weighting n python n wimpsim reader n python n astro not built skipping test test sun position n generating env shell sh n generating icetray config n configuring gfilt stl decryptor n configuring done n generating done n build files have been written to users fiedl icecube icecube combo build fiedl monopole generator coverage n reporter fiedl cc resolution fixed ts component combo core summary coverage no gcda files are found priority normal keywords coverage cmake time milestone owner type defect | 1 |
633,867 | 20,268,806,860 | IssuesEvent | 2022-02-15 14:31:29 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | Can't create security event suricata trigger with custom value | Type: Bug Priority: High | **Describe the bug**
Can't create security event suricata trigger with custom value which used to be possible
**To Reproduce**
1. Create a security event
2. Add a trigger, 'Event->Suricata Event' with value `ET POLICY iTunes User Agent`
3. You can't set this value since its not in the select options
**Expected behavior**
Should be able to set it to `ET POLICY iTunes User Agent`
| 1.0 | Can't create security event suricata trigger with custom value - **Describe the bug**
Can't create security event suricata trigger with custom value which used to be possible
**To Reproduce**
1. Create a security event
2. Add a trigger, 'Event->Suricata Event' with value `ET POLICY iTunes User Agent`
3. You can't set this value since its not in the select options
**Expected behavior**
Should be able to set it to `ET POLICY iTunes User Agent`
| non_defect | can t create security event suricata trigger with custom value describe the bug can t create security event suricata trigger with custom value which used to be possible to reproduce create a security event add a trigger event suricata event with value et policy itunes user agent you can t set this value since its not in the select options expected behavior should be able to set it to et policy itunes user agent | 0 |
250,086 | 18,873,214,399 | IssuesEvent | 2021-11-13 15:10:43 | kdblocher/bridge | https://api.github.com/repos/kdblocher/bridge | opened | Create User Stories for example usages to aid in GUI development and documentation creation | documentation | Create official user stories that can be agreed upon so that future GUI development and documentation creation will have more direction. These listed examples are just starter ideas, but the user stories will need to be carefully developed so that these activities can be explained and done via the GUI in a straightforward manner.
Case 1: First Demo. Define 1NT as 15-17 BAL, and then 2N as an 8 count with either 4333, 5 card minor, and 5 card minor with the ten. Look at number of tricks makeable in each case.
Case 2: Have a 1NT opener with a transfer response. Opener has 4333 with 3 cards in the transferred major. Responder invites with 2NT. Run stats to determine whether it is better to stay in NT rather than retreat to 3 of the major.
Case 3: Develop a system around a weak 2 bid. Examine it's effectiveness. Then alter the weak two to exclude honor 3rd in the other major.
Case 4: Conventional comparison. Compare Bergen raises to 3 of a minor as invitational with 6 or more of the minor.
Case 5: Develop a bidding tree to represent part of a full system. Walk through a mechanism in the GUI to examine set hands for a give system so that a user can identify patterns of poor bidding judgement. | 1.0 | Create User Stories for example usages to aid in GUI development and documentation creation - Create official user stories that can be agreed upon so that future GUI development and documentation creation will have more direction. These listed examples are just starter ideas, but the user stories will need to be carefully developed so that these activities can be explained and done via the GUI in a straightforward manner.
Case 1: First Demo. Define 1NT as 15-17 BAL, and then 2N as an 8 count with either 4333, 5 card minor, and 5 card minor with the ten. Look at number of tricks makeable in each case.
Case 2: Have a 1NT opener with a transfer response. Opener has 4333 with 3 cards in the transferred major. Responder invites with 2NT. Run stats to determine whether it is better to stay in NT rather than retreat to 3 of the major.
Case 3: Develop a system around a weak 2 bid. Examine it's effectiveness. Then alter the weak two to exclude honor 3rd in the other major.
Case 4: Conventional comparison. Compare Bergen raises to 3 of a minor as invitational with 6 or more of the minor.
Case 5: Develop a bidding tree to represent part of a full system. Walk through a mechanism in the GUI to examine set hands for a give system so that a user can identify patterns of poor bidding judgement. | non_defect | create user stories for example usages to aid in gui development and documentation creation create official user stories that can be agreed upon so that future gui development and documentation creation will have more direction these listed examples are just starter ideas but the user stories will need to be carefully developed so that these activities can be explained and done via the gui in a straightforward manner case first demo define as bal and then as an count with either card minor and card minor with the ten look at number of tricks makeable in each case case have a opener with a transfer response opener has with cards in the transferred major responder invites with run stats to determine whether it is better to stay in nt rather than retreat to of the major case develop a system around a weak bid examine it s effectiveness then alter the weak two to exclude honor in the other major case conventional comparison compare bergen raises to of a minor as invitational with or more of the minor case develop a bidding tree to represent part of a full system walk through a mechanism in the gui to examine set hands for a give system so that a user can identify patterns of poor bidding judgement | 0 |
9,700 | 2,615,166,062 | IssuesEvent | 2015-03-01 06:46:28 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | reaver is support dedicatory attack wps pin? | auto-migrated Priority-Triage Type-Defect | ```
answer plz.........
```
Original issue reported on code.google.com by `jainilpa...@gmail.com` on 20 May 2013 at 1:37 | 1.0 | reaver is support dedicatory attack wps pin? - ```
answer plz.........
```
Original issue reported on code.google.com by `jainilpa...@gmail.com` on 20 May 2013 at 1:37 | defect | reaver is support dedicatory attack wps pin answer plz original issue reported on code google com by jainilpa gmail com on may at | 1 |
528,347 | 15,364,791,348 | IssuesEvent | 2021-03-01 22:30:17 | code4lib/2021.code4lib.org | https://api.github.com/repos/code4lib/2021.code4lib.org | opened | Non-visible links being captured during tab navigation | Priority: Medium | - On speakers page, the link inside the speaker-info box captures the tab whether the speaker is active or not. Needs JS to set attribute for link to `tabindex="-1"` unless the speaker box is .selected, then flip to `tabindex="0"`
| 1.0 | Non-visible links being captured during tab navigation - - On speakers page, the link inside the speaker-info box captures the tab whether the speaker is active or not. Needs JS to set attribute for link to `tabindex="-1"` unless the speaker box is .selected, then flip to `tabindex="0"`
| non_defect | non visible links being captured during tab navigation on speakers page the link inside the speaker info box captures the tab whether the speaker is active or not needs js to set attribute for link to tabindex unless the speaker box is selected then flip to tabindex | 0 |
20,863 | 3,422,360,082 | IssuesEvent | 2015-12-08 22:40:38 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | malloc error in unit test | area-vm priority-unassigned triaged Type-Defect | *This issue was originally filed by perrys...@gmail.com*
_____
**What steps will reproduce the problem?**
Not exactly sure what in here is causing the error. I tried to break it down further but even taking out the '_two' seemed to resolve the malloc issue so I just extracted and reproduced the offending pattern with simple classes and unit test.
main() {
test('temp', () {
Thing.wrapper()();
});
}
class Foo {
final bool \_one;
final int \_two;
const Foo._(this._one, this._two);
static final dynamic ctor = () {
return new Foo._(true, 10);
};
}
class Thing {
static final dynamic wrapper = () {
return () {
other(null, values: null);
};
};
static final dynamic other = (Action foo, [Foo values]) {
if(values == null) { values = Foo.ctor(); }
};
}
**What is the expected output? What do you see instead?**
Thinking that a malloc isn't the level of error I should be seeing in the editor :)
unittest-suite-wait-for-done
dart(6922,0xb0185000) malloc: \*\*\* error for object 0xb0183414: pointer being freed was not allocated
\*\*\* set a breakpoint in malloc_error_break to debug
**What version of the product are you using? On what operating system?**
Mac OSX Mountain Lion
Dart Editor version 0.4.3_r20602
Dart SDK version 0.4.3.5_r20602
**Please provide any additional information below.** | 1.0 | malloc error in unit test - *This issue was originally filed by perrys...@gmail.com*
_____
**What steps will reproduce the problem?**
Not exactly sure what in here is causing the error. I tried to break it down further but even taking out the '_two' seemed to resolve the malloc issue so I just extracted and reproduced the offending pattern with simple classes and unit test.
main() {
test('temp', () {
Thing.wrapper()();
});
}
class Foo {
final bool \_one;
final int \_two;
const Foo._(this._one, this._two);
static final dynamic ctor = () {
return new Foo._(true, 10);
};
}
class Thing {
static final dynamic wrapper = () {
return () {
other(null, values: null);
};
};
static final dynamic other = (Action foo, [Foo values]) {
if(values == null) { values = Foo.ctor(); }
};
}
**What is the expected output? What do you see instead?**
Thinking that a malloc isn't the level of error I should be seeing in the editor :)
unittest-suite-wait-for-done
dart(6922,0xb0185000) malloc: \*\*\* error for object 0xb0183414: pointer being freed was not allocated
\*\*\* set a breakpoint in malloc_error_break to debug
**What version of the product are you using? On what operating system?**
Mac OSX Mountain Lion
Dart Editor version 0.4.3_r20602
Dart SDK version 0.4.3.5_r20602
**Please provide any additional information below.** | defect | malloc error in unit test this issue was originally filed by perrys gmail com what steps will reproduce the problem not exactly sure what in here is causing the error i tried to break it down further but even taking out the two seemed to resolve the malloc issue so i just extracted and reproduced the offending pattern with simple classes and unit test main nbsp nbsp test temp nbsp nbsp nbsp nbsp thing wrapper nbsp nbsp class foo nbsp nbsp final bool one nbsp nbsp final int two nbsp nbsp const foo this one this two nbsp nbsp static final dynamic ctor nbsp nbsp nbsp nbsp return new foo true nbsp nbsp class thing nbsp nbsp static final dynamic wrapper nbsp nbsp nbsp nbsp return nbsp nbsp nbsp nbsp nbsp nbsp other null values null nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp static final dynamic other action foo nbsp nbsp nbsp nbsp if values null values foo ctor nbsp nbsp what is the expected output what do you see instead thinking that a malloc isn t the level of error i should be seeing in the editor unittest suite wait for done dart malloc error for object pointer being freed was not allocated set a breakpoint in malloc error break to debug what version of the product are you using on what operating system mac osx mountain lion dart editor version dart sdk version please provide any additional information below | 1 |
21,958 | 3,587,215,275 | IssuesEvent | 2016-01-30 05:06:15 | mash99/crypto-js | https://api.github.com/repos/mash99/crypto-js | closed | Files with certain sizes cause 'Uncaught RangeError: Maximum call stack size exceeded ' with SHA1 | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Run this: dd if=/dev/urandom of=2test.raw bs=$((1024 * (124) + 128 + 57 ))
count=1
2. Load file as ArrayBuffer (in a web worker context; might not be important)
3. Pass to CryptoJS.SHA1() wrapped in CryptoJS.lib.WordArray.create()
4. Print resulting SHA1 with toString()
What is the expected output? What do you see instead?
SHA1 of input file. Instead I see:
'Uncaught RangeError: Maximum call stack size exceeded'
Note: adding or subtracting a few bytes from the random file makes SHA1 work...
What version of the product are you using? On what operating system?
3.1.2, Ubuntu 12.04 LTS 64 Bit, Chrome 24.0.1312.52
Please provide any additional information below.
```
Original issue reported on code.google.com by `wolfgang...@gmail.com` on 23 Jan 2013 at 11:26 | 1.0 | Files with certain sizes cause 'Uncaught RangeError: Maximum call stack size exceeded ' with SHA1 - ```
What steps will reproduce the problem?
1. Run this: dd if=/dev/urandom of=2test.raw bs=$((1024 * (124) + 128 + 57 ))
count=1
2. Load file as ArrayBuffer (in a web worker context; might not be important)
3. Pass to CryptoJS.SHA1() wrapped in CryptoJS.lib.WordArray.create()
4. Print resulting SHA1 with toString()
What is the expected output? What do you see instead?
SHA1 of input file. Instead I see:
'Uncaught RangeError: Maximum call stack size exceeded'
Note: adding or subtracting a few bytes from the random file makes SHA1 work...
What version of the product are you using? On what operating system?
3.1.2, Ubuntu 12.04 LTS 64 Bit, Chrome 24.0.1312.52
Please provide any additional information below.
```
Original issue reported on code.google.com by `wolfgang...@gmail.com` on 23 Jan 2013 at 11:26 | defect | files with certain sizes cause uncaught rangeerror maximum call stack size exceeded with what steps will reproduce the problem run this dd if dev urandom of raw bs count load file as arraybuffer in a web worker context might not be important pass to cryptojs wrapped in cryptojs lib wordarray create print resulting with tostring what is the expected output what do you see instead of input file instead i see uncaught rangeerror maximum call stack size exceeded note adding or subtracting a few bytes from the random file makes work what version of the product are you using on what operating system ubuntu lts bit chrome please provide any additional information below original issue reported on code google com by wolfgang gmail com on jan at | 1 |
7,290 | 3,082,724,389 | IssuesEvent | 2015-08-24 00:43:53 | california-civic-data-coalition/django-calaccess-raw-data | https://api.github.com/repos/california-civic-data-coalition/django-calaccess-raw-data | opened | Add documentation for the ``d_le_nams`` field on the ``LobbyAmendmentsCd`` database model | documentation enhancement small |
## Your mission
Add documentation for the ``d_le_nams`` field on the ``LobbyAmendmentsCd`` database model.
## Here's how
**Step 1**: Claim this ticket by leaving a comment below. Tell everyone you're ON IT!
**Step 2**: Open up the file that contains this model. It should be in <a href="https://github.com/california-civic-data-coalition/django-calaccess-raw-data/blob/master/calaccess_raw/models/lobbying.py">calaccess_raw.models.lobbying.py</a>.
**Step 3**: Hit the little pencil button in the upper-right corner of the code box to begin editing the file.

**Step 4**: Find this model and field in the file. (Clicking into the box and searching with CTRL-F can help you here.) Once you find it, we expect the field to lack the ``help_text`` field typically used in Django to explain what a field contains.
```python
effect_dt = fields.DateField(
null=True,
db_column="EFFECT_DT"
)
```
**Step 5**: In a separate tab, open up the <a href="Quilmes">official state documentation</a> and find the page that defines all the fields in this model.

**Step 6**: Find the row in that table's definition table that spells out what this field contains. If it lacks documentation. Note that in the ticket and close it now.

**Step 7**: Return to the GitHub tab.
**Step 8**: Add the state's label explaining what's in the field, to our field definition by inserting it a ``help_text`` argument. That should look something like this:
```python
effect_dt = fields.DateField(
null=True,
db_column="EFFECT_DT",
# Add a help_text argument like the one here, but put your string in instead.
help_text="The other values in record were effective as of this date"
)
```
**Step 9**: Scroll down below the code box and describe the change you've made in the commit message. Press the button below.

**Step 10**: Review your changes and create a pull request submitting them to the core team for inclusion.

That's it! Mission accomplished!
| 1.0 | Add documentation for the ``d_le_nams`` field on the ``LobbyAmendmentsCd`` database model -
## Your mission
Add documentation for the ``d_le_nams`` field on the ``LobbyAmendmentsCd`` database model.
## Here's how
**Step 1**: Claim this ticket by leaving a comment below. Tell everyone you're ON IT!
**Step 2**: Open up the file that contains this model. It should be in <a href="https://github.com/california-civic-data-coalition/django-calaccess-raw-data/blob/master/calaccess_raw/models/lobbying.py">calaccess_raw.models.lobbying.py</a>.
**Step 3**: Hit the little pencil button in the upper-right corner of the code box to begin editing the file.

**Step 4**: Find this model and field in the file. (Clicking into the box and searching with CTRL-F can help you here.) Once you find it, we expect the field to lack the ``help_text`` field typically used in Django to explain what a field contains.
```python
effect_dt = fields.DateField(
null=True,
db_column="EFFECT_DT"
)
```
**Step 5**: In a separate tab, open up the <a href="Quilmes">official state documentation</a> and find the page that defines all the fields in this model.

**Step 6**: Find the row in that table's definition table that spells out what this field contains. If it lacks documentation. Note that in the ticket and close it now.

**Step 7**: Return to the GitHub tab.
**Step 8**: Add the state's label explaining what's in the field, to our field definition by inserting it a ``help_text`` argument. That should look something like this:
```python
effect_dt = fields.DateField(
null=True,
db_column="EFFECT_DT",
# Add a help_text argument like the one here, but put your string in instead.
help_text="The other values in record were effective as of this date"
)
```
**Step 9**: Scroll down below the code box and describe the change you've made in the commit message. Press the button below.

**Step 10**: Review your changes and create a pull request submitting them to the core team for inclusion.

That's it! Mission accomplished!
| non_defect | add documentation for the d le nams field on the lobbyamendmentscd database model your mission add documentation for the d le nams field on the lobbyamendmentscd database model here s how step claim this ticket by leaving a comment below tell everyone you re on it step open up the file that contains this model it should be in a href step hit the little pencil button in the upper right corner of the code box to begin editing the file step find this model and field in the file clicking into the box and searching with ctrl f can help you here once you find it we expect the field to lack the help text field typically used in django to explain what a field contains python effect dt fields datefield null true db column effect dt step in a separate tab open up the official state documentation and find the page that defines all the fields in this model step find the row in that table s definition table that spells out what this field contains if it lacks documentation note that in the ticket and close it now step return to the github tab step add the state s label explaining what s in the field to our field definition by inserting it a help text argument that should look something like this python effect dt fields datefield null true db column effect dt add a help text argument like the one here but put your string in instead help text the other values in record were effective as of this date step scroll down below the code box and describe the change you ve made in the commit message press the button below step review your changes and create a pull request submitting them to the core team for inclusion that s it mission accomplished | 0 |
272,073 | 23,652,466,348 | IssuesEvent | 2022-08-26 08:05:01 | sourcegraph/sourcegraph | https://api.github.com/repos/sourcegraph/sourcegraph | opened | ci/flake: nightly VSCE integration tests disabled | testing dx ci/flake | ### Test
Nightly VSCE integration tests
### Example failure
https://buildkite.com/sourcegraph/sourcegraph/builds/169322
### Disabling PR
https://github.com/sourcegraph/sourcegraph/pull/40890
### Additional details
_No response_ | 1.0 | ci/flake: nightly VSCE integration tests disabled - ### Test
Nightly VSCE integration tests
### Example failure
https://buildkite.com/sourcegraph/sourcegraph/builds/169322
### Disabling PR
https://github.com/sourcegraph/sourcegraph/pull/40890
### Additional details
_No response_ | non_defect | ci flake nightly vsce integration tests disabled test nightly vsce integration tests example failure disabling pr additional details no response | 0 |
373,155 | 26,039,634,127 | IssuesEvent | 2022-12-22 09:16:36 | matplotlib/matplotlib | https://api.github.com/repos/matplotlib/matplotlib | closed | [Doc]: Development workflow doc has lots of typos and clunky sentences | Documentation | ### Documentation Link
https://matplotlib.org/devdocs/devel/development_workflow.html
### Problem
There are typos and/or grammer errors in the following lines on this page at https://github.com/matplotlib/matplotlib/blob/main/doc/devel/development_workflow.rst:
```
59 Making an new branch for each set of related changes will make it easier for
150 Enter a title for the set of changes, and some explanation of what you've done.
151 Say if there is anything you'd like particular attention for - like a
213 There's an embarrassing typo in a commit you made? Or perhaps the you
214 made several false starts you would like the posterity not to see.
282 and the history looks now like this::
317 worry that ``main`` has got a little ahead. But sometimes, the changes in
352 When all looks good you can delete your backup branch::
393 fail with an error that looks like ::
405 If this push had succeed then the commits ``A``, ``B``, and ``C`` would no
406 longer be referenced by any branch and be discarded:
420 versions from our local branch. In this case, what we want to do is ::
431 history and if anyone has fetched the old commits will have a different view
432 of history which can cause confusion.
```
and possibly others that I missed.
### Suggested improvement
Fix typos | 1.0 | [Doc]: Development workflow doc has lots of typos and clunky sentences - ### Documentation Link
https://matplotlib.org/devdocs/devel/development_workflow.html
### Problem
There are typos and/or grammer errors in the following lines on this page at https://github.com/matplotlib/matplotlib/blob/main/doc/devel/development_workflow.rst:
```
59 Making an new branch for each set of related changes will make it easier for
150 Enter a title for the set of changes, and some explanation of what you've done.
151 Say if there is anything you'd like particular attention for - like a
213 There's an embarrassing typo in a commit you made? Or perhaps the you
214 made several false starts you would like the posterity not to see.
282 and the history looks now like this::
317 worry that ``main`` has got a little ahead. But sometimes, the changes in
352 When all looks good you can delete your backup branch::
393 fail with an error that looks like ::
405 If this push had succeed then the commits ``A``, ``B``, and ``C`` would no
406 longer be referenced by any branch and be discarded:
420 versions from our local branch. In this case, what we want to do is ::
431 history and if anyone has fetched the old commits will have a different view
432 of history which can cause confusion.
```
and possibly others that I missed.
### Suggested improvement
Fix typos | non_defect | development workflow doc has lots of typos and clunky sentences documentation link problem there are typos and or grammer errors in the following lines on this page at making an new branch for each set of related changes will make it easier for enter a title for the set of changes and some explanation of what you ve done say if there is anything you d like particular attention for like a there s an embarrassing typo in a commit you made or perhaps the you made several false starts you would like the posterity not to see and the history looks now like this worry that main has got a little ahead but sometimes the changes in when all looks good you can delete your backup branch fail with an error that looks like if this push had succeed then the commits a b and c would no longer be referenced by any branch and be discarded versions from our local branch in this case what we want to do is history and if anyone has fetched the old commits will have a different view of history which can cause confusion and possibly others that i missed suggested improvement fix typos | 0 |
8,072 | 20,789,924,009 | IssuesEvent | 2022-03-17 00:07:31 | RuanScherer/api-health-checker | https://api.github.com/repos/RuanScherer/api-health-checker | closed | Configure ESLint, Prettier and Commitizen tools - frontend | architecture frontend | **Describe the solution you'd like**
Configure ESLint, Prettier and Commitizen tools in the frontend project.
| 1.0 | Configure ESLint, Prettier and Commitizen tools - frontend - **Describe the solution you'd like**
Configure ESLint, Prettier and Commitizen tools in the frontend project.
| non_defect | configure eslint prettier and commitizen tools frontend describe the solution you d like configure eslint prettier and commitizen tools in the frontend project | 0 |
74,614 | 25,212,362,742 | IssuesEvent | 2022-11-14 05:48:46 | hyperledger/iroha | https://api.github.com/repos/hyperledger/iroha | closed | [BUG] Events are emitted even if transaction fails | Bug iroha2 LTS Pre-alpha defect | ### OS and Environment
MacOs
### GIT commit hash
52dc18cd81bdc1d1906ffeecb666dd9b2eb27955
### Minimum working example / Steps to reproduce
create a transaction of multiple instructions where last one fails to execute.
### Actual result
Events are emitted for all instructions up until the first one that fails
### Expected result
No events are emitted
### Logs in JSON format
<details>
<summary>Log contents</summary>
```json
Replace this text with a JSON log,
so it doesn't grow too large and has highlighting.
```
</details>
### Who can help to reproduce?
_No response_
### Notes
Events should be aggregated and only emitted upon transaction successful completion otherwise they should be discarded | 1.0 | [BUG] Events are emitted even if transaction fails - ### OS and Environment
MacOs
### GIT commit hash
52dc18cd81bdc1d1906ffeecb666dd9b2eb27955
### Minimum working example / Steps to reproduce
create a transaction of multiple instructions where last one fails to execute.
### Actual result
Events are emitted for all instructions up until the first one that fails
### Expected result
No events are emitted
### Logs in JSON format
<details>
<summary>Log contents</summary>
```json
Replace this text with a JSON log,
so it doesn't grow too large and has highlighting.
```
</details>
### Who can help to reproduce?
_No response_
### Notes
Events should be aggregated and only emitted upon transaction successful completion otherwise they should be discarded | defect | events are emitted even if transaction fails os and environment macos git commit hash minimum working example steps to reproduce create a transaction of multiple instructions where last one fails to execute actual result events are emitted for all instructions up until the first one that fails expected result no events are emitted logs in json format log contents json replace this text with a json log so it doesn t grow too large and has highlighting who can help to reproduce no response notes events should be aggregated and only emitted upon transaction successful completion otherwise they should be discarded | 1 |
505,443 | 14,633,081,502 | IssuesEvent | 2020-12-24 00:36:06 | containrrr/watchtower | https://api.github.com/repos/containrrr/watchtower | closed | Log "Spam": "Could not do a head request | Priority: Medium Status: Available Type: Bug | **Describe the bug**
I recently updated to the latest watchtower version. But now I seem to get spammed with the error message:
```
Could not do a head request, falling back to regular pull.
```
which gets logged as well as creates an email notification every time.
**To Reproduce**
Steps to reproduce the behavior:
1. Updated watchtower to latest image
2. Deploy a container from a gitlab hosted registry
3. Let it run
4. Monitor logs and emails
**Expected behavior**
Ether a way to disable email notifications in case of this error. Or a fix for why the HEAD request is failing.
**Environment**
- Platform: Ubuntu (Linux)
- Architecture: x86_64
- Docker version: 19.03.13, build 4484c46d9d
<details>
<summary><b>Logs from running watchtower with the <code>--debug</code> option</b></summary>
```
2020-12-20T22:04:01Z [I] Running a one time update.
2020-12-20T22:04:01Z [D] Checking containers for updated images
2020-12-20T22:04:01Z [D] Retrieving running containers
2020-12-20T22:04:01Z [D] Trying to load authentication credentials.
container: /play-for-today-website
image: registry.gitlab.com/***/website:latest
2020-12-20T22:04:01Z [D] Loaded auth credentials for user gitlab+deploy-token-32212, on registry registry.gitlab.com/***/website:latest, from file /config.json
2020-12-20T22:04:01Z [D] Got image name: registry.gitlab.com/***/website:latest
2020-12-20T22:04:01Z [D] Credentials loaded
2020-12-20T22:04:01Z [D] Checking if pull is needed
container: /play-for-today-website
image: registry.gitlab.com/***/website:latest
2020-12-20T22:04:01Z [D] Building challenge URL
URL: https://registry.gitlab.com/v2/
2020-12-20T22:04:02Z [D] Got response to challenge request
header: Bearer realm="https://gitlab.com/jwt/auth",service="container_registry"
status: 401 Unauthorized
2020-12-20T22:04:02Z [D] Checking challenge header content
realm: https://gitlab.com/jwt/auth
service: container_registry
2020-12-20T22:04:02Z [D] Setting scope for auth token
image: registry.gitlab.com/***/website
scope: repository:registry.gitlab.com/***/website:pull
2020-12-20T22:04:02Z [D] Credentials found.
credentials: ******
2020-12-20T22:04:02Z [D] Parsing image ref
host: registry.gitlab.com
image: cogindo/***/website
normalized: registry.gitlab.com/***/website:latest
tag: latest
2020-12-20T22:04:02Z [D] Doing a HEAD request to fetch a digest
url: https://registry.gitlab.com/v2/***/website/manifests/latest
2020-12-20T22:04:02Z [I] Could not do a head request, falling back to regular pull.
2020-12-20T22:04:02Z [D] Reason: registry responded to head request with 401
2020-12-20T22:04:02Z [D] Pulling image
container: /play-for-today-website
image: registry.gitlab.com/***/website:latest
2020-12-20T22:04:04Z [D] No new images found for /play-for-today-website
```
</details>
**Additional context**
The issue seem to happen only for the gitlab registry. Maybe that helps fixing it.
| 1.0 | Log "Spam": "Could not do a head request - **Describe the bug**
I recently updated to the latest watchtower version. But now I seem to get spammed with the error message:
```
Could not do a head request, falling back to regular pull.
```
which gets logged as well as creates an email notification every time.
**To Reproduce**
Steps to reproduce the behavior:
1. Updated watchtower to latest image
2. Deploy a container from a gitlab hosted registry
3. Let it run
4. Monitor logs and emails
**Expected behavior**
Ether a way to disable email notifications in case of this error. Or a fix for why the HEAD request is failing.
**Environment**
- Platform: Ubuntu (Linux)
- Architecture: x86_64
- Docker version: 19.03.13, build 4484c46d9d
<details>
<summary><b>Logs from running watchtower with the <code>--debug</code> option</b></summary>
```
2020-12-20T22:04:01Z [I] Running a one time update.
2020-12-20T22:04:01Z [D] Checking containers for updated images
2020-12-20T22:04:01Z [D] Retrieving running containers
2020-12-20T22:04:01Z [D] Trying to load authentication credentials.
container: /play-for-today-website
image: registry.gitlab.com/***/website:latest
2020-12-20T22:04:01Z [D] Loaded auth credentials for user gitlab+deploy-token-32212, on registry registry.gitlab.com/***/website:latest, from file /config.json
2020-12-20T22:04:01Z [D] Got image name: registry.gitlab.com/***/website:latest
2020-12-20T22:04:01Z [D] Credentials loaded
2020-12-20T22:04:01Z [D] Checking if pull is needed
container: /play-for-today-website
image: registry.gitlab.com/***/website:latest
2020-12-20T22:04:01Z [D] Building challenge URL
URL: https://registry.gitlab.com/v2/
2020-12-20T22:04:02Z [D] Got response to challenge request
header: Bearer realm="https://gitlab.com/jwt/auth",service="container_registry"
status: 401 Unauthorized
2020-12-20T22:04:02Z [D] Checking challenge header content
realm: https://gitlab.com/jwt/auth
service: container_registry
2020-12-20T22:04:02Z [D] Setting scope for auth token
image: registry.gitlab.com/***/website
scope: repository:registry.gitlab.com/***/website:pull
2020-12-20T22:04:02Z [D] Credentials found.
credentials: ******
2020-12-20T22:04:02Z [D] Parsing image ref
host: registry.gitlab.com
image: cogindo/***/website
normalized: registry.gitlab.com/***/website:latest
tag: latest
2020-12-20T22:04:02Z [D] Doing a HEAD request to fetch a digest
url: https://registry.gitlab.com/v2/***/website/manifests/latest
2020-12-20T22:04:02Z [I] Could not do a head request, falling back to regular pull.
2020-12-20T22:04:02Z [D] Reason: registry responded to head request with 401
2020-12-20T22:04:02Z [D] Pulling image
container: /play-for-today-website
image: registry.gitlab.com/***/website:latest
2020-12-20T22:04:04Z [D] No new images found for /play-for-today-website
```
</details>
**Additional context**
The issue seem to happen only for the gitlab registry. Maybe that helps fixing it.
| non_defect | log spam could not do a head request describe the bug i recently updated to the latest watchtower version but now i seem to get spammed with the error message could not do a head request falling back to regular pull which gets logged as well as creates an email notification every time to reproduce steps to reproduce the behavior updated watchtower to latest image deploy a container from a gitlab hosted registry let it run monitor logs and emails expected behavior ether a way to disable email notifications in case of this error or a fix for why the head request is failing environment platform ubuntu linux architecture docker version build logs from running watchtower with the debug option running a one time update checking containers for updated images retrieving running containers trying to load authentication credentials container play for today website image registry gitlab com website latest loaded auth credentials for user gitlab deploy token on registry registry gitlab com website latest from file config json got image name registry gitlab com website latest credentials loaded checking if pull is needed container play for today website image registry gitlab com website latest building challenge url url got response to challenge request header bearer realm status unauthorized checking challenge header content realm service container registry setting scope for auth token image registry gitlab com website scope repository registry gitlab com website pull credentials found credentials parsing image ref host registry gitlab com image cogindo website normalized registry gitlab com website latest tag latest doing a head request to fetch a digest url could not do a head request falling back to regular pull reason registry responded to head request with pulling image container play for today website image registry gitlab com website latest no new images found for play for today website additional context the issue seem to happen only for the gitlab registry maybe that helps fixing it | 0 |
256,055 | 27,552,572,707 | IssuesEvent | 2023-03-07 15:49:58 | BrianMcDonaldWS/genie | https://api.github.com/repos/BrianMcDonaldWS/genie | opened | CVE-2019-0201 (Medium) detected in zookeeper-3.4.12.jar | security vulnerability | ## CVE-2019-0201 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>zookeeper-3.4.12.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /genie-ui/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.zookeeper/zookeeper/3.4.12/cc9c95b358202be355af8abddeb6105f089b1a8c/zookeeper-3.4.12.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.zookeeper/zookeeper/3.4.12/cc9c95b358202be355af8abddeb6105f089b1a8c/zookeeper-3.4.12.jar</p>
<p>
Dependency Hierarchy:
- spring-integration-zookeeper-5.2.2.RELEASE.jar (Root Library)
- curator-recipes-4.0.1.jar
- curator-framework-4.0.1.jar
- curator-client-4.0.1.jar
- :x: **zookeeper-3.4.12.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue is present in Apache ZooKeeper 1.0.0 to 3.4.13 and 3.5.0-alpha to 3.5.4-beta. ZooKeeper’s getACL() command doesn’t check any permission when retrieves the ACLs of the requested node and returns all information contained in the ACL Id field as plaintext string. DigestAuthenticationProvider overloads the Id field with the hash value that is used for user authentication. As a consequence, if Digest Authentication is in use, the unsalted hash value will be disclosed by getACL() request for unauthenticated or unprivileged users.
<p>Publish Date: 2019-05-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-0201>CVE-2019-0201</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://zookeeper.apache.org/security.html">https://zookeeper.apache.org/security.html</a></p>
<p>Release Date: 2019-05-23</p>
<p>Fix Resolution (org.apache.zookeeper:zookeeper): 3.4.14</p>
<p>Direct dependency fix Resolution (org.springframework.integration:spring-integration-zookeeper): 5.2.3.RELEASE</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| True | CVE-2019-0201 (Medium) detected in zookeeper-3.4.12.jar - ## CVE-2019-0201 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>zookeeper-3.4.12.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /genie-ui/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.zookeeper/zookeeper/3.4.12/cc9c95b358202be355af8abddeb6105f089b1a8c/zookeeper-3.4.12.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.zookeeper/zookeeper/3.4.12/cc9c95b358202be355af8abddeb6105f089b1a8c/zookeeper-3.4.12.jar</p>
<p>
Dependency Hierarchy:
- spring-integration-zookeeper-5.2.2.RELEASE.jar (Root Library)
- curator-recipes-4.0.1.jar
- curator-framework-4.0.1.jar
- curator-client-4.0.1.jar
- :x: **zookeeper-3.4.12.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue is present in Apache ZooKeeper 1.0.0 to 3.4.13 and 3.5.0-alpha to 3.5.4-beta. ZooKeeper’s getACL() command doesn’t check any permission when retrieves the ACLs of the requested node and returns all information contained in the ACL Id field as plaintext string. DigestAuthenticationProvider overloads the Id field with the hash value that is used for user authentication. As a consequence, if Digest Authentication is in use, the unsalted hash value will be disclosed by getACL() request for unauthenticated or unprivileged users.
<p>Publish Date: 2019-05-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-0201>CVE-2019-0201</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://zookeeper.apache.org/security.html">https://zookeeper.apache.org/security.html</a></p>
<p>Release Date: 2019-05-23</p>
<p>Fix Resolution (org.apache.zookeeper:zookeeper): 3.4.14</p>
<p>Direct dependency fix Resolution (org.springframework.integration:spring-integration-zookeeper): 5.2.3.RELEASE</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| non_defect | cve medium detected in zookeeper jar cve medium severity vulnerability vulnerable library zookeeper jar path to dependency file genie ui build gradle path to vulnerable library root gradle caches modules files org apache zookeeper zookeeper zookeeper jar root gradle caches modules files org apache zookeeper zookeeper zookeeper jar dependency hierarchy spring integration zookeeper release jar root library curator recipes jar curator framework jar curator client jar x zookeeper jar vulnerable library vulnerability details an issue is present in apache zookeeper to and alpha to beta zookeeper’s getacl command doesn’t check any permission when retrieves the acls of the requested node and returns all information contained in the acl id field as plaintext string digestauthenticationprovider overloads the id field with the hash value that is used for user authentication as a consequence if digest authentication is in use the unsalted hash value will be disclosed by getacl request for unauthenticated or unprivileged users publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache zookeeper zookeeper direct dependency fix resolution org springframework integration spring integration zookeeper release check this box to open an automated fix pr | 0 |
68,932 | 21,966,158,700 | IssuesEvent | 2022-05-24 20:32:39 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | opened | [🐛 Bug]: unhandledPromptBehavior doesn't work | I-defect needs-triaging | ### What happened?
am using python 3.9.12 with pip 22.1.1
my localhost has an index.html file with only one line of code: `<script>alert("hi")</script>`
expected result is to have the above alert dissapear but it doesnt
### How can we reproduce the issue?
```shell
from selenium.webdriver import Chrome
from selenium.webdriver import ChromeOptions
from selenium.webdriver.chrome.service import Service
import time
options = ChromeOptions()
options.unhandledPromptBehavior = "dismiss"
service = Service(executable_path="/usr/local/bin/chromedriver")
driver = Chrome(service=service,options=options)
driver.maximize_window()
driver.get("localhost")
time.sleep(5)
driver.quit()
```
### Relevant log output
```shell
no output.
```
### Operating System
macos Montery 12.3.1
### Selenium version
4.1.5
### What are the browser(s) and version(s) where you see this issue?
chrome 101.0.4951.64 arm64
### What are the browser driver(s) and version(s) where you see this issue?
Chromedriver 101.0.4951.41
### Are you using Selenium Grid?
no | 1.0 | [🐛 Bug]: unhandledPromptBehavior doesn't work - ### What happened?
am using python 3.9.12 with pip 22.1.1
my localhost has an index.html file with only one line of code: `<script>alert("hi")</script>`
expected result is to have the above alert dissapear but it doesnt
### How can we reproduce the issue?
```shell
from selenium.webdriver import Chrome
from selenium.webdriver import ChromeOptions
from selenium.webdriver.chrome.service import Service
import time
options = ChromeOptions()
options.unhandledPromptBehavior = "dismiss"
service = Service(executable_path="/usr/local/bin/chromedriver")
driver = Chrome(service=service,options=options)
driver.maximize_window()
driver.get("localhost")
time.sleep(5)
driver.quit()
```
### Relevant log output
```shell
no output.
```
### Operating System
macos Montery 12.3.1
### Selenium version
4.1.5
### What are the browser(s) and version(s) where you see this issue?
chrome 101.0.4951.64 arm64
### What are the browser driver(s) and version(s) where you see this issue?
Chromedriver 101.0.4951.41
### Are you using Selenium Grid?
no | defect | unhandledpromptbehavior doesn t work what happened am using python with pip my localhost has an index html file with only one line of code alert hi expected result is to have the above alert dissapear but it doesnt how can we reproduce the issue shell from selenium webdriver import chrome from selenium webdriver import chromeoptions from selenium webdriver chrome service import service import time options chromeoptions options unhandledpromptbehavior dismiss service service executable path usr local bin chromedriver driver chrome service service options options driver maximize window driver get localhost time sleep driver quit relevant log output shell no output operating system macos montery selenium version what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver are you using selenium grid no | 1 |
604,032 | 18,675,960,492 | IssuesEvent | 2021-10-31 15:10:01 | CMPUT301F21T26/Habit-Tracker | https://api.github.com/repos/CMPUT301F21T26/Habit-Tracker | closed | Replace default Action Bar with custom Tool Bar | Priority: Medium Base | Default Action Bar lacks flexibility and will need to be replace with a Tool Bar widget (part of the library). To remove actionbar, in the themes.xml: change Theme.MaterialComponents.DayNight.DarkActionBar to Theme.MaterialComponents.DayNight.NoActionBar. | 1.0 | Replace default Action Bar with custom Tool Bar - Default Action Bar lacks flexibility and will need to be replace with a Tool Bar widget (part of the library). To remove actionbar, in the themes.xml: change Theme.MaterialComponents.DayNight.DarkActionBar to Theme.MaterialComponents.DayNight.NoActionBar. | non_defect | replace default action bar with custom tool bar default action bar lacks flexibility and will need to be replace with a tool bar widget part of the library to remove actionbar in the themes xml change theme materialcomponents daynight darkactionbar to theme materialcomponents daynight noactionbar | 0 |
111,076 | 14,005,602,523 | IssuesEvent | 2020-10-28 18:41:55 | Opentrons/opentrons | https://api.github.com/repos/Opentrons/opentrons | opened | PD Bug: Weird tip rack flow results in white screen | :spider: SPDDRS bug protocol designer | ## Steps to re-create
1. Create a protocol with a custom tip rack
2. Delete (on purpose or accidentally) the custom tip rack from the starting deck screen
3. Did NOT update the tip rack in the pipette selection modal with the new desired tip rack | 1.0 | PD Bug: Weird tip rack flow results in white screen - ## Steps to re-create
1. Create a protocol with a custom tip rack
2. Delete (on purpose or accidentally) the custom tip rack from the starting deck screen
3. Did NOT update the tip rack in the pipette selection modal with the new desired tip rack | non_defect | pd bug weird tip rack flow results in white screen steps to re create create a protocol with a custom tip rack delete on purpose or accidentally the custom tip rack from the starting deck screen did not update the tip rack in the pipette selection modal with the new desired tip rack | 0 |
4,675 | 4,534,845,383 | IssuesEvent | 2016-09-08 15:38:22 | JuliaLang/julia | https://api.github.com/repos/JuliaLang/julia | opened | optimize @evalpoly for SIMD? | maths performance | Right now, `@evalpoly(x, coefs...)` uses Horner's rule for real `x`, and a more complicated algorithm for complex `x`, and is performance-critical for evaluation of special functions.
However, it seems likely to be faster to try and exploit SIMD instructions for these polynomial evaluations, e.g. with [Estrin's algorithm](https://en.wikipedia.org/wiki/Estrin%27s_scheme) or some other technique. (The nice thing about code generation / macros is that we can investigate fancy methods for this sort of thing.) | True | optimize @evalpoly for SIMD? - Right now, `@evalpoly(x, coefs...)` uses Horner's rule for real `x`, and a more complicated algorithm for complex `x`, and is performance-critical for evaluation of special functions.
However, it seems likely to be faster to try and exploit SIMD instructions for these polynomial evaluations, e.g. with [Estrin's algorithm](https://en.wikipedia.org/wiki/Estrin%27s_scheme) or some other technique. (The nice thing about code generation / macros is that we can investigate fancy methods for this sort of thing.) | non_defect | optimize evalpoly for simd right now evalpoly x coefs uses horner s rule for real x and a more complicated algorithm for complex x and is performance critical for evaluation of special functions however it seems likely to be faster to try and exploit simd instructions for these polynomial evaluations e g with or some other technique the nice thing about code generation macros is that we can investigate fancy methods for this sort of thing | 0 |
140,691 | 32,047,730,257 | IssuesEvent | 2023-09-23 07:10:07 | h4sh5/pypi-auto-scanner | https://api.github.com/repos/h4sh5/pypi-auto-scanner | opened | deme 0.4.4 has 3 GuardDog issues | guarddog code-execution exec-base64 | https://pypi.org/project/deme
https://inspector.pypi.io/project/deme
```{
"dependency": "deme",
"version": "0.4.4",
"result": {
"issues": 3,
"errors": {},
"results": {
"exec-base64": [
{
"location": "DEME-0.4.4/setup.py:136",
"code": " subprocess.run(\n [\"cmake\", ext.sourcedir, *cmake_args], cwd=build_temp, check=True\n )",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"code-execution": [
{
"location": "DEME-0.4.4/setup.py:144",
"code": " subprocess.run([\"make\", \"install\"], cwd=build_temp, check=True)",
"message": "This package is executing OS commands in the setup.py file"
},
{
"location": "DEME-0.4.4/thirdparty/pybind11/setup.py:139",
"code": " subprocess.run(cmd, check=True, cwd=DIR, stdout=sys.stdout, stderr=sys.stderr)",
"message": "This package is executing OS commands in the setup.py file"
}
]
},
"path": "/tmp/tmphegw186s/deme"
}
}``` | 1.0 | deme 0.4.4 has 3 GuardDog issues - https://pypi.org/project/deme
https://inspector.pypi.io/project/deme
```{
"dependency": "deme",
"version": "0.4.4",
"result": {
"issues": 3,
"errors": {},
"results": {
"exec-base64": [
{
"location": "DEME-0.4.4/setup.py:136",
"code": " subprocess.run(\n [\"cmake\", ext.sourcedir, *cmake_args], cwd=build_temp, check=True\n )",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"code-execution": [
{
"location": "DEME-0.4.4/setup.py:144",
"code": " subprocess.run([\"make\", \"install\"], cwd=build_temp, check=True)",
"message": "This package is executing OS commands in the setup.py file"
},
{
"location": "DEME-0.4.4/thirdparty/pybind11/setup.py:139",
"code": " subprocess.run(cmd, check=True, cwd=DIR, stdout=sys.stdout, stderr=sys.stderr)",
"message": "This package is executing OS commands in the setup.py file"
}
]
},
"path": "/tmp/tmphegw186s/deme"
}
}``` | non_defect | deme has guarddog issues dependency deme version result issues errors results exec location deme setup py code subprocess run n cwd build temp check true n message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n code execution location deme setup py code subprocess run cwd build temp check true message this package is executing os commands in the setup py file location deme thirdparty setup py code subprocess run cmd check true cwd dir stdout sys stdout stderr sys stderr message this package is executing os commands in the setup py file path tmp deme | 0 |
27,152 | 4,889,269,946 | IssuesEvent | 2016-11-18 09:38:49 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | INSERT ... SET ... RETURNING returns null | C: Functionality P: High T: Defect | It appears that this doesn't work:
``` java
TodosRecord persisted = jooq.insertInto(TODOS)
.set(TODOS.CREATION_TIME, currentTime)
.set(TODOS.DESCRIPTION, todo.getDescription())
.set(TODOS.ID, id)
.set(TODOS.MODIFICATION_TIME, currentTime)
.set(TODOS.TITLE, todo.getTitle())
.returning()
.fetchOne();
//persisted is always null
```
... whereas this does:
``` java
TodosRecord persisted = jooq.insertInto(TODOS)
.set(createRecord(todo))
.returning()
.fetchOne();
//Persisted is found
```
---
See also:
https://groups.google.com/forum/#!topic/jooq-user/fOsaaZq44Ac
| 1.0 | INSERT ... SET ... RETURNING returns null - It appears that this doesn't work:
``` java
TodosRecord persisted = jooq.insertInto(TODOS)
.set(TODOS.CREATION_TIME, currentTime)
.set(TODOS.DESCRIPTION, todo.getDescription())
.set(TODOS.ID, id)
.set(TODOS.MODIFICATION_TIME, currentTime)
.set(TODOS.TITLE, todo.getTitle())
.returning()
.fetchOne();
//persisted is always null
```
... whereas this does:
``` java
TodosRecord persisted = jooq.insertInto(TODOS)
.set(createRecord(todo))
.returning()
.fetchOne();
//Persisted is found
```
---
See also:
https://groups.google.com/forum/#!topic/jooq-user/fOsaaZq44Ac
| defect | insert set returning returns null it appears that this doesn t work java todosrecord persisted jooq insertinto todos set todos creation time currenttime set todos description todo getdescription set todos id id set todos modification time currenttime set todos title todo gettitle returning fetchone persisted is always null whereas this does java todosrecord persisted jooq insertinto todos set createrecord todo returning fetchone persisted is found see also | 1 |
38,203 | 8,695,528,566 | IssuesEvent | 2018-12-04 15:22:46 | jOOQ/jOOR | https://api.github.com/repos/jOOQ/jOOR | closed | Cannot compile nested or additional top-level classes | P: Medium R: Fixed T: Defect | ### Expected behavior and actual behavior:
I have this class that is loaded and compiled through jOOR:
```java
package com.example;
import org.apache.camel.Exchange;
import org.apache.camel.Processor;
import org.apache.camel.builder.RouteBuilder;
public class WithCustomProcessor extends RouteBuilder {
@Override
public void configure() throws Exception {
Processor ucp = new Processor() {
public void process(Exchange exchange) throws Exception {
String body = exchange.getIn().getBody(String.class);
body = body.toUpperCase();
exchange.getOut().setBody(body);
}
};
from("timer:tick?period=10s")
.setBody().constant("Hello")
.process(ucp)
.to("log:info");
}
}
```
But when the the method `configure` is invoked from the compiled object, the following exception is thrown:
```
Exception in thread "main" java.lang.NoClassDefFoundError: com/example/WithCustomProcessor$1
at com.example.WithCustomProcessor.configure(WithCustomProcessor.java:13)
at com.github.lburgazzoli.JoorTest.main(JoorTest.java:15)
Caused by: java.lang.ClassNotFoundException: com.example.WithCustomProcessor$1
at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:349)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
... 2 more
```
### Steps to reproduce the problem:
I have a repo with a reproducer [here](https://github.com/lburgazzoli/camel-joor-test).
Running the main class will reproduce the issue.
### Versions:
- jOOR: 0.9.9
- Java: openjdk version "1.8.0_191"
| 1.0 | Cannot compile nested or additional top-level classes - ### Expected behavior and actual behavior:
I have this class that is loaded and compiled through jOOR:
```java
package com.example;
import org.apache.camel.Exchange;
import org.apache.camel.Processor;
import org.apache.camel.builder.RouteBuilder;
public class WithCustomProcessor extends RouteBuilder {
@Override
public void configure() throws Exception {
Processor ucp = new Processor() {
public void process(Exchange exchange) throws Exception {
String body = exchange.getIn().getBody(String.class);
body = body.toUpperCase();
exchange.getOut().setBody(body);
}
};
from("timer:tick?period=10s")
.setBody().constant("Hello")
.process(ucp)
.to("log:info");
}
}
```
But when the the method `configure` is invoked from the compiled object, the following exception is thrown:
```
Exception in thread "main" java.lang.NoClassDefFoundError: com/example/WithCustomProcessor$1
at com.example.WithCustomProcessor.configure(WithCustomProcessor.java:13)
at com.github.lburgazzoli.JoorTest.main(JoorTest.java:15)
Caused by: java.lang.ClassNotFoundException: com.example.WithCustomProcessor$1
at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:349)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
... 2 more
```
### Steps to reproduce the problem:
I have a repo with a reproducer [here](https://github.com/lburgazzoli/camel-joor-test).
Running the main class will reproduce the issue.
### Versions:
- jOOR: 0.9.9
- Java: openjdk version "1.8.0_191"
| defect | cannot compile nested or additional top level classes expected behavior and actual behavior i have this class that is loaded and compiled through joor java package com example import org apache camel exchange import org apache camel processor import org apache camel builder routebuilder public class withcustomprocessor extends routebuilder override public void configure throws exception processor ucp new processor public void process exchange exchange throws exception string body exchange getin getbody string class body body touppercase exchange getout setbody body from timer tick period setbody constant hello process ucp to log info but when the the method configure is invoked from the compiled object the following exception is thrown exception in thread main java lang noclassdeffounderror com example withcustomprocessor at com example withcustomprocessor configure withcustomprocessor java at com github lburgazzoli joortest main joortest java caused by java lang classnotfoundexception com example withcustomprocessor at java net urlclassloader findclass urlclassloader java at java lang classloader loadclass classloader java at sun misc launcher appclassloader loadclass launcher java at java lang classloader loadclass classloader java more steps to reproduce the problem i have a repo with a reproducer running the main class will reproduce the issue versions joor java openjdk version | 1 |
37,185 | 8,288,778,203 | IssuesEvent | 2018-09-19 13:03:13 | hazelcast/hazelcast-jet | https://api.github.com/repos/hazelcast/hazelcast-jet | opened | Logs for ReceiverTasklet contain no context | core defect | The logger name doesn't include vertex name or the job name
```
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800``` | 1.0 | Logs for ReceiverTasklet contain no context - The logger name doesn't include vertex name or the job name
```
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800
00:01:25,509 TRACE || - [ReceiverTasklet] hz._hzInstance_1_jet.cached.thread-4 - receiveWindowCompressed=800``` | defect | logs for receivertasklet contain no context the logger name doesn t include vertex name or the job name trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed trace hz hzinstance jet cached thread receivewindowcompressed | 1 |
117,075 | 17,407,947,288 | IssuesEvent | 2021-08-03 08:36:57 | elikkatzgit/TestingPOM | https://api.github.com/repos/elikkatzgit/TestingPOM | closed | CVE-2018-5968 (High) detected in jackson-databind-2.7.2.jar - autoclosed | security vulnerability | ## CVE-2018-5968 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind through 2.8.11 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 and CVE-2017-17485 deserialization flaws. This is exploitable via two different gadgets that bypass a blacklist.
<p>Publish Date: 2018-01-22
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968>CVE-2018-5968</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968</a></p>
<p>Release Date: 2018-01-22</p>
<p>Fix Resolution: 2.8.11.1, 2.9.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.2","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.11.1, 2.9.4"}],"vulnerabilityIdentifier":"CVE-2018-5968","vulnerabilityDetails":"FasterXML jackson-databind through 2.8.11 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 and CVE-2017-17485 deserialization flaws. This is exploitable via two different gadgets that bypass a blacklist.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2018-5968","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-5968 (High) detected in jackson-databind-2.7.2.jar - autoclosed - ## CVE-2018-5968 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind through 2.8.11 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 and CVE-2017-17485 deserialization flaws. This is exploitable via two different gadgets that bypass a blacklist.
<p>Publish Date: 2018-01-22
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968>CVE-2018-5968</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968</a></p>
<p>Release Date: 2018-01-22</p>
<p>Fix Resolution: 2.8.11.1, 2.9.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.2","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.11.1, 2.9.4"}],"vulnerabilityIdentifier":"CVE-2018-5968","vulnerabilityDetails":"FasterXML jackson-databind through 2.8.11 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 and CVE-2017-17485 deserialization flaws. This is exploitable via two different gadgets that bypass a blacklist.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2018-5968","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_defect | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind through and x through allows unauthenticated remote code execution because of an incomplete fix for the cve and cve deserialization flaws this is exploitable via two different gadgets that bypass a blacklist publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind through and x through allows unauthenticated remote code execution because of an incomplete fix for the cve and cve deserialization flaws this is exploitable via two different gadgets that bypass a blacklist vulnerabilityurl | 0 |
64,908 | 18,960,439,243 | IssuesEvent | 2021-11-19 03:35:37 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | typo in jOOQ-academy setup script | T: Defect | There is an extraneous comma at the end of line 34 (`fk_t_book_author_id`) in the db-h2.sql setup script:
https://github.com/jOOQ/jOOQ/blob/e77df40080b9f73a2da1b7436ee875ad2973731c/jOOQ-examples/jOOQ-academy/src/main/resources/db-h2.sql#L25-L36
This causes `maven clean install` in jOOQ-academy to fail:
```
[ERROR] CONSTRAINT PK_T_BOOK PRIMARY KEY (ID),
[ERROR] CONSTRAINT FK_T_BOOK_AUTHOR_ID FOREIGN KEY (AUTHOR_ID) REFERENCES AUTHOR(ID),
[ERROR] )[*]
[ERROR] "; expected "identifier"; SQL statement:
[ERROR]
```
### Versions
- jOOQ: 3.16.0 | 1.0 | typo in jOOQ-academy setup script - There is an extraneous comma at the end of line 34 (`fk_t_book_author_id`) in the db-h2.sql setup script:
https://github.com/jOOQ/jOOQ/blob/e77df40080b9f73a2da1b7436ee875ad2973731c/jOOQ-examples/jOOQ-academy/src/main/resources/db-h2.sql#L25-L36
This causes `maven clean install` in jOOQ-academy to fail:
```
[ERROR] CONSTRAINT PK_T_BOOK PRIMARY KEY (ID),
[ERROR] CONSTRAINT FK_T_BOOK_AUTHOR_ID FOREIGN KEY (AUTHOR_ID) REFERENCES AUTHOR(ID),
[ERROR] )[*]
[ERROR] "; expected "identifier"; SQL statement:
[ERROR]
```
### Versions
- jOOQ: 3.16.0 | defect | typo in jooq academy setup script there is an extraneous comma at the end of line fk t book author id in the db sql setup script this causes maven clean install in jooq academy to fail constraint pk t book primary key id constraint fk t book author id foreign key author id references author id expected identifier sql statement versions jooq | 1 |
59,889 | 17,023,280,386 | IssuesEvent | 2021-07-03 01:12:20 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | change look of highway=road in potlatch | Component: potlatch (flash editor) Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 11.01am, Tuesday, 29th July 2008]**
It seems Potlatch does not recognise ways tagged as highway=road, so they appear as a grey line in the editor. This makes them hard to spot when aerial imagery is turned on. | 1.0 | change look of highway=road in potlatch - **[Submitted to the original trac issue database at 11.01am, Tuesday, 29th July 2008]**
It seems Potlatch does not recognise ways tagged as highway=road, so they appear as a grey line in the editor. This makes them hard to spot when aerial imagery is turned on. | defect | change look of highway road in potlatch it seems potlatch does not recognise ways tagged as highway road so they appear as a grey line in the editor this makes them hard to spot when aerial imagery is turned on | 1 |
10,100 | 14,528,458,704 | IssuesEvent | 2020-12-14 16:33:41 | OpenEnergyPlatform/tutorial | https://api.github.com/repos/OpenEnergyPlatform/tutorial | closed | Include in Tutorials first steps | requirement_specification specification_sheet | The Tutorials start with an Installed Anaconda version. It would be good to include the first steps for first time users: setup of virtual environment etc.
If possible it would be nice if the tutorials could be grouped into: first time user, some python experience, expert and hence start at the respective points. | 1.0 | Include in Tutorials first steps - The Tutorials start with an Installed Anaconda version. It would be good to include the first steps for first time users: setup of virtual environment etc.
If possible it would be nice if the tutorials could be grouped into: first time user, some python experience, expert and hence start at the respective points. | non_defect | include in tutorials first steps the tutorials start with an installed anaconda version it would be good to include the first steps for first time users setup of virtual environment etc if possible it would be nice if the tutorials could be grouped into first time user some python experience expert and hence start at the respective points | 0 |
40,622 | 10,075,011,645 | IssuesEvent | 2019-07-24 13:25:08 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | ExpressionChangedAfterItHasBeenCheckedError when preselect a p-radioButton | defect | **I'm submitting a ...**
```
[x] bug report
[ ] feature request
[ ] support request
```
**Stackblitz demo**
Minimal Angular 8/PrimeNg 8 app that demonstrates the issue
https://stackblitz.com/edit/angular-zhmbkf
**Current behavior**
Initially getting `ExpressionChangedAfterItHasBeenCheckedError` (exactly like in #3100) after upgrading to Angular 8 and PrimeNG 8. In Angular 7/PrimeNg 7 the error did not occur.
_My 2 cents:_
Looks like setting `static` to `true` in
```
@ViewChild('rb', { static: false }) inputViewChild: ElementRef;
```
(https://github.com/primefaces/primeng/blob/master/src/app/components/radiobutton/radiobutton.ts#L57)
resolves the issue.
**Expected behavior**
Well, no error would be nice :-)
**Minimal reproduction of the problem with instructions**
Just open the Stackblitz and you will see the error in the console
**Please tell us about your environment:**
<!-- Ubuntu Linux, Chromium 75 -->
* **Angular version:** 8.1.1
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 8.0.1
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** Chromium 75.0.3770.90
* **Language:** TypeScript 3.4
| 1.0 | ExpressionChangedAfterItHasBeenCheckedError when preselect a p-radioButton - **I'm submitting a ...**
```
[x] bug report
[ ] feature request
[ ] support request
```
**Stackblitz demo**
Minimal Angular 8/PrimeNg 8 app that demonstrates the issue
https://stackblitz.com/edit/angular-zhmbkf
**Current behavior**
Initially getting `ExpressionChangedAfterItHasBeenCheckedError` (exactly like in #3100) after upgrading to Angular 8 and PrimeNG 8. In Angular 7/PrimeNg 7 the error did not occur.
_My 2 cents:_
Looks like setting `static` to `true` in
```
@ViewChild('rb', { static: false }) inputViewChild: ElementRef;
```
(https://github.com/primefaces/primeng/blob/master/src/app/components/radiobutton/radiobutton.ts#L57)
resolves the issue.
**Expected behavior**
Well, no error would be nice :-)
**Minimal reproduction of the problem with instructions**
Just open the Stackblitz and you will see the error in the console
**Please tell us about your environment:**
<!-- Ubuntu Linux, Chromium 75 -->
* **Angular version:** 8.1.1
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 8.0.1
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** Chromium 75.0.3770.90
* **Language:** TypeScript 3.4
| defect | expressionchangedafterithasbeencheckederror when preselect a p radiobutton i m submitting a bug report feature request support request stackblitz demo minimal angular primeng app that demonstrates the issue current behavior initially getting expressionchangedafterithasbeencheckederror exactly like in after upgrading to angular and primeng in angular primeng the error did not occur my cents looks like setting static to true in viewchild rb static false inputviewchild elementref resolves the issue expected behavior well no error would be nice minimal reproduction of the problem with instructions just open the stackblitz and you will see the error in the console please tell us about your environment angular version primeng version browser chromium language typescript | 1 |
27,563 | 5,048,378,835 | IssuesEvent | 2016-12-20 12:43:25 | TASVideos/BizHawk | https://api.github.com/repos/TASVideos/BizHawk | closed | The audio of Chinese version of Metroid Fusion is glitchy | auto-migrated Core-VBANext Type-Defect | ```
Bizhawk 1.8.2
Yes, I mean the official Chinese version.
The audio is fine in both English and Japanese version.
Not sure about the European version.
```
Original issue reported on code.google.com by `Fortr...@gmail.com` on 11 Oct 2014 at 3:31
| 1.0 | The audio of Chinese version of Metroid Fusion is glitchy - ```
Bizhawk 1.8.2
Yes, I mean the official Chinese version.
The audio is fine in both English and Japanese version.
Not sure about the European version.
```
Original issue reported on code.google.com by `Fortr...@gmail.com` on 11 Oct 2014 at 3:31
| defect | the audio of chinese version of metroid fusion is glitchy bizhawk yes i mean the official chinese version the audio is fine in both english and japanese version not sure about the european version original issue reported on code google com by fortr gmail com on oct at | 1 |
416,732 | 28,097,698,583 | IssuesEvent | 2023-03-30 16:58:38 | golang/go | https://api.github.com/repos/golang/go | closed | x/tools/go/analysis: pkgsite documentation for golang.org/x/tools/go/analysis/passes/* often lacks details | Documentation Tools pkgsite | I am not sure if this is a feature request for pkgsite, or it's just because the go analysis analyzer code was structured in a way not friendly with go documentation. Analyzers often place the helpful details in Doc constant, but this long doc is hidden from documentation
https://pkg.go.dev/golang.org/x/tools/go/analysis/passes/nilness

```
$ go doc
package nilness // import "golang.org/x/tools/go/analysis/passes/nilness"
Package nilness inspects the control-flow graph of an SSA function and reports
errors such as nil pointer dereferences and degenerate nil pointer comparisons.
const Doc = ...
var Analyzer = &analysis.Analyzer{ ... }
$ go doc Doc
```
Would be nice if the documentation page of analyzers presents sufficient details.
The pattern I saw in similar situation was to generate the package doc (doc.go) from the types but not sure if that's the best way.
cc @adonovan @golang/tools-team | 1.0 | x/tools/go/analysis: pkgsite documentation for golang.org/x/tools/go/analysis/passes/* often lacks details - I am not sure if this is a feature request for pkgsite, or it's just because the go analysis analyzer code was structured in a way not friendly with go documentation. Analyzers often place the helpful details in Doc constant, but this long doc is hidden from documentation
https://pkg.go.dev/golang.org/x/tools/go/analysis/passes/nilness

```
$ go doc
package nilness // import "golang.org/x/tools/go/analysis/passes/nilness"
Package nilness inspects the control-flow graph of an SSA function and reports
errors such as nil pointer dereferences and degenerate nil pointer comparisons.
const Doc = ...
var Analyzer = &analysis.Analyzer{ ... }
$ go doc Doc
```
Would be nice if the documentation page of analyzers presents sufficient details.
The pattern I saw in similar situation was to generate the package doc (doc.go) from the types but not sure if that's the best way.
cc @adonovan @golang/tools-team | non_defect | x tools go analysis pkgsite documentation for golang org x tools go analysis passes often lacks details i am not sure if this is a feature request for pkgsite or it s just because the go analysis analyzer code was structured in a way not friendly with go documentation analyzers often place the helpful details in doc constant but this long doc is hidden from documentation go doc package nilness import golang org x tools go analysis passes nilness package nilness inspects the control flow graph of an ssa function and reports errors such as nil pointer dereferences and degenerate nil pointer comparisons const doc var analyzer analysis analyzer go doc doc would be nice if the documentation page of analyzers presents sufficient details the pattern i saw in similar situation was to generate the package doc doc go from the types but not sure if that s the best way cc adonovan golang tools team | 0 |
27,063 | 12,510,687,569 | IssuesEvent | 2020-06-02 19:07:26 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | No updates available for 'azure-cli-ml' - extension not available | Machine Learning Service Attention question | > ### `az feedback` auto-generates most of the information requested below, as of CLI version 2.0.62
**Describe the bug**
<!--- A clear and concise description of what the bug is. --->
Until yesterday azure-cli-ml extension was available to be installed through command az extension install -n azure-cli-ml
**To Reproduce**
<!--- Steps to reproduce the behavior. --->
type: "**az extension install -n azure-cli-ml**" or "**az extension update -n azure-cli-ml**"
This error message appear:
**No updates available for 'azure-cli-ml'. Use --debug for more information.**
**Expected behavior**
<!--- A clear and concise description of what you expected to happen. --->
Azure CLI ML extension should be installed successfully
**Environment summary**
<!--- Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows) --->
Any platform with Azure CLI
**Additional context**
<!--- Add any other context about the problem here. --->
There's additional information log:
**urllib3.connectionpool : Starting new HTTPS connection (1): aka.ms:443
urllib3.connectionpool : https://aka.ms:443 "GET /azure-cli-extension-index-v1 HTTP/1.1" 301 0
urllib3.connectionpool : Starting new HTTPS connection (1): azurecliextensionsync.blob.core.windows.net:443
urllib3.connectionpool : https://azurecliextensionsync.blob.core.windows.net:443 "GET /index1/index.json HTTP/1.1" 200 309315
Candidates ['azure_cli_ml-1.6.0-py3-none-any.whl']
Skipping 'azure_cli_ml-1.6.0-py3-none-any.whl' as not universal wheel.We do not currently support platform specific extension detection. They can be installed with the full URL https://azurecliext.blob.core.windows.net/release/azure_cli_ml-1.6.0-py3-none-any.whl
Candidates []
Candidates []
No suitable extensions found.
cli.azure.cli.core.util : No updates available for 'azure-cli-ml'. Use --debug for more information.
No updates available for 'azure-cli-ml'. Use --debug for more information.**
| 1.0 | No updates available for 'azure-cli-ml' - extension not available - > ### `az feedback` auto-generates most of the information requested below, as of CLI version 2.0.62
**Describe the bug**
<!--- A clear and concise description of what the bug is. --->
Until yesterday azure-cli-ml extension was available to be installed through command az extension install -n azure-cli-ml
**To Reproduce**
<!--- Steps to reproduce the behavior. --->
type: "**az extension install -n azure-cli-ml**" or "**az extension update -n azure-cli-ml**"
This error message appear:
**No updates available for 'azure-cli-ml'. Use --debug for more information.**
**Expected behavior**
<!--- A clear and concise description of what you expected to happen. --->
Azure CLI ML extension should be installed successfully
**Environment summary**
<!--- Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows) --->
Any platform with Azure CLI
**Additional context**
<!--- Add any other context about the problem here. --->
There's additional information log:
**urllib3.connectionpool : Starting new HTTPS connection (1): aka.ms:443
urllib3.connectionpool : https://aka.ms:443 "GET /azure-cli-extension-index-v1 HTTP/1.1" 301 0
urllib3.connectionpool : Starting new HTTPS connection (1): azurecliextensionsync.blob.core.windows.net:443
urllib3.connectionpool : https://azurecliextensionsync.blob.core.windows.net:443 "GET /index1/index.json HTTP/1.1" 200 309315
Candidates ['azure_cli_ml-1.6.0-py3-none-any.whl']
Skipping 'azure_cli_ml-1.6.0-py3-none-any.whl' as not universal wheel.We do not currently support platform specific extension detection. They can be installed with the full URL https://azurecliext.blob.core.windows.net/release/azure_cli_ml-1.6.0-py3-none-any.whl
Candidates []
Candidates []
No suitable extensions found.
cli.azure.cli.core.util : No updates available for 'azure-cli-ml'. Use --debug for more information.
No updates available for 'azure-cli-ml'. Use --debug for more information.**
| non_defect | no updates available for azure cli ml extension not available az feedback auto generates most of the information requested below as of cli version describe the bug until yesterday azure cli ml extension was available to be installed through command az extension install n azure cli ml to reproduce type az extension install n azure cli ml or az extension update n azure cli ml this error message appear no updates available for azure cli ml use debug for more information expected behavior azure cli ml extension should be installed successfully environment summary any platform with azure cli additional context there s additional information log connectionpool starting new https connection aka ms connectionpool get azure cli extension index http connectionpool starting new https connection azurecliextensionsync blob core windows net connectionpool get index json http candidates skipping azure cli ml none any whl as not universal wheel we do not currently support platform specific extension detection they can be installed with the full url candidates candidates no suitable extensions found cli azure cli core util no updates available for azure cli ml use debug for more information no updates available for azure cli ml use debug for more information | 0 |
64,010 | 18,107,106,444 | IssuesEvent | 2021-09-22 20:26:51 | rropen/SFM | https://api.github.com/repos/rropen/SFM | opened | Defect: Fix Broken Code Coverage | defect frontend | **Topic**
Code Coverage is running, but is incorrect.
**Service**
Frontend
**Steps**
- [ ] Utils is heavily tested, but showing as not tested. Other sections of app are not tested but show as being tested.
| 1.0 | Defect: Fix Broken Code Coverage - **Topic**
Code Coverage is running, but is incorrect.
**Service**
Frontend
**Steps**
- [ ] Utils is heavily tested, but showing as not tested. Other sections of app are not tested but show as being tested.
| defect | defect fix broken code coverage topic code coverage is running but is incorrect service frontend steps utils is heavily tested but showing as not tested other sections of app are not tested but show as being tested | 1 |
45,749 | 13,055,639,807 | IssuesEvent | 2020-07-30 02:17:51 | svalinn/DAGMC | https://api.github.com/repos/svalinn/DAGMC | opened | DAGMC not compatible with Geant4 10.6 | Type: Defect | There appear to have been some changes in Geant4 10.6 that break our ability to build Dag-Geant4 applications. Stuff like this appears during build:
```
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc: In function ‘int main(int, char**)’:
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:84:5: error: ‘G4UIExecutive’ was not declared in this scope
84 | G4UIExecutive* ui = new G4UIExecutive(argc, argv, "tcsh");
| ^~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:84:20: error: ‘ui’ was not declared in this scope
84 | G4UIExecutive* ui = new G4UIExecutive(argc, argv, "tcsh");
| ^~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:84:29: error: expected type-specifier before ‘G4UIExecutive’
84 | G4UIExecutive* ui = new G4UIExecutive(argc, argv, "tcsh");
| ^~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:87:12: error: type ‘<type error>’ argument given to ‘delete’, expected pointer
87 | delete ui;
| ^~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:89:5: error: ‘G4VisManager’ was not declared in this scope; did you mean ‘G4VPCManager’?
89 | G4VisManager* visManager = new G4VisExecutive;
| ^~~~~~~~~~~~
| G4VPCManager
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:89:19: error: ‘visManager’ was not declared in this scope; did you mean ‘scManager’?
89 | G4VisManager* visManager = new G4VisExecutive;
| ^~~~~~~~~~
| scManager
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:89:36: error: ‘G4VisExecutive’ does not name a type
89 | G4VisManager* visManager = new G4VisExecutive;
| ^~~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:92:5: error: ‘G4UIExecutive’ was not declared in this scope
92 | G4UIExecutive* UI = new G4UIExecutive(argc, argv);
| ^~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:92:20: error: ‘UI’ was not declared in this scope
92 | G4UIExecutive* UI = new G4UIExecutive(argc, argv);
| ^~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:92:29: error: expected type-specifier before ‘G4UIExecutive’
92 | G4UIExecutive* UI = new G4UIExecutive(argc, argv);
| ^~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:96:12: error: type ‘<type error>’ argument given to ‘delete’, expected pointer
96 | delete visManager;
| ^~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:97:12: error: type ‘<type error>’ argument given to ‘delete’, expected pointer
97 | delete UI;
| ^~
``` | 1.0 | DAGMC not compatible with Geant4 10.6 - There appear to have been some changes in Geant4 10.6 that break our ability to build Dag-Geant4 applications. Stuff like this appears during build:
```
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc: In function ‘int main(int, char**)’:
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:84:5: error: ‘G4UIExecutive’ was not declared in this scope
84 | G4UIExecutive* ui = new G4UIExecutive(argc, argv, "tcsh");
| ^~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:84:20: error: ‘ui’ was not declared in this scope
84 | G4UIExecutive* ui = new G4UIExecutive(argc, argv, "tcsh");
| ^~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:84:29: error: expected type-specifier before ‘G4UIExecutive’
84 | G4UIExecutive* ui = new G4UIExecutive(argc, argv, "tcsh");
| ^~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:87:12: error: type ‘<type error>’ argument given to ‘delete’, expected pointer
87 | delete ui;
| ^~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:89:5: error: ‘G4VisManager’ was not declared in this scope; did you mean ‘G4VPCManager’?
89 | G4VisManager* visManager = new G4VisExecutive;
| ^~~~~~~~~~~~
| G4VPCManager
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:89:19: error: ‘visManager’ was not declared in this scope; did you mean ‘scManager’?
89 | G4VisManager* visManager = new G4VisExecutive;
| ^~~~~~~~~~
| scManager
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:89:36: error: ‘G4VisExecutive’ does not name a type
89 | G4VisManager* visManager = new G4VisExecutive;
| ^~~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:92:5: error: ‘G4UIExecutive’ was not declared in this scope
92 | G4UIExecutive* UI = new G4UIExecutive(argc, argv);
| ^~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:92:20: error: ‘UI’ was not declared in this scope
92 | G4UIExecutive* UI = new G4UIExecutive(argc, argv);
| ^~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:92:29: error: expected type-specifier before ‘G4UIExecutive’
92 | G4UIExecutive* UI = new G4UIExecutive(argc, argv);
| ^~~~~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:96:12: error: type ‘<type error>’ argument given to ‘delete’, expected pointer
96 | delete visManager;
| ^~~~~~~~~~
/home/lucas/build/native/DAGMC-moab-5.1.0/src/src/geant4/app/exampleN01.cc:97:12: error: type ‘<type error>’ argument given to ‘delete’, expected pointer
97 | delete UI;
| ^~
``` | defect | dagmc not compatible with there appear to have been some changes in that break our ability to build dag applications stuff like this appears during build home lucas build native dagmc moab src src app cc in function ‘int main int char ’ home lucas build native dagmc moab src src app cc error ‘ ’ was not declared in this scope ui new argc argv tcsh home lucas build native dagmc moab src src app cc error ‘ui’ was not declared in this scope ui new argc argv tcsh home lucas build native dagmc moab src src app cc error expected type specifier before ‘ ’ ui new argc argv tcsh home lucas build native dagmc moab src src app cc error type ‘ ’ argument given to ‘delete’ expected pointer delete ui home lucas build native dagmc moab src src app cc error ‘ ’ was not declared in this scope did you mean ‘ ’ vismanager new home lucas build native dagmc moab src src app cc error ‘vismanager’ was not declared in this scope did you mean ‘scmanager’ vismanager new scmanager home lucas build native dagmc moab src src app cc error ‘ ’ does not name a type vismanager new home lucas build native dagmc moab src src app cc error ‘ ’ was not declared in this scope ui new argc argv home lucas build native dagmc moab src src app cc error ‘ui’ was not declared in this scope ui new argc argv home lucas build native dagmc moab src src app cc error expected type specifier before ‘ ’ ui new argc argv home lucas build native dagmc moab src src app cc error type ‘ ’ argument given to ‘delete’ expected pointer delete vismanager home lucas build native dagmc moab src src app cc error type ‘ ’ argument given to ‘delete’ expected pointer delete ui | 1 |
6,530 | 5,510,682,447 | IssuesEvent | 2017-03-17 00:56:38 | google/blockly | https://api.github.com/repos/google/blockly | opened | Performance regression when dragging blocks out of the flyout | bug performance | tldr: composite layers got quite expensive (~.2ms -> 4ms) when dragging a block out of the flyout.
This is present on develop. Marking as a bug for now since it is a big enough regression that I'd like to figure it out before we push to develop.
See https://github.com/LLK/scratch-blocks/issues/835 for more context.
| True | Performance regression when dragging blocks out of the flyout - tldr: composite layers got quite expensive (~.2ms -> 4ms) when dragging a block out of the flyout.
This is present on develop. Marking as a bug for now since it is a big enough regression that I'd like to figure it out before we push to develop.
See https://github.com/LLK/scratch-blocks/issues/835 for more context.
| non_defect | performance regression when dragging blocks out of the flyout tldr composite layers got quite expensive when dragging a block out of the flyout this is present on develop marking as a bug for now since it is a big enough regression that i d like to figure it out before we push to develop see for more context | 0 |
30,993 | 4,229,547,983 | IssuesEvent | 2016-07-04 08:18:02 | pythonapis/6ZJYP2PXGY5CWP2LWTZZFRIL | https://api.github.com/repos/pythonapis/6ZJYP2PXGY5CWP2LWTZZFRIL | closed | RncyRD2ffPklGRc42o56iYPDqv5FntO8jmphmjGt5EXom3ENuPHI1jFe3iEuzuYiNvDdB7XCdifkK+n03BjdHMBuCstR1TxSv28XKwmgfvcvCigCXLNS5AMszCmULaUEDVqf8WZ3oslKah/YeToEAUdleELccXhrvunsofMtTUA= | design | uX8P4Brikz9ZTgzQvv7KWic/PeU4YUuwYDQAU1KVLCV8X6PLfx+Fud6fitUXyYHy6w8tDqycKRnE4SAq3dZ7FM/J4ucL0T7/3PalaLN448NebYUgGaYdr1x/wVsOE7RI+ZDaoPn5ynro6IO9m3RuNWt8PAF9jTeM3xCEonyc8vpXhALMpQTMfYGceW/xnXXQSSK+rAxdAzuJiibBSY9Vp1A7KQZgMZ2G/Xe6DggnCa2zvRd7Kj2JM/vWK+EgABCXKpJpa1qLJEnc0j2ARjKVpXqQLHh1dcPclDxN4vknUSIZ0pybiClATrnL4n0SiOtWgLdvk1o0b0rAi6h/noiwDzQ/Kn1eD6sY055DDRjixvBQnFdswGymvL+3wASfGSpEWU/fzY0pWt8rkbpHQSRUgGoAmpk4QFENblKoWu08clcuWOyQ7MZn5vPX33Ez/W2/QuPp4K/jCUf7JYvSWxf63Zq6eQyQkC/RGEMR85z2usIIolQVIN9OpjcIDdM64kvxh7GAybXvBqvKhmbfPWBIC9Q/6zWw45VGVN2JEz/80hkMpZ0/ezia843WgLoeA47CUySWEHyv+vzxsBgz3owMi6ScTeBXWvCRZx6c1Eas75FrwSn0MUrJekyZL4oZlf9qBUjLtTDJ2rwopEgjRAfdYkrv80ICSA/FjXS0aBqIK0TL5g1tVmTpcPz0yxs4jLMvURPNxekVXwsOfiaM/saIijYiaD+FqHfowZrcjPyphcW7zD0WSh/5855O3E6h9Nq8 | 1.0 | RncyRD2ffPklGRc42o56iYPDqv5FntO8jmphmjGt5EXom3ENuPHI1jFe3iEuzuYiNvDdB7XCdifkK+n03BjdHMBuCstR1TxSv28XKwmgfvcvCigCXLNS5AMszCmULaUEDVqf8WZ3oslKah/YeToEAUdleELccXhrvunsofMtTUA= - uX8P4Brikz9ZTgzQvv7KWic/PeU4YUuwYDQAU1KVLCV8X6PLfx+Fud6fitUXyYHy6w8tDqycKRnE4SAq3dZ7FM/J4ucL0T7/3PalaLN448NebYUgGaYdr1x/wVsOE7RI+ZDaoPn5ynro6IO9m3RuNWt8PAF9jTeM3xCEonyc8vpXhALMpQTMfYGceW/xnXXQSSK+rAxdAzuJiibBSY9Vp1A7KQZgMZ2G/Xe6DggnCa2zvRd7Kj2JM/vWK+EgABCXKpJpa1qLJEnc0j2ARjKVpXqQLHh1dcPclDxN4vknUSIZ0pybiClATrnL4n0SiOtWgLdvk1o0b0rAi6h/noiwDzQ/Kn1eD6sY055DDRjixvBQnFdswGymvL+3wASfGSpEWU/fzY0pWt8rkbpHQSRUgGoAmpk4QFENblKoWu08clcuWOyQ7MZn5vPX33Ez/W2/QuPp4K/jCUf7JYvSWxf63Zq6eQyQkC/RGEMR85z2usIIolQVIN9OpjcIDdM64kvxh7GAybXvBqvKhmbfPWBIC9Q/6zWw45VGVN2JEz/80hkMpZ0/ezia843WgLoeA47CUySWEHyv+vzxsBgz3owMi6ScTeBXWvCRZx6c1Eas75FrwSn0MUrJekyZL4oZlf9qBUjLtTDJ2rwopEgjRAfdYkrv80ICSA/FjXS0aBqIK0TL5g1tVmTpcPz0yxs4jLMvURPNxekVXwsOfiaM/saIijYiaD+FqHfowZrcjPyphcW7zD0WSh/5855O3E6h9Nq8 | non_defect | yetoeaudleelccxhrvunsofmttua xnxxqssk vwk noiwdzq saiijyiad | 0 |
5,665 | 2,610,192,927 | IssuesEvent | 2015-02-26 19:00:56 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 纠结脸上色斑怎么去小窍门 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
容貌对一个人是至关重要,如果脸上长黄褐斑是所有人所困��
�的问题。随着社会的发展,人们生活水平的提高,人们的压�
��就会越来越大,经常熬夜加班,饮食不规律,精神紧张,就
会在脸上长出很多黄褐斑,色斑形成的原因这个问题让很多��
�都很头疼。脸上色斑怎么去小窍门,
《客户案例》
什么去黄褐斑最好,
在我的生活当中,要问我对自己最不满意的地方是什么,想��
�,就是我的一脸黄褐斑了。我的皮肤属于比较白的那种女生�
��在很小的时候,鼻梁就有一点黄褐斑,但那时斑是若隐若现
,不多也不明显。后来毕业工作了,因为做的是经常对着电��
�的文员工作,电脑辐射本身对皮肤伤害很大,我平常又没有�
��别注意清洁皮肤,再加上日常饮食不注意,时常还熬夜看书
……种种原因,我的黄褐斑加重了,鼻梁上的斑也开始扩散��
�面部。平常喜欢照镜子穿衣打扮的我,开始害怕仔细看镜子�
��的自己。爱美的我,一想到自己脸上布满黄褐斑的情景,就
不寒而栗。</br>
可以说我为了祛斑真的费了不少的功夫,就在我在经济��
�身心都遭受了很多打击之后,一个网友给我介绍天然精华黛�
��薇尔。我在黛芙薇尔的官网上咨询了专家,专家告诉我,先
天性的黄褐斑也可以治愈,此刻我仿佛找到了救命的稻草。��
�客服人员的指导下,我购买了两个周期的黛芙薇尔。</br>
收到黛芙薇尔在我使用了有15天左右,脸上的斑就开始慢
慢的淡化了,我是又惊又喜,一蹦一跳的,感觉自己那会儿��
�像小孩子偷使用了蜂蜜一样,,现在脸上的斑已经没有了,�
��谢谢黛芙薇尔了。专家们告诉我平时还应该注意。保持情绪
的稳定,防止日晒,睡眠要充足等。第一盒使用完效果真的��
�不是有太大的变化,但是,为了自己脸上的斑能去掉,还是�
��专家们的意见吧,前后共使用了两个周期的。没想到两个周
期后我脸上的黄褐斑就真的没有了。而且皮肤也变的好很多��
�我整天照着镜子,心里美滋滋的问着自己这是我吗?对,这是
我。再也不是那个满脸斑点的“斑长”了,谢谢黛芙薇尔。
阅读了脸上色斑怎么去小窍门,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
脸上色斑怎么去小窍门,同时为您分享祛斑小方法
果酸:生化制剂,低浓度(5-10%)时,对皮肤有减少角朊细胞
粘连,能有效地渗透皮肤,使堆积在皮肤上的角质层脱落,��
�肤表面显得光泽、亮丽,可清除毛囊口堵塞的角化物,使皮�
��分泌物能通畅地向外排泄。果酸还具有细胞再生性、保湿性
,以及具有改善皮肤质地的作用。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:36 | 1.0 | 纠结脸上色斑怎么去小窍门 - ```
《摘要》
容貌对一个人是至关重要,如果脸上长黄褐斑是所有人所困��
�的问题。随着社会的发展,人们生活水平的提高,人们的压�
��就会越来越大,经常熬夜加班,饮食不规律,精神紧张,就
会在脸上长出很多黄褐斑,色斑形成的原因这个问题让很多��
�都很头疼。脸上色斑怎么去小窍门,
《客户案例》
什么去黄褐斑最好,
在我的生活当中,要问我对自己最不满意的地方是什么,想��
�,就是我的一脸黄褐斑了。我的皮肤属于比较白的那种女生�
��在很小的时候,鼻梁就有一点黄褐斑,但那时斑是若隐若现
,不多也不明显。后来毕业工作了,因为做的是经常对着电��
�的文员工作,电脑辐射本身对皮肤伤害很大,我平常又没有�
��别注意清洁皮肤,再加上日常饮食不注意,时常还熬夜看书
……种种原因,我的黄褐斑加重了,鼻梁上的斑也开始扩散��
�面部。平常喜欢照镜子穿衣打扮的我,开始害怕仔细看镜子�
��的自己。爱美的我,一想到自己脸上布满黄褐斑的情景,就
不寒而栗。</br>
可以说我为了祛斑真的费了不少的功夫,就在我在经济��
�身心都遭受了很多打击之后,一个网友给我介绍天然精华黛�
��薇尔。我在黛芙薇尔的官网上咨询了专家,专家告诉我,先
天性的黄褐斑也可以治愈,此刻我仿佛找到了救命的稻草。��
�客服人员的指导下,我购买了两个周期的黛芙薇尔。</br>
收到黛芙薇尔在我使用了有15天左右,脸上的斑就开始慢
慢的淡化了,我是又惊又喜,一蹦一跳的,感觉自己那会儿��
�像小孩子偷使用了蜂蜜一样,,现在脸上的斑已经没有了,�
��谢谢黛芙薇尔了。专家们告诉我平时还应该注意。保持情绪
的稳定,防止日晒,睡眠要充足等。第一盒使用完效果真的��
�不是有太大的变化,但是,为了自己脸上的斑能去掉,还是�
��专家们的意见吧,前后共使用了两个周期的。没想到两个周
期后我脸上的黄褐斑就真的没有了。而且皮肤也变的好很多��
�我整天照着镜子,心里美滋滋的问着自己这是我吗?对,这是
我。再也不是那个满脸斑点的“斑长”了,谢谢黛芙薇尔。
阅读了脸上色斑怎么去小窍门,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
脸上色斑怎么去小窍门,同时为您分享祛斑小方法
果酸:生化制剂,低浓度(5-10%)时,对皮肤有减少角朊细胞
粘连,能有效地渗透皮肤,使堆积在皮肤上的角质层脱落,��
�肤表面显得光泽、亮丽,可清除毛囊口堵塞的角化物,使皮�
��分泌物能通畅地向外排泄。果酸还具有细胞再生性、保湿性
,以及具有改善皮肤质地的作用。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:36 | defect | 纠结脸上色斑怎么去小窍门 《摘要》 容貌对一个人是至关重要,如果脸上长黄褐斑是所有人所困�� �的问题。随着社会的发展,人们生活水平的提高,人们的压� ��就会越来越大,经常熬夜加班,饮食不规律,精神紧张,就 会在脸上长出很多黄褐斑,色斑形成的原因这个问题让很多�� �都很头疼。脸上色斑怎么去小窍门, 《客户案例》 什么去黄褐斑最好 在我的生活当中,要问我对自己最不满意的地方是什么,想�� �,就是我的一脸黄褐斑了。我的皮肤属于比较白的那种女生� ��在很小的时候,鼻梁就有一点黄褐斑,但那时斑是若隐若现 ,不多也不明显。后来毕业工作了,因为做的是经常对着电�� �的文员工作,电脑辐射本身对皮肤伤害很大,我平常又没有� ��别注意清洁皮肤,再加上日常饮食不注意,时常还熬夜看书 ……种种原因,我的黄褐斑加重了,鼻梁上的斑也开始扩散�� �面部。平常喜欢照镜子穿衣打扮的我,开始害怕仔细看镜子� ��的自己。爱美的我,一想到自己脸上布满黄褐斑的情景,就 不寒而栗。 可以说我为了祛斑真的费了不少的功夫,就在我在经济�� �身心都遭受了很多打击之后,一个网友给我介绍天然精华黛� ��薇尔。我在黛芙薇尔的官网上咨询了专家,专家告诉我,先 天性的黄褐斑也可以治愈,此刻我仿佛找到了救命的稻草。�� �客服人员的指导下,我购买了两个周期的黛芙薇尔。 ,脸上的斑就开始慢 慢的淡化了,我是又惊又喜,一蹦一跳的,感觉自己那会儿�� �像小孩子偷使用了蜂蜜一样,,现在脸上的斑已经没有了,� ��谢谢黛芙薇尔了。专家们告诉我平时还应该注意。保持情绪 的稳定,防止日晒,睡眠要充足等。第一盒使用完效果真的�� �不是有太大的变化,但是,为了自己脸上的斑能去掉,还是� ��专家们的意见吧,前后共使用了两个周期的。没想到两个周 期后我脸上的黄褐斑就真的没有了。而且皮肤也变的好很多�� �我整天照着镜子,心里美滋滋的问着自己这是我吗 对,这是 我。再也不是那个满脸斑点的“斑长”了,谢谢黛芙薇尔。 阅读了脸上色斑怎么去小窍门,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上色斑怎么去小窍门,同时为您分享祛斑小方法 果酸:生化制剂,低浓度( )时,对皮肤有减少角朊细胞 粘连,能有效地渗透皮肤,使堆积在皮肤上的角质层脱落,�� �肤表面显得光泽、亮丽,可清除毛囊口堵塞的角化物,使皮� ��分泌物能通畅地向外排泄。果酸还具有细胞再生性、保湿性 ,以及具有改善皮肤质地的作用。 original issue reported on code google com by additive gmail com on jul at | 1 |
17,251 | 2,986,836,759 | IssuesEvent | 2015-07-20 08:12:23 | cultibox/cultibox | https://api.github.com/repos/cultibox/cultibox | closed | [LOGS] Gestion des accents dans highchart | Component-cultiboxUI Priority-Low Type-Defect Type-ToTest | ```
Lorsqu'on fait un export des courbes, les légendes avec des accents
apparaissent au format HTML.
```
Original issue reported on code.google.com by `alliaume...@gmail.com` on 10 Feb 2015 at 9:01 | 1.0 | [LOGS] Gestion des accents dans highchart - ```
Lorsqu'on fait un export des courbes, les légendes avec des accents
apparaissent au format HTML.
```
Original issue reported on code.google.com by `alliaume...@gmail.com` on 10 Feb 2015 at 9:01 | defect | gestion des accents dans highchart lorsqu on fait un export des courbes les légendes avec des accents apparaissent au format html original issue reported on code google com by alliaume gmail com on feb at | 1 |
37,921 | 8,579,442,656 | IssuesEvent | 2018-11-13 09:11:42 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Review Fields.indexOf() and similar methods for performance | C: Functionality E: All Editions P: Medium T: Defect | More details will follow
----
See also:
https://groups.google.com/forum/#!topic/jooq-user/5Dt-gN68158 | 1.0 | Review Fields.indexOf() and similar methods for performance - More details will follow
----
See also:
https://groups.google.com/forum/#!topic/jooq-user/5Dt-gN68158 | defect | review fields indexof and similar methods for performance more details will follow see also | 1 |
158,746 | 13,746,599,296 | IssuesEvent | 2020-10-06 06:06:13 | msandfor/10-Easy-Steps | https://api.github.com/repos/msandfor/10-Easy-Steps | closed | [Hacktoberfest]: Add "Open Source Summit" to the Conferences Section | beginner beginner-friendly documentation :memo: first-contribution good first issue hacktoberfest :children_crossing: help wanted :hand: up-for-grabs | 🆕🐥☝ Beginners Only:
This issue is reserved for people who are new to Open Source. We know that the process of creating a pull request is the biggest barrier for new contributors. This issue is for you 💝
To do:
* Add *Open Source Summit* to the Conferences Section
* See below for link to step-by-step tutorial
* Go down to Conferences section - it starts at around line 173
* Using the following template, put in the correct info for that conference (Search for it on the internet to find the correct details) underneath the last row in the table, ensuring there is still an empty row before the next table.
`| [Conference Name](Website Address) | Conference Name or Organising Org | 10 | Description of the Conferences, found on their website |`
* Add your details in the contributors section at the bottom of the page using the same formatting as the last one
📋 *Step by Step*
[Step-By-Step Tutorial](https://github.com/msandfor/10-Easy-Steps/blob/master/docs/GitHub-browser-tutorial.md)
Remember to observe the [Code of Conduct](https://github.com/msandfor/10-Easy-Steps/blob/master/code_of_conduct.md) - by which we mean, in short - Be Kind, Be Positive, Be Helpful
| 1.0 | [Hacktoberfest]: Add "Open Source Summit" to the Conferences Section - 🆕🐥☝ Beginners Only:
This issue is reserved for people who are new to Open Source. We know that the process of creating a pull request is the biggest barrier for new contributors. This issue is for you 💝
To do:
* Add *Open Source Summit* to the Conferences Section
* See below for link to step-by-step tutorial
* Go down to Conferences section - it starts at around line 173
* Using the following template, put in the correct info for that conference (Search for it on the internet to find the correct details) underneath the last row in the table, ensuring there is still an empty row before the next table.
`| [Conference Name](Website Address) | Conference Name or Organising Org | 10 | Description of the Conferences, found on their website |`
* Add your details in the contributors section at the bottom of the page using the same formatting as the last one
📋 *Step by Step*
[Step-By-Step Tutorial](https://github.com/msandfor/10-Easy-Steps/blob/master/docs/GitHub-browser-tutorial.md)
Remember to observe the [Code of Conduct](https://github.com/msandfor/10-Easy-Steps/blob/master/code_of_conduct.md) - by which we mean, in short - Be Kind, Be Positive, Be Helpful
| non_defect | add open source summit to the conferences section 🆕🐥☝ beginners only this issue is reserved for people who are new to open source we know that the process of creating a pull request is the biggest barrier for new contributors this issue is for you 💝 to do add open source summit to the conferences section see below for link to step by step tutorial go down to conferences section it starts at around line using the following template put in the correct info for that conference search for it on the internet to find the correct details underneath the last row in the table ensuring there is still an empty row before the next table website address conference name or organising org description of the conferences found on their website add your details in the contributors section at the bottom of the page using the same formatting as the last one 📋 step by step remember to observe the by which we mean in short be kind be positive be helpful | 0 |
67,341 | 20,961,606,243 | IssuesEvent | 2022-03-27 21:48:17 | abedmaatalla/sipdroid | https://api.github.com/repos/abedmaatalla/sipdroid | closed | TLS support in v 3.4beta | Priority-Medium Type-Defect auto-migrated | ```
looking at code of SipDroid v 3.4 beta for tls support, it looks like tls is
kept false at Initialization of the SipProvider.Later at sending message method
, tcp return conn id, udp return null, while using tls return "Unsupported
Protocol (tls), Message discarded". Does that means tls is still not supported?
As enabling tls protocol it still uses tcp and enabling tls in the code the
Register request is not going.
```
Original issue reported on code.google.com by `sendimpo...@gmail.com` on 27 Jan 2014 at 5:10
| 1.0 | TLS support in v 3.4beta - ```
looking at code of SipDroid v 3.4 beta for tls support, it looks like tls is
kept false at Initialization of the SipProvider.Later at sending message method
, tcp return conn id, udp return null, while using tls return "Unsupported
Protocol (tls), Message discarded". Does that means tls is still not supported?
As enabling tls protocol it still uses tcp and enabling tls in the code the
Register request is not going.
```
Original issue reported on code.google.com by `sendimpo...@gmail.com` on 27 Jan 2014 at 5:10
| defect | tls support in v looking at code of sipdroid v beta for tls support it looks like tls is kept false at initialization of the sipprovider later at sending message method tcp return conn id udp return null while using tls return unsupported protocol tls message discarded does that means tls is still not supported as enabling tls protocol it still uses tcp and enabling tls in the code the register request is not going original issue reported on code google com by sendimpo gmail com on jan at | 1 |
32,826 | 6,953,397,141 | IssuesEvent | 2017-12-06 20:52:47 | Dzhuneyt/jquery-tubular | https://api.github.com/repos/Dzhuneyt/jquery-tubular | closed | In browser SAFARI on IMAC | auto-migrated Priority-Medium Type-Defect | ```
Video dont go PLAY
```
Original issue reported on code.google.com by `nando.pi...@gmail.com` on 26 Nov 2013 at 1:28
| 1.0 | In browser SAFARI on IMAC - ```
Video dont go PLAY
```
Original issue reported on code.google.com by `nando.pi...@gmail.com` on 26 Nov 2013 at 1:28
| defect | in browser safari on imac video dont go play original issue reported on code google com by nando pi gmail com on nov at | 1 |
434,188 | 30,445,934,359 | IssuesEvent | 2023-07-15 17:09:50 | helpwave/services | https://api.github.com/repos/helpwave/services | closed | Document everything | documentation priority: low | Currently there are a lot of moving things, but when the storm has calmed we should document everything to make onboarding easier. | 1.0 | Document everything - Currently there are a lot of moving things, but when the storm has calmed we should document everything to make onboarding easier. | non_defect | document everything currently there are a lot of moving things but when the storm has calmed we should document everything to make onboarding easier | 0 |
273,857 | 23,790,084,478 | IssuesEvent | 2022-09-02 13:48:47 | cosmos/cosmos-sdk | https://api.github.com/repos/cosmos/cosmos-sdk | closed | Improve testing DevX | T: Tests tooling T: Dev UX | ## Summary
As a developer I want an easy way to run and manage tests. Moreover, I want to have a clear separation of unit tests, integration tests and functional tests. Finally, unit tests must be very quick to run and don't involve lot of dependencies.
### NOTE
* This is a **meta-issue**.
* If you have a quick question - please make a comment.
* Feel free to add a proposal (prefer to do it through a comment)
* Don't discuss details - this we can do in a sub-issue.
## Problem Definition
Current setup is far from good (in terms of DevX):
+ `make test-unit` takes loooot of time, and it's more than unit tests
+ We should be able to compile everything (currently we need to be careful to remember to put all build flags to test-unit job).
+ We don't have single test command to build and tests everything (ALL files).
## Proposal
- [ ] Add a make job to compile all tests without running them. #7364, ...
- [x] Make sure tests are cached (currently they are not if we use `make`) #7314
- [ ] Use `testing.T.Skip` instead of build flags whenever possible to skip some tests.
- [ ] Make sure unit tests are proper unit tests.
- [ ] Use test suites more extensively (this will also help with tests separation).
- [ ] Improve test output summary (@alessio working on it)
- [ ] Limit amount of logs in tests https://github.com/cosmos/cosmos-sdk/issues/7257
____
#### For Admin Use
- [x] Not duplicate issue
- [x] Appropriate labels applied
- [ ] Appropriate contributors tagged
- [ ] Contributor assigned/self-assigned
| 1.0 | Improve testing DevX - ## Summary
As a developer I want an easy way to run and manage tests. Moreover, I want to have a clear separation of unit tests, integration tests and functional tests. Finally, unit tests must be very quick to run and don't involve lot of dependencies.
### NOTE
* This is a **meta-issue**.
* If you have a quick question - please make a comment.
* Feel free to add a proposal (prefer to do it through a comment)
* Don't discuss details - this we can do in a sub-issue.
## Problem Definition
Current setup is far from good (in terms of DevX):
+ `make test-unit` takes loooot of time, and it's more than unit tests
+ We should be able to compile everything (currently we need to be careful to remember to put all build flags to test-unit job).
+ We don't have single test command to build and tests everything (ALL files).
## Proposal
- [ ] Add a make job to compile all tests without running them. #7364, ...
- [x] Make sure tests are cached (currently they are not if we use `make`) #7314
- [ ] Use `testing.T.Skip` instead of build flags whenever possible to skip some tests.
- [ ] Make sure unit tests are proper unit tests.
- [ ] Use test suites more extensively (this will also help with tests separation).
- [ ] Improve test output summary (@alessio working on it)
- [ ] Limit amount of logs in tests https://github.com/cosmos/cosmos-sdk/issues/7257
____
#### For Admin Use
- [x] Not duplicate issue
- [x] Appropriate labels applied
- [ ] Appropriate contributors tagged
- [ ] Contributor assigned/self-assigned
| non_defect | improve testing devx summary as a developer i want an easy way to run and manage tests moreover i want to have a clear separation of unit tests integration tests and functional tests finally unit tests must be very quick to run and don t involve lot of dependencies note this is a meta issue if you have a quick question please make a comment feel free to add a proposal prefer to do it through a comment don t discuss details this we can do in a sub issue problem definition current setup is far from good in terms of devx make test unit takes loooot of time and it s more than unit tests we should be able to compile everything currently we need to be careful to remember to put all build flags to test unit job we don t have single test command to build and tests everything all files proposal add a make job to compile all tests without running them make sure tests are cached currently they are not if we use make use testing t skip instead of build flags whenever possible to skip some tests make sure unit tests are proper unit tests use test suites more extensively this will also help with tests separation improve test output summary alessio working on it limit amount of logs in tests for admin use not duplicate issue appropriate labels applied appropriate contributors tagged contributor assigned self assigned | 0 |
424,348 | 29,048,943,270 | IssuesEvent | 2023-05-13 23:47:42 | janderland/fdbq | https://api.github.com/repos/janderland/fdbq | closed | Revise the readme for initial release. | documentation | - [x] Differences between lang, API, & app (including versioning).
- [x] Syntax explanation.
- [x] How to build & run tests.
- [x] #151
- [ ] Visitor pattern.
- [x] Value serialization. | 1.0 | Revise the readme for initial release. - - [x] Differences between lang, API, & app (including versioning).
- [x] Syntax explanation.
- [x] How to build & run tests.
- [x] #151
- [ ] Visitor pattern.
- [x] Value serialization. | non_defect | revise the readme for initial release differences between lang api app including versioning syntax explanation how to build run tests visitor pattern value serialization | 0 |
76,225 | 15,495,881,453 | IssuesEvent | 2021-03-11 01:39:56 | DODHI5/react-kanban | https://api.github.com/repos/DODHI5/react-kanban | opened | CVE-2018-6342 (High) detected in react-dev-utils-5.0.0.tgz | security vulnerability | ## CVE-2018-6342 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>react-dev-utils-5.0.0.tgz</b></p></summary>
<p>Webpack utilities used by Create React App</p>
<p>Library home page: <a href="https://registry.npmjs.org/react-dev-utils/-/react-dev-utils-5.0.0.tgz">https://registry.npmjs.org/react-dev-utils/-/react-dev-utils-5.0.0.tgz</a></p>
<p>Path to dependency file: /react-kanban/package.json</p>
<p>Path to vulnerable library: react-kanban/node_modules/react-dev-utils/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.1.1.tgz (Root Library)
- :x: **react-dev-utils-5.0.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
react-dev-utils on Windows allows developers to run a local webserver for accepting various commands, including a command to launch an editor. The input to that command was not properly sanitized, allowing an attacker who can make a network request to the server (either via CSRF or by direct request) to execute arbitrary commands on the targeted system. This issue affects multiple branches: 1.x.x prior to 1.0.4, 2.x.x prior to 2.0.2, 3.x.x prior to 3.1.2, 4.x.x prior to 4.2.2, and 5.x.x prior to 5.0.2.
<p>Publish Date: 2018-12-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-6342>CVE-2018-6342</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-6342">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-6342</a></p>
<p>Release Date: 2018-12-31</p>
<p>Fix Resolution: 1.0.4,2.0.2,3.1.2,4.2.2,5.0.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-6342 (High) detected in react-dev-utils-5.0.0.tgz - ## CVE-2018-6342 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>react-dev-utils-5.0.0.tgz</b></p></summary>
<p>Webpack utilities used by Create React App</p>
<p>Library home page: <a href="https://registry.npmjs.org/react-dev-utils/-/react-dev-utils-5.0.0.tgz">https://registry.npmjs.org/react-dev-utils/-/react-dev-utils-5.0.0.tgz</a></p>
<p>Path to dependency file: /react-kanban/package.json</p>
<p>Path to vulnerable library: react-kanban/node_modules/react-dev-utils/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.1.1.tgz (Root Library)
- :x: **react-dev-utils-5.0.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
react-dev-utils on Windows allows developers to run a local webserver for accepting various commands, including a command to launch an editor. The input to that command was not properly sanitized, allowing an attacker who can make a network request to the server (either via CSRF or by direct request) to execute arbitrary commands on the targeted system. This issue affects multiple branches: 1.x.x prior to 1.0.4, 2.x.x prior to 2.0.2, 3.x.x prior to 3.1.2, 4.x.x prior to 4.2.2, and 5.x.x prior to 5.0.2.
<p>Publish Date: 2018-12-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-6342>CVE-2018-6342</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-6342">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-6342</a></p>
<p>Release Date: 2018-12-31</p>
<p>Fix Resolution: 1.0.4,2.0.2,3.1.2,4.2.2,5.0.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in react dev utils tgz cve high severity vulnerability vulnerable library react dev utils tgz webpack utilities used by create react app library home page a href path to dependency file react kanban package json path to vulnerable library react kanban node modules react dev utils package json dependency hierarchy react scripts tgz root library x react dev utils tgz vulnerable library vulnerability details react dev utils on windows allows developers to run a local webserver for accepting various commands including a command to launch an editor the input to that command was not properly sanitized allowing an attacker who can make a network request to the server either via csrf or by direct request to execute arbitrary commands on the targeted system this issue affects multiple branches x x prior to x x prior to x x prior to x x prior to and x x prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
5,526 | 2,610,189,304 | IssuesEvent | 2015-02-26 18:59:56 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 关于如何能去除色斑 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
愉快跟不愉快的回忆,比如一个硬币的两面,存在于我们的��
�一段情感里。就像那个著名的“蝴蝶效应”,如果你经常记�
��不愉快的人、不愉快的事,生活就跟着变得不愉快起来。相
反,有些女人却能在跟老公吵架的时候及其她求婚时的表情��
�他怀抱的温暖。这里的“吵”是一种乐观、积极的沟通方式�
��这样的女人即便是面临命运得不测风云,也不会唉声叹气,
而当它是动力。面带微笑、坦然自处,男人有乐观女人的相��
�,一生都将阳光灿烂。如何能去除色斑,
《客户案例》
小兰<br>
我以前的皮肤挺白皙细腻的,由于常年在外打拼,生活��
�律也不协调,近年来脸上长出了很多的色斑,颈部也有很多�
��斑点大小不一,呈深褐色,并随季节变化,冬天浅,夏天深
。老公都说我现在像老太太了,这句话刺激了我,所以四处��
�找祛斑产品,发誓一定要祛斑成功。<br>
一年多以来,我试了无数种祛斑产品,尝试了形形色色��
�祛斑方法,可那斑丝毫没有改变。了精华,做过二氧化碳激�
��,斑点不但没变小,反而变大变深了,实在让我无可奈何。
后来,一个朋友推荐我买「黛芙薇尔精华液」,她说她一个��
�友就是用这个去掉色斑的,效果挺不错的。所以我就打电话�
��专家咨询,专家告诉我色斑包括黄褐斑、黑斑、黄褐斑和老
年斑等,属色素障碍性皮肤病。该组疾病的发病机理不外乎��
�素细胞数量增多或其活性增强。必须从根源入手,方可治愈�
��专家还说了许多,很专业也很详细,我就订购了一个周期的
「黛芙薇尔精华液」。<br>
半个多月以后,我脸上的斑点就开始变淡,颜色不再那��
�深,从暗褐色变成了淡褐色。我就又订购了一个周期。两个�
��的样子,鼻子两侧和脸颊上一块块的色斑缩小了不少,颜色
更加淡,小块些的斑点就像花生米变芝麻,不但变小了,而��
�颜色也变浅,黑印的颜色慢慢褪下了那层脏脏污污,皮肤开�
��露出肉色。看到色斑淡化得不错,我接着订购了第三个周期
的套装,在使用套装过程中尽量避免日晒,脸上的色斑淡化��
�成绩不错,皮肤不会出现一块黑、一块暗褐色的,基本上的�
��点都变成了浅褐色,看起来一点都不明显,稍微打点粉底或
者隔离霜就看不出来了。<br>
看着没有那些斑斑点点的脸,我的心里那个开心啊。老��
�也不再说我像老太太了,是「黛芙薇尔精华液」让我的肌肤�
��复了年轻。
阅读了如何能去除色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
如何能去除色斑,同时为您分享祛斑小方法
去斑方法,西红柿汁:将西红柿榨汁,加入一汤匙甘油,每日2
到3次,每次敷10分钟,然后 清水洗净。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:42 | 1.0 | 关于如何能去除色斑 - ```
《摘要》
愉快跟不愉快的回忆,比如一个硬币的两面,存在于我们的��
�一段情感里。就像那个著名的“蝴蝶效应”,如果你经常记�
��不愉快的人、不愉快的事,生活就跟着变得不愉快起来。相
反,有些女人却能在跟老公吵架的时候及其她求婚时的表情��
�他怀抱的温暖。这里的“吵”是一种乐观、积极的沟通方式�
��这样的女人即便是面临命运得不测风云,也不会唉声叹气,
而当它是动力。面带微笑、坦然自处,男人有乐观女人的相��
�,一生都将阳光灿烂。如何能去除色斑,
《客户案例》
小兰<br>
我以前的皮肤挺白皙细腻的,由于常年在外打拼,生活��
�律也不协调,近年来脸上长出了很多的色斑,颈部也有很多�
��斑点大小不一,呈深褐色,并随季节变化,冬天浅,夏天深
。老公都说我现在像老太太了,这句话刺激了我,所以四处��
�找祛斑产品,发誓一定要祛斑成功。<br>
一年多以来,我试了无数种祛斑产品,尝试了形形色色��
�祛斑方法,可那斑丝毫没有改变。了精华,做过二氧化碳激�
��,斑点不但没变小,反而变大变深了,实在让我无可奈何。
后来,一个朋友推荐我买「黛芙薇尔精华液」,她说她一个��
�友就是用这个去掉色斑的,效果挺不错的。所以我就打电话�
��专家咨询,专家告诉我色斑包括黄褐斑、黑斑、黄褐斑和老
年斑等,属色素障碍性皮肤病。该组疾病的发病机理不外乎��
�素细胞数量增多或其活性增强。必须从根源入手,方可治愈�
��专家还说了许多,很专业也很详细,我就订购了一个周期的
「黛芙薇尔精华液」。<br>
半个多月以后,我脸上的斑点就开始变淡,颜色不再那��
�深,从暗褐色变成了淡褐色。我就又订购了一个周期。两个�
��的样子,鼻子两侧和脸颊上一块块的色斑缩小了不少,颜色
更加淡,小块些的斑点就像花生米变芝麻,不但变小了,而��
�颜色也变浅,黑印的颜色慢慢褪下了那层脏脏污污,皮肤开�
��露出肉色。看到色斑淡化得不错,我接着订购了第三个周期
的套装,在使用套装过程中尽量避免日晒,脸上的色斑淡化��
�成绩不错,皮肤不会出现一块黑、一块暗褐色的,基本上的�
��点都变成了浅褐色,看起来一点都不明显,稍微打点粉底或
者隔离霜就看不出来了。<br>
看着没有那些斑斑点点的脸,我的心里那个开心啊。老��
�也不再说我像老太太了,是「黛芙薇尔精华液」让我的肌肤�
��复了年轻。
阅读了如何能去除色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
如何能去除色斑,同时为您分享祛斑小方法
去斑方法,西红柿汁:将西红柿榨汁,加入一汤匙甘油,每日2
到3次,每次敷10分钟,然后 清水洗净。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:42 | defect | 关于如何能去除色斑 《摘要》 愉快跟不愉快的回忆,比如一个硬币的两面,存在于我们的�� �一段情感里。就像那个著名的“蝴蝶效应”,如果你经常记� ��不愉快的人、不愉快的事,生活就跟着变得不愉快起来。相 反,有些女人却能在跟老公吵架的时候及其她求婚时的表情�� �他怀抱的温暖。这里的“吵”是一种乐观、积极的沟通方式� ��这样的女人即便是面临命运得不测风云,也不会唉声叹气, 而当它是动力。面带微笑、坦然自处,男人有乐观女人的相�� �,一生都将阳光灿烂。如何能去除色斑, 《客户案例》 小兰 我以前的皮肤挺白皙细腻的,由于常年在外打拼,生活�� �律也不协调,近年来脸上长出了很多的色斑,颈部也有很多� ��斑点大小不一,呈深褐色,并随季节变化,冬天浅,夏天深 。老公都说我现在像老太太了,这句话刺激了我,所以四处�� �找祛斑产品,发誓一定要祛斑成功。 一年多以来,我试了无数种祛斑产品,尝试了形形色色�� �祛斑方法,可那斑丝毫没有改变。了精华,做过二氧化碳激� ��,斑点不但没变小,反而变大变深了,实在让我无可奈何。 后来,一个朋友推荐我买「黛芙薇尔精华液」,她说她一个�� �友就是用这个去掉色斑的,效果挺不错的。所以我就打电话� ��专家咨询,专家告诉我色斑包括黄褐斑、黑斑、黄褐斑和老 年斑等,属色素障碍性皮肤病。该组疾病的发病机理不外乎�� �素细胞数量增多或其活性增强。必须从根源入手,方可治愈� ��专家还说了许多,很专业也很详细,我就订购了一个周期的 「黛芙薇尔精华液」。 半个多月以后,我脸上的斑点就开始变淡,颜色不再那�� �深,从暗褐色变成了淡褐色。我就又订购了一个周期。两个� ��的样子,鼻子两侧和脸颊上一块块的色斑缩小了不少,颜色 更加淡,小块些的斑点就像花生米变芝麻,不但变小了,而�� �颜色也变浅,黑印的颜色慢慢褪下了那层脏脏污污,皮肤开� ��露出肉色。看到色斑淡化得不错,我接着订购了第三个周期 的套装,在使用套装过程中尽量避免日晒,脸上的色斑淡化�� �成绩不错,皮肤不会出现一块黑、一块暗褐色的,基本上的� ��点都变成了浅褐色,看起来一点都不明显,稍微打点粉底或 者隔离霜就看不出来了。 看着没有那些斑斑点点的脸,我的心里那个开心啊。老�� �也不再说我像老太太了,是「黛芙薇尔精华液」让我的肌肤� ��复了年轻。 阅读了如何能去除色斑,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 如何能去除色斑,同时为您分享祛斑小方法 去斑方法 西红柿汁:将西红柿榨汁,加入一汤匙甘油, , ,然后 清水洗净。 original issue reported on code google com by additive gmail com on jul at | 1 |
391,254 | 26,883,400,924 | IssuesEvent | 2023-02-05 22:22:19 | fga-eps-mds/2022-2-Alectrion-DOC | https://api.github.com/repos/fga-eps-mds/2022-2-Alectrion-DOC | closed | Atualização diagrama lógico | documentation ARQ EPS | **Descrição**
Atualiza diagrama lógico.
**Tarefas :**
- [ ] Atualização do diagrama lógico
| 1.0 | Atualização diagrama lógico - **Descrição**
Atualiza diagrama lógico.
**Tarefas :**
- [ ] Atualização do diagrama lógico
| non_defect | atualização diagrama lógico descrição atualiza diagrama lógico tarefas atualização do diagrama lógico | 0 |
13,191 | 2,736,573,396 | IssuesEvent | 2015-04-19 15:34:26 | tspence/csharp-csv-reader | https://api.github.com/repos/tspence/csharp-csv-reader | closed | What options are supported by this DLL? | auto-migrated Priority-Medium Type-Defect | ```
Does it handles Double Quotes, Single Quotes, and a column having large amount
of text which contains commas?
```
Original issue reported on code.google.com by `babji.su...@gmail.com` on 16 Oct 2013 at 5:10 | 1.0 | What options are supported by this DLL? - ```
Does it handles Double Quotes, Single Quotes, and a column having large amount
of text which contains commas?
```
Original issue reported on code.google.com by `babji.su...@gmail.com` on 16 Oct 2013 at 5:10 | defect | what options are supported by this dll does it handles double quotes single quotes and a column having large amount of text which contains commas original issue reported on code google com by babji su gmail com on oct at | 1 |
32,756 | 7,595,778,054 | IssuesEvent | 2018-04-27 07:15:54 | exercism/java | https://api.github.com/repos/exercism/java | closed | minesweeper: remove `final` from tests | code good first patch | According to our [policies](https://github.com/exercism/java/blob/master/POLICIES.md#avoid-using-final) we want to avoid using `final` in user facing code as it can be confusing to people new to Java.
Please remove all uses of `final` in the [minesweeper tests](https://github.com/exercism/java/blob/master/exercises/minesweeper/src/test/java/MinesweeperBoardTest.java). | 1.0 | minesweeper: remove `final` from tests - According to our [policies](https://github.com/exercism/java/blob/master/POLICIES.md#avoid-using-final) we want to avoid using `final` in user facing code as it can be confusing to people new to Java.
Please remove all uses of `final` in the [minesweeper tests](https://github.com/exercism/java/blob/master/exercises/minesweeper/src/test/java/MinesweeperBoardTest.java). | non_defect | minesweeper remove final from tests according to our we want to avoid using final in user facing code as it can be confusing to people new to java please remove all uses of final in the | 0 |
57,383 | 15,761,126,752 | IssuesEvent | 2021-03-31 09:40:14 | primefaces/primereact | https://api.github.com/repos/primefaces/primereact | closed | A maximizable dialog cannot be maximized properly after its size or position has been changed | defect | **I'm submitting a ...** (check one with "x")
```
[x] bug report
[ ] feature request
[ ] support request
```
**Codesandbox Case (Bug Reports)**
https://codesandbox.io/s/6jbh8
**Current behavior**
A maximizable dialog has a strange behavior after its size or position has been changed:
- if we change the size of the dialog and then try to maximize it, the maximize action changes just the width of the dialog to full width, but the height remains the same
- if we drag the dialog to a new position, triggering the maximize action will maximize the dialog but its position will not be aligned with the window but is shifted by the same amount as the non-maximized dialog was moved around. This sometimes leads to situations where part of the dialog is non-reachable/non-visible.
**Expected behavior**
To be able to have maximized the dialog even after moving it around or changing its size.
* **React version:**
17.0.1
* **PrimeReact version:**
6.2.1
* **Browser:**
all
* **Language:**
ES6/7 | ES5 | 1.0 | A maximizable dialog cannot be maximized properly after its size or position has been changed - **I'm submitting a ...** (check one with "x")
```
[x] bug report
[ ] feature request
[ ] support request
```
**Codesandbox Case (Bug Reports)**
https://codesandbox.io/s/6jbh8
**Current behavior**
A maximizable dialog has a strange behavior after its size or position has been changed:
- if we change the size of the dialog and then try to maximize it, the maximize action changes just the width of the dialog to full width, but the height remains the same
- if we drag the dialog to a new position, triggering the maximize action will maximize the dialog but its position will not be aligned with the window but is shifted by the same amount as the non-maximized dialog was moved around. This sometimes leads to situations where part of the dialog is non-reachable/non-visible.
**Expected behavior**
To be able to have maximized the dialog even after moving it around or changing its size.
* **React version:**
17.0.1
* **PrimeReact version:**
6.2.1
* **Browser:**
all
* **Language:**
ES6/7 | ES5 | defect | a maximizable dialog cannot be maximized properly after its size or position has been changed i m submitting a check one with x bug report feature request support request codesandbox case bug reports current behavior a maximizable dialog has a strange behavior after its size or position has been changed if we change the size of the dialog and then try to maximize it the maximize action changes just the width of the dialog to full width but the height remains the same if we drag the dialog to a new position triggering the maximize action will maximize the dialog but its position will not be aligned with the window but is shifted by the same amount as the non maximized dialog was moved around this sometimes leads to situations where part of the dialog is non reachable non visible expected behavior to be able to have maximized the dialog even after moving it around or changing its size react version primereact version browser all language | 1 |
413,893 | 27,972,831,926 | IssuesEvent | 2023-03-25 07:55:27 | Alfarezie/SI-44-08-KELOMPOK-269 | https://api.github.com/repos/Alfarezie/SI-44-08-KELOMPOK-269 | closed | [S1 - PBI1] Membuat halaman registrasi | documentation | Sebagai pasien, saya ingin melakukan pendaftaran pada halaman registrasi website Sigma sehingga akun saya akan terdaftar di database
(Activity diagram)

| 1.0 | [S1 - PBI1] Membuat halaman registrasi - Sebagai pasien, saya ingin melakukan pendaftaran pada halaman registrasi website Sigma sehingga akun saya akan terdaftar di database
(Activity diagram)

| non_defect | membuat halaman registrasi sebagai pasien saya ingin melakukan pendaftaran pada halaman registrasi website sigma sehingga akun saya akan terdaftar di database activity diagram | 0 |
50,257 | 13,187,404,186 | IssuesEvent | 2020-08-13 03:18:31 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | fix simulation so all bots report "green" (Trac #394) | Migrated from Trac combo simulation defect |
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/394
, reported by nega and owned by kjmeagher_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-05-30T16:43:18",
"description": "",
"reporter": "nega",
"cc": "blaufuss olivas",
"resolution": "fixed",
"_ts": "1338396198000000",
"component": "combo simulation",
"summary": "fix simulation so all bots report \"green\"",
"priority": "normal",
"keywords": "",
"time": "2012-05-17T14:20:40",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
| 1.0 | fix simulation so all bots report "green" (Trac #394) -
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/394
, reported by nega and owned by kjmeagher_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-05-30T16:43:18",
"description": "",
"reporter": "nega",
"cc": "blaufuss olivas",
"resolution": "fixed",
"_ts": "1338396198000000",
"component": "combo simulation",
"summary": "fix simulation so all bots report \"green\"",
"priority": "normal",
"keywords": "",
"time": "2012-05-17T14:20:40",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
| defect | fix simulation so all bots report green trac migrated from reported by nega and owned by kjmeagher json status closed changetime description reporter nega cc blaufuss olivas resolution fixed ts component combo simulation summary fix simulation so all bots report green priority normal keywords time milestone owner kjmeagher type defect | 1 |
18,852 | 3,089,697,396 | IssuesEvent | 2015-08-25 23:05:25 | google/googlemock | https://api.github.com/repos/google/googlemock | closed | Crash caused by missing lock | auto-migrated Priority-Medium Type-Defect | ```
FunctionMockerBase<F>::AddNewExpectation() modifies untyped_expectations_
without locking g_gmock_mutex.
It can cause a crash if another thread calls the same function or
FindMatchingExpectationLocked().
Google Mock version: 1.7.0
```
Original issue reported on code.google.com by `n...@tresorit.com` on 11 Nov 2013 at 6:49 | 1.0 | Crash caused by missing lock - ```
FunctionMockerBase<F>::AddNewExpectation() modifies untyped_expectations_
without locking g_gmock_mutex.
It can cause a crash if another thread calls the same function or
FindMatchingExpectationLocked().
Google Mock version: 1.7.0
```
Original issue reported on code.google.com by `n...@tresorit.com` on 11 Nov 2013 at 6:49 | defect | crash caused by missing lock functionmockerbase addnewexpectation modifies untyped expectations without locking g gmock mutex it can cause a crash if another thread calls the same function or findmatchingexpectationlocked google mock version original issue reported on code google com by n tresorit com on nov at | 1 |
47,966 | 13,067,338,531 | IssuesEvent | 2020-07-31 00:08:40 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | [genie-icetray] tests taking 1000+ minutes to run (Trac #1536) | Migrated from Trac combo simulation defect | The test process automatically quits after 1200 seconds w/o output. Tests should only take a few 10s of seconds at most.
```text
9088 ? R 983:28 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py
15646 ? R 914:11 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py
27741 ? R 299:10 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py
30551 ? R 1056:17 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py
```
Migrated from https://code.icecube.wisc.edu/ticket/1536
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:35",
"description": "The test process automatically quits after 1200 seconds w/o output. Tests should only take a few 10s of seconds at most.\n\n{{{\n 9088 ? R 983:28 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py\n15646 ? R 914:11 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py\n27741 ? R 299:10 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py\n30551 ? R 1056:17 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py\n}}}",
"reporter": "nega",
"cc": "kclark",
"resolution": "duplicate",
"_ts": "1550067215093672",
"component": "combo simulation",
"summary": "[genie-icetray] tests taking 1000+ minutes to run",
"priority": "normal",
"keywords": "",
"time": "2016-01-29T21:16:50",
"milestone": "",
"owner": "melanie.day",
"type": "defect"
}
```
| 1.0 | [genie-icetray] tests taking 1000+ minutes to run (Trac #1536) - The test process automatically quits after 1200 seconds w/o output. Tests should only take a few 10s of seconds at most.
```text
9088 ? R 983:28 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py
15646 ? R 914:11 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py
27741 ? R 299:10 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py
30551 ? R 1056:17 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py
```
Migrated from https://code.icecube.wisc.edu/ticket/1536
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:35",
"description": "The test process automatically quits after 1200 seconds w/o output. Tests should only take a few 10s of seconds at most.\n\n{{{\n 9088 ? R 983:28 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py\n15646 ? R 914:11 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py\n27741 ? R 299:10 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py\n30551 ? R 1056:17 python /build/buildslave/morax_cvmfs/Scientific_Linux_6__cvmfs_/source/genie-icetray/resources/test/GENIETest.py\n}}}",
"reporter": "nega",
"cc": "kclark",
"resolution": "duplicate",
"_ts": "1550067215093672",
"component": "combo simulation",
"summary": "[genie-icetray] tests taking 1000+ minutes to run",
"priority": "normal",
"keywords": "",
"time": "2016-01-29T21:16:50",
"milestone": "",
"owner": "melanie.day",
"type": "defect"
}
```
| defect | tests taking minutes to run trac the test process automatically quits after seconds w o output tests should only take a few of seconds at most text r python build buildslave morax cvmfs scientific linux cvmfs source genie icetray resources test genietest py r python build buildslave morax cvmfs scientific linux cvmfs source genie icetray resources test genietest py r python build buildslave morax cvmfs scientific linux cvmfs source genie icetray resources test genietest py r python build buildslave morax cvmfs scientific linux cvmfs source genie icetray resources test genietest py migrated from json status closed changetime description the test process automatically quits after seconds w o output tests should only take a few of seconds at most n n n r python build buildslave morax cvmfs scientific linux cvmfs source genie icetray resources test genietest py r python build buildslave morax cvmfs scientific linux cvmfs source genie icetray resources test genietest py r python build buildslave morax cvmfs scientific linux cvmfs source genie icetray resources test genietest py r python build buildslave morax cvmfs scientific linux cvmfs source genie icetray resources test genietest py n reporter nega cc kclark resolution duplicate ts component combo simulation summary tests taking minutes to run priority normal keywords time milestone owner melanie day type defect | 1 |
10,706 | 13,501,854,591 | IssuesEvent | 2020-09-13 05:00:05 | amor71/LiuAlgoTrader | https://api.github.com/repos/amor71/LiuAlgoTrader | closed | liquidation of non-updated stocks | in-process | sell stocks before the end of trading session, even if there are no update/trades | 1.0 | liquidation of non-updated stocks - sell stocks before the end of trading session, even if there are no update/trades | non_defect | liquidation of non updated stocks sell stocks before the end of trading session even if there are no update trades | 0 |
194,432 | 22,261,985,943 | IssuesEvent | 2022-06-10 01:56:44 | Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034 | https://api.github.com/repos/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034 | reopened | CVE-2020-29661 (High) detected in linuxlinux-4.19.239 | security vulnerability | ## CVE-2020-29661 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.239</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034/commit/19525e8c58fe9ba0d7cb0f7a1a87d31d30380de6">19525e8c58fe9ba0d7cb0f7a1a87d31d30380de6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/tty_jobctrl.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/tty_jobctrl.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A locking issue was discovered in the tty subsystem of the Linux kernel through 5.9.13. drivers/tty/tty_jobctrl.c allows a use-after-free attack against TIOCSPGRP, aka CID-54ffccbf053b.
<p>Publish Date: 2020-12-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-29661>CVE-2020-29661</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gregkh/linux/releases/tag/v5.10-rc7">https://github.com/gregkh/linux/releases/tag/v5.10-rc7</a></p>
<p>Release Date: 2020-12-09</p>
<p>Fix Resolution: v5.10-rc7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-29661 (High) detected in linuxlinux-4.19.239 - ## CVE-2020-29661 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.239</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034/commit/19525e8c58fe9ba0d7cb0f7a1a87d31d30380de6">19525e8c58fe9ba0d7cb0f7a1a87d31d30380de6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/tty_jobctrl.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/tty_jobctrl.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A locking issue was discovered in the tty subsystem of the Linux kernel through 5.9.13. drivers/tty/tty_jobctrl.c allows a use-after-free attack against TIOCSPGRP, aka CID-54ffccbf053b.
<p>Publish Date: 2020-12-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-29661>CVE-2020-29661</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gregkh/linux/releases/tag/v5.10-rc7">https://github.com/gregkh/linux/releases/tag/v5.10-rc7</a></p>
<p>Release Date: 2020-12-09</p>
<p>Fix Resolution: v5.10-rc7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers tty tty jobctrl c drivers tty tty jobctrl c vulnerability details a locking issue was discovered in the tty subsystem of the linux kernel through drivers tty tty jobctrl c allows a use after free attack against tiocspgrp aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
312,728 | 23,440,861,162 | IssuesEvent | 2022-08-15 14:45:00 | thm-mni-ii/graph-algo-ptas | https://api.github.com/repos/thm-mni-ii/graph-algo-ptas | closed | Custom badges for docs and benchmark | documentation benchmark | Currently the README badges for docs and benchmark look the same and can only be distinguished using the alt text. This should be changed. | 1.0 | Custom badges for docs and benchmark - Currently the README badges for docs and benchmark look the same and can only be distinguished using the alt text. This should be changed. | non_defect | custom badges for docs and benchmark currently the readme badges for docs and benchmark look the same and can only be distinguished using the alt text this should be changed | 0 |
21,920 | 3,587,215,040 | IssuesEvent | 2016-01-30 05:06:04 | mash99/crypto-js | https://api.github.com/repos/mash99/crypto-js | closed | Most of the hashs wont works regarding file encoding | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Create a text file containing aÿþa
2. Convert it to AINSII/UTF8/UTF8-NOBOM/UTF16/.....
3. Check the hash returned and compare with a Checksum application.
What is the expected output? What do you see instead?
The CryptoJS lib force conversion to UTF8 on the input, it will return wrong
hash...
What version of the product are you using? On what operating system?
Firefox 31 / 3.1.2
Please provide any additional information below.
How to fix it ? Simply don't use UTF8 encode on the input because you really
don't need it, you would need UTF8 encode if you wanted to "SHOW" the content,
but you really don't need to UTF8 encode the file to get a hash of it.... Or
you will get a very wrong hash.
I attached a screenshot showing how to fix it for SHA3.js file, however, you
will have the same issues in almost all if not all the others hashs
implementation. I got the same issue with SHA-256.
For SHA3 : q to e
For SHA256 : l to k
Easy to realize if you look at my screenshot.
Note that is a "temp-fix", you may need the UTF8 somewhere else in the file for
w/e reason.
```
Original issue reported on code.google.com by `contact@sundark.eu` on 3 Aug 2014 at 2:52
Attachments:
* [show.png](https://storage.googleapis.com/google-code-attachments/crypto-js/issue-139/comment-0/show.png)
| 1.0 | Most of the hashs wont works regarding file encoding - ```
What steps will reproduce the problem?
1. Create a text file containing aÿþa
2. Convert it to AINSII/UTF8/UTF8-NOBOM/UTF16/.....
3. Check the hash returned and compare with a Checksum application.
What is the expected output? What do you see instead?
The CryptoJS lib force conversion to UTF8 on the input, it will return wrong
hash...
What version of the product are you using? On what operating system?
Firefox 31 / 3.1.2
Please provide any additional information below.
How to fix it ? Simply don't use UTF8 encode on the input because you really
don't need it, you would need UTF8 encode if you wanted to "SHOW" the content,
but you really don't need to UTF8 encode the file to get a hash of it.... Or
you will get a very wrong hash.
I attached a screenshot showing how to fix it for SHA3.js file, however, you
will have the same issues in almost all if not all the others hashs
implementation. I got the same issue with SHA-256.
For SHA3 : q to e
For SHA256 : l to k
Easy to realize if you look at my screenshot.
Note that is a "temp-fix", you may need the UTF8 somewhere else in the file for
w/e reason.
```
Original issue reported on code.google.com by `contact@sundark.eu` on 3 Aug 2014 at 2:52
Attachments:
* [show.png](https://storage.googleapis.com/google-code-attachments/crypto-js/issue-139/comment-0/show.png)
| defect | most of the hashs wont works regarding file encoding what steps will reproduce the problem create a text file containing aÿþa convert it to ainsii nobom check the hash returned and compare with a checksum application what is the expected output what do you see instead the cryptojs lib force conversion to on the input it will return wrong hash what version of the product are you using on what operating system firefox please provide any additional information below how to fix it simply don t use encode on the input because you really don t need it you would need encode if you wanted to show the content but you really don t need to encode the file to get a hash of it or you will get a very wrong hash i attached a screenshot showing how to fix it for js file however you will have the same issues in almost all if not all the others hashs implementation i got the same issue with sha for q to e for l to k easy to realize if you look at my screenshot note that is a temp fix you may need the somewhere else in the file for w e reason original issue reported on code google com by contact sundark eu on aug at attachments | 1 |
114,454 | 11,848,116,756 | IssuesEvent | 2020-03-24 13:16:24 | sliu107/alpha | https://api.github.com/repos/sliu107/alpha | opened | Use ref frame for the sequnce diagram in DG | severity.Low type.DocumentationBug | For each feature, we repeat the process of `setup()` in the corresponding sequence diagram.
To reduce repetition, we can draw a master sequence diagram and use ref frame to break it to multiple sub diagrams of different feature.
| 1.0 | Use ref frame for the sequnce diagram in DG - For each feature, we repeat the process of `setup()` in the corresponding sequence diagram.
To reduce repetition, we can draw a master sequence diagram and use ref frame to break it to multiple sub diagrams of different feature.
| non_defect | use ref frame for the sequnce diagram in dg for each feature we repeat the process of setup in the corresponding sequence diagram to reduce repetition we can draw a master sequence diagram and use ref frame to break it to multiple sub diagrams of different feature | 0 |
84,055 | 16,444,453,026 | IssuesEvent | 2021-05-20 17:49:23 | smeas/Beer-and-Plunder | https://api.github.com/repos/smeas/Beer-and-Plunder | closed | Showing the different states of how much health a table has. | 2p code | **Description**
Show how damaged the table is.
**Subtasks**
- [x] Showing cracks throughout the table appearing as the health goes below some thresholds.
- [x] Five states: whole table, slightly damaged, more damaged, very damaged, completely broken. | 1.0 | Showing the different states of how much health a table has. - **Description**
Show how damaged the table is.
**Subtasks**
- [x] Showing cracks throughout the table appearing as the health goes below some thresholds.
- [x] Five states: whole table, slightly damaged, more damaged, very damaged, completely broken. | non_defect | showing the different states of how much health a table has description show how damaged the table is subtasks showing cracks throughout the table appearing as the health goes below some thresholds five states whole table slightly damaged more damaged very damaged completely broken | 0 |
460,782 | 13,218,118,126 | IssuesEvent | 2020-08-17 08:09:21 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.ebay.co.uk - site is not usable | browser-firefox engine-gecko priority-important type-no-css | <!-- @browser: Firefox 80.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:80.0) Gecko/20100101 Firefox/80.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/56520 -->
**URL**: https://www.ebay.co.uk/
**Browser / Version**: Firefox 80.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Internet Explorer
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
will not load in firefox only internet explorer
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/8/d6d86bae-8d9a-4106-a5d4-cd54a75c724a.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200807195315</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/8/6034bf33-ce5f-463f-ac97-27a135c5087e)
Submitted in the name of `@sean`
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.ebay.co.uk - site is not usable - <!-- @browser: Firefox 80.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:80.0) Gecko/20100101 Firefox/80.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/56520 -->
**URL**: https://www.ebay.co.uk/
**Browser / Version**: Firefox 80.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Internet Explorer
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
will not load in firefox only internet explorer
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/8/d6d86bae-8d9a-4106-a5d4-cd54a75c724a.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200807195315</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/8/6034bf33-ce5f-463f-ac97-27a135c5087e)
Submitted in the name of `@sean`
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_defect | site is not usable url browser version firefox operating system windows tested another browser yes internet explorer problem type site is not usable description page not loading correctly steps to reproduce will not load in firefox only internet explorer view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false submitted in the name of sean from with ❤️ | 0 |
44,619 | 12,294,002,505 | IssuesEvent | 2020-05-10 21:38:26 | IBM/CAST | https://api.github.com/repos/IBM/CAST | closed | bbcmd cancel segfault | Comp: Burst Buffer PhaseFound: Customer Sev: 3 Status: Closed Type: Defect | **Describe the bug**
When running inside of an interactive allocation and trying to cancel an ongoing transfer with multiple contributors, I got the following segfault:
```
bbcmd cancel --target=0- --jobstepid=2 --handle=1567724507
Segmentation fault (core dumped)
```
In this job, 250 hosts completed, and 6 were stuck INPROGRESS.
**Environment (please complete the following information):**
ibm-burstbuffer-tools-1.4.0-1756.ppc64le
ibm-burstbuffer-1.4.1-1774.ppc64le
ibm-burstbuffer-mn-1.4.0-1756.ppc64le
ibm-burstbuffer-tests-1.4.0-1756.ppc64le
ibm-burstbuffer-lsf-1.4.1-1774.ppc64le | 1.0 | bbcmd cancel segfault - **Describe the bug**
When running inside of an interactive allocation and trying to cancel an ongoing transfer with multiple contributors, I got the following segfault:
```
bbcmd cancel --target=0- --jobstepid=2 --handle=1567724507
Segmentation fault (core dumped)
```
In this job, 250 hosts completed, and 6 were stuck INPROGRESS.
**Environment (please complete the following information):**
ibm-burstbuffer-tools-1.4.0-1756.ppc64le
ibm-burstbuffer-1.4.1-1774.ppc64le
ibm-burstbuffer-mn-1.4.0-1756.ppc64le
ibm-burstbuffer-tests-1.4.0-1756.ppc64le
ibm-burstbuffer-lsf-1.4.1-1774.ppc64le | defect | bbcmd cancel segfault describe the bug when running inside of an interactive allocation and trying to cancel an ongoing transfer with multiple contributors i got the following segfault bbcmd cancel target jobstepid handle segmentation fault core dumped in this job hosts completed and were stuck inprogress environment please complete the following information ibm burstbuffer tools ibm burstbuffer ibm burstbuffer mn ibm burstbuffer tests ibm burstbuffer lsf | 1 |
79,945 | 29,735,760,315 | IssuesEvent | 2023-06-14 00:39:01 | openslide/openslide | https://api.github.com/repos/openslide/openslide | opened | Some images ignored in DICOM JP2K sample slides | defect | ### Operating system
Fedora 38
### Platform
x86_64
### OpenSlide version
Git main
### Slide format
DICOM
### Issue details
The JP2K [sample slides](https://openslide.cs.cmu.edu/download/openslide-testdata/DICOM/) (`CMU-1-JP2K-33005` and `JP2K-33003-1`), which were converted from Aperio slides with `com.pixelmed.convert.TIFFToDicom`, have a few issues:
1. OpenSlide ignores the highest-resolution level and all the associated images because it expects their `ImageType`s to be `ORIGINAL`, but in this case they're `DERIVED`.
1. It also ignores the Aperio thumbnail, which has an image flavor of `THUMBNAIL`. I'm not seeing that flavor defined [here](https://dicom.innolitics.com/ciods/vl-whole-slide-microscopy-image/whole-slide-microscopy-image/00080008), though.
1. If modified to accept the overview or thumbnail images, OpenSlide fails with `Not a JPEG file: starts with 0x19 0x00`. I haven't debugged this.
Should we fix these? I assume it's reasonable to accept `DERIVED` images when reading converted data (especially since we silently accept only the lower-resolution levels right now). Not sure about the others.
cc @jcupitt | 1.0 | Some images ignored in DICOM JP2K sample slides - ### Operating system
Fedora 38
### Platform
x86_64
### OpenSlide version
Git main
### Slide format
DICOM
### Issue details
The JP2K [sample slides](https://openslide.cs.cmu.edu/download/openslide-testdata/DICOM/) (`CMU-1-JP2K-33005` and `JP2K-33003-1`), which were converted from Aperio slides with `com.pixelmed.convert.TIFFToDicom`, have a few issues:
1. OpenSlide ignores the highest-resolution level and all the associated images because it expects their `ImageType`s to be `ORIGINAL`, but in this case they're `DERIVED`.
1. It also ignores the Aperio thumbnail, which has an image flavor of `THUMBNAIL`. I'm not seeing that flavor defined [here](https://dicom.innolitics.com/ciods/vl-whole-slide-microscopy-image/whole-slide-microscopy-image/00080008), though.
1. If modified to accept the overview or thumbnail images, OpenSlide fails with `Not a JPEG file: starts with 0x19 0x00`. I haven't debugged this.
Should we fix these? I assume it's reasonable to accept `DERIVED` images when reading converted data (especially since we silently accept only the lower-resolution levels right now). Not sure about the others.
cc @jcupitt | defect | some images ignored in dicom sample slides operating system fedora platform openslide version git main slide format dicom issue details the cmu and which were converted from aperio slides with com pixelmed convert tifftodicom have a few issues openslide ignores the highest resolution level and all the associated images because it expects their imagetype s to be original but in this case they re derived it also ignores the aperio thumbnail which has an image flavor of thumbnail i m not seeing that flavor defined though if modified to accept the overview or thumbnail images openslide fails with not a jpeg file starts with i haven t debugged this should we fix these i assume it s reasonable to accept derived images when reading converted data especially since we silently accept only the lower resolution levels right now not sure about the others cc jcupitt | 1 |
763,710 | 26,770,936,060 | IssuesEvent | 2023-01-31 14:05:19 | zowe/vscode-extension-for-zowe | https://api.github.com/repos/zowe/vscode-extension-for-zowe | closed | USS command fails when JCL submitted via VScode | bug Research Needed priority-medium | <!--
Before opening a new issue, please search our existing issues: https://github.com/zowe/vscode-extension-for-zowe/issues
-->
**Describe the bug**
Same USS related JCL submitted through VScode fails with RC=0012.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to job, submit mount zfs JCL as sample below
//MOUNT EXEC PGM=IKJEFT01,REGION=0M,COND=(0,LT)
//SYSEXEC DD DISP=SHR,DSN=SYS1.SBPXEXEC
//SYSTSPRT DD SYSOUT=*
//SYSTSIN DD *
PROFILE MSGID WTPMSG
oshell umask 0022; +
mkdir -p /z/mg894880/zs20/pax
MOUNT +
FILESYSTEM('MG894880.ZS20.ZFS') +
MOUNTPOINT('/z/mg894880/zs20/pax') +
MODE(RDWR) TYPE(ZFS) PARM('AGGRGROW')
//*
2. go to jobs and see error log with RC=12
**Expected behavior**
Job should complete
with RC=00
**Screenshots**



**Desktop (please complete the following information):**
OS: Windows 10
Zowe Explorer Version: 1.72
(Optional) Zowe CLI Version: 7.5.1
(Optional) Are you using Secure Credential Store?
<!-- If you have SCS installed, please read the documentation for Integrating SCS within Zowe Explorer: https://github.com/zowe/vscode-extension-for-zowe/blob/main/docs/README-Security.md -->
**Additional context**
<!-- Add any other context about the problem here. -->
| 1.0 | USS command fails when JCL submitted via VScode - <!--
Before opening a new issue, please search our existing issues: https://github.com/zowe/vscode-extension-for-zowe/issues
-->
**Describe the bug**
Same USS related JCL submitted through VScode fails with RC=0012.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to job, submit mount zfs JCL as sample below
//MOUNT EXEC PGM=IKJEFT01,REGION=0M,COND=(0,LT)
//SYSEXEC DD DISP=SHR,DSN=SYS1.SBPXEXEC
//SYSTSPRT DD SYSOUT=*
//SYSTSIN DD *
PROFILE MSGID WTPMSG
oshell umask 0022; +
mkdir -p /z/mg894880/zs20/pax
MOUNT +
FILESYSTEM('MG894880.ZS20.ZFS') +
MOUNTPOINT('/z/mg894880/zs20/pax') +
MODE(RDWR) TYPE(ZFS) PARM('AGGRGROW')
//*
2. go to jobs and see error log with RC=12
**Expected behavior**
Job should complete
with RC=00
**Screenshots**



**Desktop (please complete the following information):**
OS: Windows 10
Zowe Explorer Version: 1.72
(Optional) Zowe CLI Version: 7.5.1
(Optional) Are you using Secure Credential Store?
<!-- If you have SCS installed, please read the documentation for Integrating SCS within Zowe Explorer: https://github.com/zowe/vscode-extension-for-zowe/blob/main/docs/README-Security.md -->
**Additional context**
<!-- Add any other context about the problem here. -->
| non_defect | uss command fails when jcl submitted via vscode before opening a new issue please search our existing issues describe the bug same uss related jcl submitted through vscode fails with rc to reproduce steps to reproduce the behavior go to job submit mount zfs jcl as sample below mount exec pgm region cond lt sysexec dd disp shr dsn sbpxexec systsprt dd sysout systsin dd profile msgid wtpmsg oshell umask mkdir p z pax mount filesystem zfs mountpoint z pax mode rdwr type zfs parm aggrgrow go to jobs and see error log with rc expected behavior job should complete with rc screenshots desktop please complete the following information os windows zowe explorer version optional zowe cli version optional are you using secure credential store additional context | 0 |
10,908 | 2,622,846,742 | IssuesEvent | 2015-03-04 08:02:58 | max99x/pagemon-chrome-ext | https://api.github.com/repos/max99x/pagemon-chrome-ext | closed | Updates stops after some time | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
I use the extension to monitor about 10 channels on youtube, at intervals of
2-5 minutes. The detection works well for some time but after an interval
varying from minutes to several hours, the extension stops the upgrading. After
this point even though I try to update manually, it does not. The only way back
to work is to disable and re-enable the extension, or restart the program. When
everything is back to normal for some time.
I have had this problem for some time and realize that the main factor that
causes the failure is the heavy use of my connection. So imagine that the
failure may have something to do with attempts to update suffer with timeouts.
So to reproduce the problem is important to test in conditions of connection
unavailable for some time.
What is the expected output? What do you see instead?
Some form for the extension retry the update when realize that the desired
interval has passed.
What version of the Chrome are you using? On what operating system?
Google Chrome 7.0.517 in Win7-64. But the problem ever happens since i start to
use the extension in version 5.
Please provide any additional information below.
Im using custom mode with selector in monitor, i tried without, but the problem
remained.
```
Original issue reported on code.google.com by `gandalf....@gmail.com` on 11 Sep 2010 at 10:18 | 1.0 | Updates stops after some time - ```
What steps will reproduce the problem?
I use the extension to monitor about 10 channels on youtube, at intervals of
2-5 minutes. The detection works well for some time but after an interval
varying from minutes to several hours, the extension stops the upgrading. After
this point even though I try to update manually, it does not. The only way back
to work is to disable and re-enable the extension, or restart the program. When
everything is back to normal for some time.
I have had this problem for some time and realize that the main factor that
causes the failure is the heavy use of my connection. So imagine that the
failure may have something to do with attempts to update suffer with timeouts.
So to reproduce the problem is important to test in conditions of connection
unavailable for some time.
What is the expected output? What do you see instead?
Some form for the extension retry the update when realize that the desired
interval has passed.
What version of the Chrome are you using? On what operating system?
Google Chrome 7.0.517 in Win7-64. But the problem ever happens since i start to
use the extension in version 5.
Please provide any additional information below.
Im using custom mode with selector in monitor, i tried without, but the problem
remained.
```
Original issue reported on code.google.com by `gandalf....@gmail.com` on 11 Sep 2010 at 10:18 | defect | updates stops after some time what steps will reproduce the problem i use the extension to monitor about channels on youtube at intervals of minutes the detection works well for some time but after an interval varying from minutes to several hours the extension stops the upgrading after this point even though i try to update manually it does not the only way back to work is to disable and re enable the extension or restart the program when everything is back to normal for some time i have had this problem for some time and realize that the main factor that causes the failure is the heavy use of my connection so imagine that the failure may have something to do with attempts to update suffer with timeouts so to reproduce the problem is important to test in conditions of connection unavailable for some time what is the expected output what do you see instead some form for the extension retry the update when realize that the desired interval has passed what version of the chrome are you using on what operating system google chrome in but the problem ever happens since i start to use the extension in version please provide any additional information below im using custom mode with selector in monitor i tried without but the problem remained original issue reported on code google com by gandalf gmail com on sep at | 1 |
51,053 | 13,188,065,340 | IssuesEvent | 2020-08-13 05:27:17 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | cmake - absolute vs relative paths in *.gcno files (Trac #1891) | Migrated from Trac cmake defect | Previously we've been generating coverage on `stheno` (Ubuntu 14.04 + CVMFS). This included `gcc-4.8.4`, `cmake-2.8.12.2`, and `lcov-1.10`. We constantly received the dreaded "unexpected end of file" errors while `lcov` was reading *.gcno files. All signs pointed to upgrading `gcc` and `lcov`, so we moved coverage builds to `yoaguai` (Ubuntu 16.04) which includes `gcc-5.4`, `cmake-3.5.1` and `lcov-1.12`.
The *.gnco files generated on `stheno` contain **absolute paths** to header files, while *.gcno files generated on `yaoguai` contain **relative paths** to header files, where `CMakeFiles/` is the base. The former obviated the need for `lcov`'s `-b/--base-directory` option. The later requires that the `-b` option be updated.
This change may break combinations of older software. This ticket serves as documentation for that case.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1891">https://code.icecube.wisc.edu/ticket/1891</a>, reported by nega and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T00:26:47",
"description": "Previously we've been generating coverage on `stheno` (Ubuntu 14.04 + CVMFS). This included `gcc-4.8.4`, `cmake-2.8.12.2`, and `lcov-1.10`. We constantly received the dreaded \"unexpected end of file\" errors while `lcov` was reading *.gcno files. All signs pointed to upgrading `gcc` and `lcov`, so we moved coverage builds to `yoaguai` (Ubuntu 16.04) which includes `gcc-5.4`, `cmake-3.5.1` and `lcov-1.12`.\n\nThe *.gnco files generated on `stheno` contain **absolute paths** to header files, while *.gcno files generated on `yaoguai` contain **relative paths** to header files, where `CMakeFiles/` is the base. The former obviated the need for `lcov`'s `-b/--base-directory` option. The later requires that the `-b` option be updated.\n\nThis change may break combinations of older software. This ticket serves as documentation for that case.",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1547252807474793",
"component": "cmake",
"summary": "cmake - absolute vs relative paths in *.gcno files",
"priority": "normal",
"keywords": "",
"time": "2016-10-12T20:28:18",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | cmake - absolute vs relative paths in *.gcno files (Trac #1891) - Previously we've been generating coverage on `stheno` (Ubuntu 14.04 + CVMFS). This included `gcc-4.8.4`, `cmake-2.8.12.2`, and `lcov-1.10`. We constantly received the dreaded "unexpected end of file" errors while `lcov` was reading *.gcno files. All signs pointed to upgrading `gcc` and `lcov`, so we moved coverage builds to `yoaguai` (Ubuntu 16.04) which includes `gcc-5.4`, `cmake-3.5.1` and `lcov-1.12`.
The *.gnco files generated on `stheno` contain **absolute paths** to header files, while *.gcno files generated on `yaoguai` contain **relative paths** to header files, where `CMakeFiles/` is the base. The former obviated the need for `lcov`'s `-b/--base-directory` option. The later requires that the `-b` option be updated.
This change may break combinations of older software. This ticket serves as documentation for that case.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1891">https://code.icecube.wisc.edu/ticket/1891</a>, reported by nega and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T00:26:47",
"description": "Previously we've been generating coverage on `stheno` (Ubuntu 14.04 + CVMFS). This included `gcc-4.8.4`, `cmake-2.8.12.2`, and `lcov-1.10`. We constantly received the dreaded \"unexpected end of file\" errors while `lcov` was reading *.gcno files. All signs pointed to upgrading `gcc` and `lcov`, so we moved coverage builds to `yoaguai` (Ubuntu 16.04) which includes `gcc-5.4`, `cmake-3.5.1` and `lcov-1.12`.\n\nThe *.gnco files generated on `stheno` contain **absolute paths** to header files, while *.gcno files generated on `yaoguai` contain **relative paths** to header files, where `CMakeFiles/` is the base. The former obviated the need for `lcov`'s `-b/--base-directory` option. The later requires that the `-b` option be updated.\n\nThis change may break combinations of older software. This ticket serves as documentation for that case.",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1547252807474793",
"component": "cmake",
"summary": "cmake - absolute vs relative paths in *.gcno files",
"priority": "normal",
"keywords": "",
"time": "2016-10-12T20:28:18",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | cmake absolute vs relative paths in gcno files trac previously we ve been generating coverage on stheno ubuntu cvmfs this included gcc cmake and lcov we constantly received the dreaded unexpected end of file errors while lcov was reading gcno files all signs pointed to upgrading gcc and lcov so we moved coverage builds to yoaguai ubuntu which includes gcc cmake and lcov the gnco files generated on stheno contain absolute paths to header files while gcno files generated on yaoguai contain relative paths to header files where cmakefiles is the base the former obviated the need for lcov s b base directory option the later requires that the b option be updated this change may break combinations of older software this ticket serves as documentation for that case migrated from json status closed changetime description previously we ve been generating coverage on stheno ubuntu cvmfs this included gcc cmake and lcov we constantly received the dreaded unexpected end of file errors while lcov was reading gcno files all signs pointed to upgrading gcc and lcov so we moved coverage builds to yoaguai ubuntu which includes gcc cmake and lcov n nthe gnco files generated on stheno contain absolute paths to header files while gcno files generated on yaoguai contain relative paths to header files where cmakefiles is the base the former obviated the need for lcov s b base directory option the later requires that the b option be updated n nthis change may break combinations of older software this ticket serves as documentation for that case reporter nega cc resolution fixed ts component cmake summary cmake absolute vs relative paths in gcno files priority normal keywords time milestone owner nega type defect | 1 |
14,899 | 2,831,390,193 | IssuesEvent | 2015-05-24 15:55:02 | nobodyguy/dslrdashboard | https://api.github.com/repos/nobodyguy/dslrdashboard | closed | D800 only takes first of 5 pics in focus stacking mode | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. plug in D800
2. turn off red AF so it's white, set to take 5 images focus step 10
3. hit start focus stacking.
What is the expected output? What do you see instead?
I expect the camera to take 5 pictures but it just takes one then does nothing
What version of the product are you using? On what operating system?
vo.30.22 beta android 4.0.3
Please provide any additional information below.
I expect that when I hit start focus stacking the camera will take 5 pictures
by itself. Do I need to do something after pic # 1 is taken?
```
Original issue reported on code.google.com by `normu2...@gmail.com` on 26 Jul 2013 at 2:15 | 1.0 | D800 only takes first of 5 pics in focus stacking mode - ```
What steps will reproduce the problem?
1. plug in D800
2. turn off red AF so it's white, set to take 5 images focus step 10
3. hit start focus stacking.
What is the expected output? What do you see instead?
I expect the camera to take 5 pictures but it just takes one then does nothing
What version of the product are you using? On what operating system?
vo.30.22 beta android 4.0.3
Please provide any additional information below.
I expect that when I hit start focus stacking the camera will take 5 pictures
by itself. Do I need to do something after pic # 1 is taken?
```
Original issue reported on code.google.com by `normu2...@gmail.com` on 26 Jul 2013 at 2:15 | defect | only takes first of pics in focus stacking mode what steps will reproduce the problem plug in turn off red af so it s white set to take images focus step hit start focus stacking what is the expected output what do you see instead i expect the camera to take pictures but it just takes one then does nothing what version of the product are you using on what operating system vo beta android please provide any additional information below i expect that when i hit start focus stacking the camera will take pictures by itself do i need to do something after pic is taken original issue reported on code google com by gmail com on jul at | 1 |
161,175 | 12,532,995,734 | IssuesEvent | 2020-06-04 16:50:07 | QMCPACK/qmcpack | https://api.github.com/repos/QMCPACK/qmcpack | opened | Unit test for cuBLAS_missing_functions | enhancement testing | **Is your feature request related to a problem? Please describe.**
This batched BLAS functionality should be tested, as discussed in #2486
**Describe the solution you'd like**
Standalone unit test.
Would be good to write at the "missing function" level if practical so that the same code can be used to test other back end implementations.
**Describe alternatives you've considered**
Use vendor libraries, but they aren't implemented.
**Additional context**
| 1.0 | Unit test for cuBLAS_missing_functions - **Is your feature request related to a problem? Please describe.**
This batched BLAS functionality should be tested, as discussed in #2486
**Describe the solution you'd like**
Standalone unit test.
Would be good to write at the "missing function" level if practical so that the same code can be used to test other back end implementations.
**Describe alternatives you've considered**
Use vendor libraries, but they aren't implemented.
**Additional context**
| non_defect | unit test for cublas missing functions is your feature request related to a problem please describe this batched blas functionality should be tested as discussed in describe the solution you d like standalone unit test would be good to write at the missing function level if practical so that the same code can be used to test other back end implementations describe alternatives you ve considered use vendor libraries but they aren t implemented additional context | 0 |
46,690 | 24,675,532,421 | IssuesEvent | 2022-10-18 16:40:49 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | [Perf] Linux/arm64: 5 Regressions on 9/27/2022 12:44:07 AM | arch-arm64 area-System.Runtime tenet-performance tenet-performance-benchmarks untriaged | ### Run Information
Architecture | arm64
-- | --
OS | ubuntu 20.04
Baseline | [f8b032cd4fd5f544211d37ac25816cbd4deb67b7](https://github.com/dotnet/runtime/commit/f8b032cd4fd5f544211d37ac25816cbd4deb67b7)
Compare | [7fc8fb56327e9696ce9baa80f997e3f50675af1f](https://github.com/dotnet/runtime/commit/7fc8fb56327e9696ce9baa80f997e3f50675af1f)
Diff | [Diff](https://github.com/dotnet/runtime/compare/f8b032cd4fd5f544211d37ac25816cbd4deb67b7...7fc8fb56327e9696ce9baa80f997e3f50675af1f)
### Regressions in System.IO.Tests.Perf_File
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[CopyTo - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.CopyTo(size%3a%20512).html>) | 24.92 μs | 35.59 μs | 1.43 | 0.51 | False | | |
[CopyTo - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.CopyTo(size%3a%204096).html>) | 25.38 μs | 34.41 μs | 1.36 | 0.50 | False | | |
[ReadAllBytes - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.ReadAllBytes(size%3a%2016384).html>) | 6.31 μs | 9.05 μs | 1.43 | 0.56 | False | | |
[ReadAllBytes - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.ReadAllBytes(size%3a%20512).html>) | 4.39 μs | 6.72 μs | 1.53 | 0.46 | False | | |
[ReadAllBytes - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.ReadAllBytes(size%3a%204096).html>) | 5.04 μs | 8.14 μs | 1.62 | 0.56 | False | | |





[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/09_29_2022/refs/heads/main_arm64_ubuntu%2020.04_Regression/System.IO.Tests.Perf_File.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.IO.Tests.Perf_File*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-17dabeb9-6ca6-448d-b556-1f1217625f855669ac01cdf40dba2/e01b6325-7d76-4976-a45c-37d3dad0f58d.zip?sv=2021-08-06&se=2022-10-26T17%3A03%3A45Z&sr=c&sp=rl&sig=EMq0cmilyKiKKtFqa7t9oN9BcQvJC8%2Bvzdi6w6yolE4%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-b7534f40-895a-4caf-ba36-fe4c4bd87c72410d39468ba46ac91/6099fb78-384f-4658-9d14-60f24968399e.zip?sv=2021-08-06&se=2022-10-27T06%3A00%3A40Z&sr=c&sp=rl&sig=ssn6yyCcvfeEWg6847fyTKpglOIgUJqKKpYSzhdBb74%3D>)
### Histogram
### Edge Detector Info
[Collection Data](<https://pvscmdupload.blob.core.windows.net/autofilereport/reports/09_29_2022/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.ReadAllBytes(size- 4096).txt?sv=2018-03-28&sr=c&sig=LG%2FuZBS96wm3PV3z1k%2FG4jh6a9y8FAY3G0pihsGACTQ%3D&se=2022-10-23T12%3A15%3A03Z&sp=racwdl>)
#### System.IO.Tests.Perf_File.CopyTo(size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 35.594804657061694 > 26.09059472858944.
IsChangePoint: Marked as a change because one of 8/23/2022 6:13:58 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -13.613009703729203 (T) = (0 -36511.77374580777) / Math.Sqrt((1566868.0637298285 / (42)) + (5175036.831680358 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.44591350254795964 = (25251.69982953161 - 36511.77374580777) / 25251.69982953161 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.IO.Tests.Perf_File.CopyTo(size: 4096)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 34.40838088609899 > 26.208162399025255.
IsChangePoint: Marked as a change because one of 8/23/2022 11:36:32 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -15.495849004882762 (T) = (0 -35435.653598683675) / Math.Sqrt((872448.9772897566 / (42)) + (3165962.1549307415 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.3931895323425536 = (25434.91231885803 - 35435.653598683675) / 25434.91231885803 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.IO.Tests.Perf_File.ReadAllBytes(size: 16384)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 9.045063286511875 > 6.659189162809567.
IsChangePoint: Marked as a change because one of 8/23/2022 11:36:32 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -13.899161751698783 (T) = (0 -9163.779053103408) / Math.Sqrt((65225.11082022433 / (42)) + (304692.97658290406 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.4326243571176506 = (6396.498152202543 - 9163.779053103408) / 6396.498152202543 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.IO.Tests.Perf_File.ReadAllBytes(size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 6.71578038848459 > 4.6212490147787175.
IsChangePoint: Marked as a change because one of 8/23/2022 11:36:32 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -11.115603545421076 (T) = (0 -7349.238629170683) / Math.Sqrt((14760.799831971182 / (42)) + (538954.3655574843 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.6490678119208728 = (4456.601830466946 - 7349.238629170683) / 4456.601830466946 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.IO.Tests.Perf_File.ReadAllBytes(size: 4096)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 8.144012421752738 > 5.22310083708533.
IsChangePoint: Marked as a change because one of 8/23/2022 11:36:32 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -15.336957768962828 (T) = (0 -7967.533070545622) / Math.Sqrt((71696.64768731919 / (42)) + (282607.2829072366 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.5883932274721652 = (5016.096098083649 - 7967.533070545622) / 5016.096098083649 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
| True | [Perf] Linux/arm64: 5 Regressions on 9/27/2022 12:44:07 AM - ### Run Information
Architecture | arm64
-- | --
OS | ubuntu 20.04
Baseline | [f8b032cd4fd5f544211d37ac25816cbd4deb67b7](https://github.com/dotnet/runtime/commit/f8b032cd4fd5f544211d37ac25816cbd4deb67b7)
Compare | [7fc8fb56327e9696ce9baa80f997e3f50675af1f](https://github.com/dotnet/runtime/commit/7fc8fb56327e9696ce9baa80f997e3f50675af1f)
Diff | [Diff](https://github.com/dotnet/runtime/compare/f8b032cd4fd5f544211d37ac25816cbd4deb67b7...7fc8fb56327e9696ce9baa80f997e3f50675af1f)
### Regressions in System.IO.Tests.Perf_File
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[CopyTo - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.CopyTo(size%3a%20512).html>) | 24.92 μs | 35.59 μs | 1.43 | 0.51 | False | | |
[CopyTo - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.CopyTo(size%3a%204096).html>) | 25.38 μs | 34.41 μs | 1.36 | 0.50 | False | | |
[ReadAllBytes - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.ReadAllBytes(size%3a%2016384).html>) | 6.31 μs | 9.05 μs | 1.43 | 0.56 | False | | |
[ReadAllBytes - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.ReadAllBytes(size%3a%20512).html>) | 4.39 μs | 6.72 μs | 1.53 | 0.46 | False | | |
[ReadAllBytes - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.ReadAllBytes(size%3a%204096).html>) | 5.04 μs | 8.14 μs | 1.62 | 0.56 | False | | |





[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/09_29_2022/refs/heads/main_arm64_ubuntu%2020.04_Regression/System.IO.Tests.Perf_File.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.IO.Tests.Perf_File*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-17dabeb9-6ca6-448d-b556-1f1217625f855669ac01cdf40dba2/e01b6325-7d76-4976-a45c-37d3dad0f58d.zip?sv=2021-08-06&se=2022-10-26T17%3A03%3A45Z&sr=c&sp=rl&sig=EMq0cmilyKiKKtFqa7t9oN9BcQvJC8%2Bvzdi6w6yolE4%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-b7534f40-895a-4caf-ba36-fe4c4bd87c72410d39468ba46ac91/6099fb78-384f-4658-9d14-60f24968399e.zip?sv=2021-08-06&se=2022-10-27T06%3A00%3A40Z&sr=c&sp=rl&sig=ssn6yyCcvfeEWg6847fyTKpglOIgUJqKKpYSzhdBb74%3D>)
### Histogram
### Edge Detector Info
[Collection Data](<https://pvscmdupload.blob.core.windows.net/autofilereport/reports/09_29_2022/refs/heads/main_arm64_ubuntu 20.04/System.IO.Tests.Perf_File.ReadAllBytes(size- 4096).txt?sv=2018-03-28&sr=c&sig=LG%2FuZBS96wm3PV3z1k%2FG4jh6a9y8FAY3G0pihsGACTQ%3D&se=2022-10-23T12%3A15%3A03Z&sp=racwdl>)
#### System.IO.Tests.Perf_File.CopyTo(size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 35.594804657061694 > 26.09059472858944.
IsChangePoint: Marked as a change because one of 8/23/2022 6:13:58 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -13.613009703729203 (T) = (0 -36511.77374580777) / Math.Sqrt((1566868.0637298285 / (42)) + (5175036.831680358 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.44591350254795964 = (25251.69982953161 - 36511.77374580777) / 25251.69982953161 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.IO.Tests.Perf_File.CopyTo(size: 4096)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 34.40838088609899 > 26.208162399025255.
IsChangePoint: Marked as a change because one of 8/23/2022 11:36:32 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -15.495849004882762 (T) = (0 -35435.653598683675) / Math.Sqrt((872448.9772897566 / (42)) + (3165962.1549307415 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.3931895323425536 = (25434.91231885803 - 35435.653598683675) / 25434.91231885803 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.IO.Tests.Perf_File.ReadAllBytes(size: 16384)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 9.045063286511875 > 6.659189162809567.
IsChangePoint: Marked as a change because one of 8/23/2022 11:36:32 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -13.899161751698783 (T) = (0 -9163.779053103408) / Math.Sqrt((65225.11082022433 / (42)) + (304692.97658290406 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.4326243571176506 = (6396.498152202543 - 9163.779053103408) / 6396.498152202543 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.IO.Tests.Perf_File.ReadAllBytes(size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 6.71578038848459 > 4.6212490147787175.
IsChangePoint: Marked as a change because one of 8/23/2022 11:36:32 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -11.115603545421076 (T) = (0 -7349.238629170683) / Math.Sqrt((14760.799831971182 / (42)) + (538954.3655574843 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.6490678119208728 = (4456.601830466946 - 7349.238629170683) / 4456.601830466946 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.IO.Tests.Perf_File.ReadAllBytes(size: 4096)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 8.144012421752738 > 5.22310083708533.
IsChangePoint: Marked as a change because one of 8/23/2022 11:36:32 PM, 9/26/2022 7:45:50 PM, 9/29/2022 12:14:35 AM falls between 9/20/2022 9:08:11 AM and 9/29/2022 12:14:35 AM.
IsRegressionStdDev: Marked as regression because -15.336957768962828 (T) = (0 -7967.533070545622) / Math.Sqrt((71696.64768731919 / (42)) + (282607.2829072366 / (8))) is less than -2.010634757623041 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (42) + (8) - 2, .025) and -0.5883932274721652 = (5016.096098083649 - 7967.533070545622) / 5016.096098083649 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
| non_defect | linux regressions on am run information architecture os ubuntu baseline compare diff regressions in system io tests perf file benchmark baseline test test base test quality edge detector baseline ir compare ir ir ratio baseline etl compare etl μs μs false μs μs false μs μs false μs μs false μs μs false repro cmd git clone py performance scripts benchmarks ci py f filter system io tests perf file payloads histogram edge detector info system io tests perf file copyto size log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm pm am falls between am and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so ischangeedgedetector marked not as a regression because edge detector said so system io tests perf file copyto size log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm pm am falls between am and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so ischangeedgedetector marked not as a regression because edge detector said so system io tests perf file readallbytes size log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm pm am falls between am and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so ischangeedgedetector marked not as a regression because edge detector said so system io tests perf file readallbytes size log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm pm am falls between am and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so ischangeedgedetector marked not as a regression because edge detector said so system io tests perf file readallbytes size log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm pm am falls between am and am isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so ischangeedgedetector marked not as a regression because edge detector said so docs | 0 |
1,963 | 2,603,974,151 | IssuesEvent | 2015-02-24 19:01:04 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳疱疹的临床症状 | auto-migrated Priority-Medium Type-Defect | ```
沈阳疱疹的临床症状〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:10 | 1.0 | 沈阳疱疹的临床症状 - ```
沈阳疱疹的临床症状〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:10 | defect | 沈阳疱疹的临床症状 沈阳疱疹的临床症状〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at | 1 |
25,107 | 4,204,627,125 | IssuesEvent | 2016-06-28 10:36:49 | opencaching/opencaching-pl | https://api.github.com/repos/opencaching/opencaching-pl | closed | Cache owner can't change type of his logs | Component_CacheLog Priority_Low Type_Defect | From Forum (parlando - https://forum.opencaching.pl/viewtopic.php?p=134617#p134617):
> Z czego wynika brak możliwości zmiany komentarza na "wykonano serwis" lub inny typ wpisu?

| 1.0 | Cache owner can't change type of his logs - From Forum (parlando - https://forum.opencaching.pl/viewtopic.php?p=134617#p134617):
> Z czego wynika brak możliwości zmiany komentarza na "wykonano serwis" lub inny typ wpisu?

| defect | cache owner can t change type of his logs from forum parlando z czego wynika brak możliwości zmiany komentarza na wykonano serwis lub inny typ wpisu | 1 |
83,633 | 24,107,857,969 | IssuesEvent | 2022-09-20 08:56:47 | openvinotoolkit/openvino | https://api.github.com/repos/openvinotoolkit/openvino | closed | Build error when build python api from resource | bug category: build support_request | ##### System information (version)
<!-- Example
- OpenVINO => 2022.1
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2017
- Problem classification: Model Conversion
- Framework: TensorFlow (if applicable)
- Model name: ResNet50 (if applicable)
-->
- OpenVINO=> 2022.1
- Operating System / Platform => windows11
- Compiler => vs2022
- Problem classification =>build error
##### Detailed description
Hi, when I want to build python api from resource follow https://github.com/openvinotoolkit/openvino/wiki/BuildingForWindows.
I used `` cmake -DENABLE_TESTS=ON -DENABLE_FUNCTIONAL_TESTS=ON -DENABLE_PYTHON=ON -DPYTHON_EXECUTABLE="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\python.exe" -DPYTHON_LIBRARY="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\libs\python38.lib" -DPYTHON_INCLUDE_DIR="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\include" -DENABLE_OV_ONNX_FRONTEND=ON .. `` then `` cmake --build . --config Release -j 8 ``
but there were some wrong :
```
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(613,19): message : see previous definition of 'private: static FeStat FrontEndMockPy::m_stat' [C:\openvino\build\src\bindings\pyth
on\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): error C2491: 'FrontEndMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock
\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(429,22): message : see previous definition of 'private: static ModelStat InputModelMockPy::m_stat' [C:\openvino\build\src\bindings
\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): error C2491: 'InputModelMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mo
ck\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(91,22): message : see previous definition of 'private: static PlaceStat PlaceMockPy::m_stat' [C:\openvino\build\src\bindings\pytho
n\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): error C2491: 'PlaceMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock\mo
ck_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(17,53): error C2491: 'GetAPIVersion': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\o
penvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(21,45): error C2491: 'GetFrontEndData': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend
\openvino_mock_py_frontend.vcxproj]
```
Any help? Thanks!
##### Steps to reproduce
<!--
Describe your problem and steps you've done before you got to this point.
to add code example fence it with triple backticks and optional file extension
```.cpp
// C++ code example
```
or attach as .txt or .zip file
-->
##### Issue submission checklist
- [ ] I report the issue, it's not a question
<!--
OpenVINO team works with support forum, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenVINO documentation: https://docs.openvino.ai/
* OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit
* OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino
-->
- [ ] There is reproducer code and related data files: images, videos, models, etc.
<!--
The best reproducer -- test case for OpenVINO that we can add to the library.
-->
| 1.0 | Build error when build python api from resource - ##### System information (version)
<!-- Example
- OpenVINO => 2022.1
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2017
- Problem classification: Model Conversion
- Framework: TensorFlow (if applicable)
- Model name: ResNet50 (if applicable)
-->
- OpenVINO=> 2022.1
- Operating System / Platform => windows11
- Compiler => vs2022
- Problem classification =>build error
##### Detailed description
Hi, when I want to build python api from resource follow https://github.com/openvinotoolkit/openvino/wiki/BuildingForWindows.
I used `` cmake -DENABLE_TESTS=ON -DENABLE_FUNCTIONAL_TESTS=ON -DENABLE_PYTHON=ON -DPYTHON_EXECUTABLE="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\python.exe" -DPYTHON_LIBRARY="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\libs\python38.lib" -DPYTHON_INCLUDE_DIR="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\include" -DENABLE_OV_ONNX_FRONTEND=ON .. `` then `` cmake --build . --config Release -j 8 ``
but there were some wrong :
```
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(613,19): message : see previous definition of 'private: static FeStat FrontEndMockPy::m_stat' [C:\openvino\build\src\bindings\pyth
on\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): error C2491: 'FrontEndMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock
\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(429,22): message : see previous definition of 'private: static ModelStat InputModelMockPy::m_stat' [C:\openvino\build\src\bindings
\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): error C2491: 'InputModelMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mo
ck\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(91,22): message : see previous definition of 'private: static PlaceStat PlaceMockPy::m_stat' [C:\openvino\build\src\bindings\pytho
n\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): error C2491: 'PlaceMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock\mo
ck_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(17,53): error C2491: 'GetAPIVersion': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\o
penvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(21,45): error C2491: 'GetFrontEndData': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend
\openvino_mock_py_frontend.vcxproj]
```
Any help? Thanks!
##### Steps to reproduce
<!--
Describe your problem and steps you've done before you got to this point.
to add code example fence it with triple backticks and optional file extension
```.cpp
// C++ code example
```
or attach as .txt or .zip file
-->
##### Issue submission checklist
- [ ] I report the issue, it's not a question
<!--
OpenVINO team works with support forum, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenVINO documentation: https://docs.openvino.ai/
* OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit
* OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino
-->
- [ ] There is reproducer code and related data files: images, videos, models, etc.
<!--
The best reproducer -- test case for OpenVINO that we can add to the library.
-->
| non_defect | build error when build python api from resource system information version example openvino operating system platform windows bit compiler visual studio problem classification model conversion framework tensorflow if applicable model name if applicable openvino operating system platform compiler problem classification build error detailed description hi when i want to build python api from resource follow i used cmake denable tests on denable functional tests on denable python on dpython executable c users local admin appdata local programs python python exe dpython library c users local admin appdata local programs python libs lib dpython include dir c users local admin appdata local programs python include denable ov onnx frontend on then cmake build config release j but there were some wrong c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static festat frontendmockpy m stat c openvino build src bindings pyth on tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error frontendmockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static modelstat inputmodelmockpy m stat c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error inputmodelmockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mo ck mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static placestat placemockpy m stat c openvino build src bindings pytho n tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error placemockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mock mo ck py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error getapiversion definition of dllimport function not allowed c openvino build src bindings python tests mock mock py frontend o penvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error getfrontenddata definition of dllimport function not allowed c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj any help thanks steps to reproduce describe your problem and steps you ve done before you got to this point to add code example fence it with triple backticks and optional file extension cpp c code example or attach as txt or zip file issue submission checklist i report the issue it s not a question openvino team works with support forum stack overflow and other communities to discuss problems tickets with question without real issue statement will be closed i checked the problem with documentation faq open issues stack overflow etc and have not found solution places to check openvino documentation openvino forum openvino issue tracker stack overflow branch there is reproducer code and related data files images videos models etc the best reproducer test case for openvino that we can add to the library | 0 |
201,714 | 15,219,427,032 | IssuesEvent | 2021-02-17 19:11:13 | jupyterhub/jupyterhub | https://api.github.com/repos/jupyterhub/jupyterhub | closed | JupyterHub PAM authentication with google two-factor login | configuration needs: testing | This is for the default install of Jupyterhub. I'm setting this up on a linux AMI .
The login requires both a username and a google-authenticator time based token. When I ssh in it asks for both.
When I installed RStudio server-- it just automagically asks for the username and not the token.
When I run Jupyterhub it always fails to authenticate against the user -- I believe it's because of the two factor authnetication.
Is there a way to get Jupyterhub to just use the password (like RStudio does) or ask for both the username and password?
I tried copying the entry in `/etc/pam.d` for rstudio into jupyterhub but that wasn't effective. For reference, here's the rstudio `/etc/pam.d/rstudio` file:
```
#%PAM-1.0
auth requisite pam_succeed_if.so uid >= 500 quiet
auth required pam_unix.so nodelay
account required pam_unix.so
```
Perhaps I need to restart a service after creating `/etc/pam.d/jupyterhub`? | 1.0 | JupyterHub PAM authentication with google two-factor login - This is for the default install of Jupyterhub. I'm setting this up on a linux AMI .
The login requires both a username and a google-authenticator time based token. When I ssh in it asks for both.
When I installed RStudio server-- it just automagically asks for the username and not the token.
When I run Jupyterhub it always fails to authenticate against the user -- I believe it's because of the two factor authnetication.
Is there a way to get Jupyterhub to just use the password (like RStudio does) or ask for both the username and password?
I tried copying the entry in `/etc/pam.d` for rstudio into jupyterhub but that wasn't effective. For reference, here's the rstudio `/etc/pam.d/rstudio` file:
```
#%PAM-1.0
auth requisite pam_succeed_if.so uid >= 500 quiet
auth required pam_unix.so nodelay
account required pam_unix.so
```
Perhaps I need to restart a service after creating `/etc/pam.d/jupyterhub`? | non_defect | jupyterhub pam authentication with google two factor login this is for the default install of jupyterhub i m setting this up on a linux ami the login requires both a username and a google authenticator time based token when i ssh in it asks for both when i installed rstudio server it just automagically asks for the username and not the token when i run jupyterhub it always fails to authenticate against the user i believe it s because of the two factor authnetication is there a way to get jupyterhub to just use the password like rstudio does or ask for both the username and password i tried copying the entry in etc pam d for rstudio into jupyterhub but that wasn t effective for reference here s the rstudio etc pam d rstudio file pam auth requisite pam succeed if so uid quiet auth required pam unix so nodelay account required pam unix so perhaps i need to restart a service after creating etc pam d jupyterhub | 0 |
214,980 | 16,621,015,308 | IssuesEvent | 2021-06-03 01:00:48 | sugarlabs/musicblocks | https://api.github.com/repos/sugarlabs/musicblocks | closed | Show = Current Pitch not working | Issue-Bug WF6-Needs testing | Using Show = Current Pitch in Code like this

Results in
```
Uncaught TypeError: Cannot read property 'toString' of undefined
at Turtle.doShowText (turtle.js:845)
at Logo.processShow (logo.js:524)
at logo.js:1954
```
(And it does not print the text at the location of the mouse, as expected.)
Test File:
[Show Current Pitch.html.zip](https://github.com/sugarlabs/musicblocks/files/6565138/Show.Current.Pitch.html.zip)
| 1.0 | Show = Current Pitch not working - Using Show = Current Pitch in Code like this

Results in
```
Uncaught TypeError: Cannot read property 'toString' of undefined
at Turtle.doShowText (turtle.js:845)
at Logo.processShow (logo.js:524)
at logo.js:1954
```
(And it does not print the text at the location of the mouse, as expected.)
Test File:
[Show Current Pitch.html.zip](https://github.com/sugarlabs/musicblocks/files/6565138/Show.Current.Pitch.html.zip)
| non_defect | show current pitch not working using show current pitch in code like this results in uncaught typeerror cannot read property tostring of undefined at turtle doshowtext turtle js at logo processshow logo js at logo js and it does not print the text at the location of the mouse as expected test file | 0 |
19,654 | 3,228,794,948 | IssuesEvent | 2015-10-12 07:06:29 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Hazelcast 3.5.*: MigrationListener migrationStarted/migrationCompleted is called twice | Team: Core Type: Defect | In hazelcast version 3.4.* the ```migrationStarted/migrationCompleted``` was only called once when a partition migration was done. In 3.5.* these methods get's called twice for each partition.
That's my test:
```
import com.hazelcast.config.Config;
import com.hazelcast.config.NetworkConfig;
import com.hazelcast.core.Hazelcast;
import com.hazelcast.core.HazelcastInstance;
import com.hazelcast.core.IMap;
import com.hazelcast.core.MigrationEvent;
import com.hazelcast.core.MigrationListener;
import com.hazelcast.instance.GroupProperties;
import org.junit.Test;
import java.util.HashSet;
import java.util.Set;
public class HazelcastTest {
@Test
public void migrationListener() throws InterruptedException {
Config config = new Config();
config.setProperty(GroupProperties.PROP_VERSION_CHECK_ENABLED, "" + false);
config.setProperty(GroupProperties.PROP_PREFER_IPv4_STACK, "" + true);
config.setProperty(GroupProperties.PROP_PARTITION_COUNT, "" + 10);
NetworkConfig netConfig = new NetworkConfig();
Set<String> interfaces = new HashSet<>();
interfaces.add("127.0.0.1");
netConfig.getJoin().getMulticastConfig().setEnabled(false);
netConfig.getJoin().getTcpIpConfig().addMember("127.0.0.1").setEnabled(true);
netConfig.setPort(55555);
netConfig.getInterfaces().addInterface("127.0.0.1").setEnabled(true);
config.setNetworkConfig(netConfig);
config.getGroupConfig().setName("cluster-test" + 55555);
HazelcastInstance instance1 = Hazelcast.newHazelcastInstance(config);
IMap<String, String> map1 = instance1.getMap("foo");
map1.put("bla", "bar");
map1.put("foo", "bar");
map1.put("bar", "bar");
instance1.getPartitionService().addMigrationListener(new MigrationListener() {
@Override
public void migrationStarted(MigrationEvent migrationEvent) {
System.out.println("migration: " + migrationEvent);
}
@Override
public void migrationCompleted(MigrationEvent migrationEvent) {
System.out.println("migration: " + migrationEvent);
}
@Override
public void migrationFailed(MigrationEvent migrationEvent) {
System.out.println("migration: " + migrationEvent);
}
});
HazelcastInstance instance2 = Hazelcast.newHazelcastInstance(config);
Thread.sleep(500);
}
}
```
Output in 3.4.2:
```
migration: MigrationEvent{partitionId=0, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=0, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
```
and in 3.5.2
```
migration: MigrationEvent{partitionId=0, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=0, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=0, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=0, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
```
| 1.0 | Hazelcast 3.5.*: MigrationListener migrationStarted/migrationCompleted is called twice - In hazelcast version 3.4.* the ```migrationStarted/migrationCompleted``` was only called once when a partition migration was done. In 3.5.* these methods get's called twice for each partition.
That's my test:
```
import com.hazelcast.config.Config;
import com.hazelcast.config.NetworkConfig;
import com.hazelcast.core.Hazelcast;
import com.hazelcast.core.HazelcastInstance;
import com.hazelcast.core.IMap;
import com.hazelcast.core.MigrationEvent;
import com.hazelcast.core.MigrationListener;
import com.hazelcast.instance.GroupProperties;
import org.junit.Test;
import java.util.HashSet;
import java.util.Set;
public class HazelcastTest {
@Test
public void migrationListener() throws InterruptedException {
Config config = new Config();
config.setProperty(GroupProperties.PROP_VERSION_CHECK_ENABLED, "" + false);
config.setProperty(GroupProperties.PROP_PREFER_IPv4_STACK, "" + true);
config.setProperty(GroupProperties.PROP_PARTITION_COUNT, "" + 10);
NetworkConfig netConfig = new NetworkConfig();
Set<String> interfaces = new HashSet<>();
interfaces.add("127.0.0.1");
netConfig.getJoin().getMulticastConfig().setEnabled(false);
netConfig.getJoin().getTcpIpConfig().addMember("127.0.0.1").setEnabled(true);
netConfig.setPort(55555);
netConfig.getInterfaces().addInterface("127.0.0.1").setEnabled(true);
config.setNetworkConfig(netConfig);
config.getGroupConfig().setName("cluster-test" + 55555);
HazelcastInstance instance1 = Hazelcast.newHazelcastInstance(config);
IMap<String, String> map1 = instance1.getMap("foo");
map1.put("bla", "bar");
map1.put("foo", "bar");
map1.put("bar", "bar");
instance1.getPartitionService().addMigrationListener(new MigrationListener() {
@Override
public void migrationStarted(MigrationEvent migrationEvent) {
System.out.println("migration: " + migrationEvent);
}
@Override
public void migrationCompleted(MigrationEvent migrationEvent) {
System.out.println("migration: " + migrationEvent);
}
@Override
public void migrationFailed(MigrationEvent migrationEvent) {
System.out.println("migration: " + migrationEvent);
}
});
HazelcastInstance instance2 = Hazelcast.newHazelcastInstance(config);
Thread.sleep(500);
}
}
```
Output in 3.4.2:
```
migration: MigrationEvent{partitionId=0, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=0, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
```
and in 3.5.2
```
migration: MigrationEvent{partitionId=0, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=0, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=0, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=0, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=1, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=2, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=3, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=STARTED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
migration: MigrationEvent{partitionId=4, status=COMPLETED, oldOwner=Member [127.0.0.1]:55555 this, newOwner=Member [127.0.0.1]:55556}
```
| defect | hazelcast migrationlistener migrationstarted migrationcompleted is called twice in hazelcast version the migrationstarted migrationcompleted was only called once when a partition migration was done in these methods get s called twice for each partition that s my test import com hazelcast config config import com hazelcast config networkconfig import com hazelcast core hazelcast import com hazelcast core hazelcastinstance import com hazelcast core imap import com hazelcast core migrationevent import com hazelcast core migrationlistener import com hazelcast instance groupproperties import org junit test import java util hashset import java util set public class hazelcasttest test public void migrationlistener throws interruptedexception config config new config config setproperty groupproperties prop version check enabled false config setproperty groupproperties prop prefer stack true config setproperty groupproperties prop partition count networkconfig netconfig new networkconfig set interfaces new hashset interfaces add netconfig getjoin getmulticastconfig setenabled false netconfig getjoin gettcpipconfig addmember setenabled true netconfig setport netconfig getinterfaces addinterface setenabled true config setnetworkconfig netconfig config getgroupconfig setname cluster test hazelcastinstance hazelcast newhazelcastinstance config imap getmap foo put bla bar put foo bar put bar bar getpartitionservice addmigrationlistener new migrationlistener override public void migrationstarted migrationevent migrationevent system out println migration migrationevent override public void migrationcompleted migrationevent migrationevent system out println migration migrationevent override public void migrationfailed migrationevent migrationevent system out println migration migrationevent hazelcastinstance hazelcast newhazelcastinstance config thread sleep output in migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member and in migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status started oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member migration migrationevent partitionid status completed oldowner member this newowner member | 1 |
16,342 | 2,889,776,138 | IssuesEvent | 2015-06-13 19:01:15 | damonkohler/sl4a | https://api.github.com/repos/damonkohler/sl4a | opened | Scripts started by external activities should prompt the user to allow execution | auto-migrated Priority-Medium Type-Defect | _From @GoogleCodeExporter on May 31, 2015 11:28_
```
Not prompting the user is a potential security hole.
```
Original issue reported on code.google.com by `damonkoh...@gmail.com` on 6 Nov 2010 at 4:19
_Copied from original issue: damonkohler/android-scripting#468_ | 1.0 | Scripts started by external activities should prompt the user to allow execution - _From @GoogleCodeExporter on May 31, 2015 11:28_
```
Not prompting the user is a potential security hole.
```
Original issue reported on code.google.com by `damonkoh...@gmail.com` on 6 Nov 2010 at 4:19
_Copied from original issue: damonkohler/android-scripting#468_ | defect | scripts started by external activities should prompt the user to allow execution from googlecodeexporter on may not prompting the user is a potential security hole original issue reported on code google com by damonkoh gmail com on nov at copied from original issue damonkohler android scripting | 1 |
45,986 | 13,055,833,413 | IssuesEvent | 2020-07-30 02:52:19 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | GEANT4 tool (Trac #380) | Incomplete Migration Migrated from Trac combo core defect | Migrated from https://code.icecube.wisc.edu/ticket/380
```json
{
"status": "closed",
"changetime": "2014-11-22T18:26:26",
"description": "Having trouble installing GEANT4 and therefore verifying that g4-tankresponse builds in IceSim. geant4_4.9.5 seems to install just fine on Ubuntu 11.10, but cmake can't find it when building g4-tankresponse. It can't find \"liblist\" which just doesn't exist in 4.9.5 where cmake is looking for it. Not sure if this is a build issue or whether geant4.cmake just needs to be adapted for 4.9.5.\n\nCan't seem to build 4.9.3 either as it dies with the error:\n\n---> Verifying checksum(s) for geant4_4.9.3\nError: Checksum (md5) mismatch for G4ABLA.3.0.tar.gz\nError: Checksum (sha1) mismatch for G4ABLA.3.0.tar.gz\nError: Target com.apple.checksum returned: Unable to verify file checksums\n\nTrying 4.9.4 to see if that's any better...\n\n",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1416680786794380",
"component": "combo core",
"summary": "GEANT4 tool",
"priority": "normal",
"keywords": "",
"time": "2012-03-13T04:08:42",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | GEANT4 tool (Trac #380) - Migrated from https://code.icecube.wisc.edu/ticket/380
```json
{
"status": "closed",
"changetime": "2014-11-22T18:26:26",
"description": "Having trouble installing GEANT4 and therefore verifying that g4-tankresponse builds in IceSim. geant4_4.9.5 seems to install just fine on Ubuntu 11.10, but cmake can't find it when building g4-tankresponse. It can't find \"liblist\" which just doesn't exist in 4.9.5 where cmake is looking for it. Not sure if this is a build issue or whether geant4.cmake just needs to be adapted for 4.9.5.\n\nCan't seem to build 4.9.3 either as it dies with the error:\n\n---> Verifying checksum(s) for geant4_4.9.3\nError: Checksum (md5) mismatch for G4ABLA.3.0.tar.gz\nError: Checksum (sha1) mismatch for G4ABLA.3.0.tar.gz\nError: Target com.apple.checksum returned: Unable to verify file checksums\n\nTrying 4.9.4 to see if that's any better...\n\n",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1416680786794380",
"component": "combo core",
"summary": "GEANT4 tool",
"priority": "normal",
"keywords": "",
"time": "2012-03-13T04:08:42",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| defect | tool trac migrated from json status closed changetime description having trouble installing and therefore verifying that tankresponse builds in icesim seems to install just fine on ubuntu but cmake can t find it when building tankresponse it can t find liblist which just doesn t exist in where cmake is looking for it not sure if this is a build issue or whether cmake just needs to be adapted for n ncan t seem to build either as it dies with the error n n verifying checksum s for nerror checksum mismatch for tar gz nerror checksum mismatch for tar gz nerror target com apple checksum returned unable to verify file checksums n ntrying to see if that s any better n n reporter olivas cc resolution fixed ts component combo core summary tool priority normal keywords time milestone owner nega type defect | 1 |
41,966 | 10,728,804,872 | IssuesEvent | 2019-10-28 14:33:27 | vector-im/riot-web | https://api.github.com/repos/vector-im/riot-web | closed | Incorrect warnings on registration with no IS configured | bug defect phase:2 privacy privacy-sprint | Outdated warning text in the registration flow saying you can't reset password.
<img width="425" alt="2019-10-09 at 14 25" src="https://user-images.githubusercontent.com/279572/66485408-b0991480-eaa0-11e9-9085-a587ed9f1264.png">
<img width="797" alt="2019-10-09 at 14 25" src="https://user-images.githubusercontent.com/279572/66485438-b8f14f80-eaa0-11e9-9ef8-24d6bb5414cc.png"> | 1.0 | Incorrect warnings on registration with no IS configured - Outdated warning text in the registration flow saying you can't reset password.
<img width="425" alt="2019-10-09 at 14 25" src="https://user-images.githubusercontent.com/279572/66485408-b0991480-eaa0-11e9-9085-a587ed9f1264.png">
<img width="797" alt="2019-10-09 at 14 25" src="https://user-images.githubusercontent.com/279572/66485438-b8f14f80-eaa0-11e9-9ef8-24d6bb5414cc.png"> | defect | incorrect warnings on registration with no is configured outdated warning text in the registration flow saying you can t reset password img width alt at src img width alt at src | 1 |
21,733 | 3,549,040,977 | IssuesEvent | 2016-01-20 16:35:10 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | "Invalid constructor name" error doesn't appear after incremental analysis | area-analyzer Priority-Medium Type-Defect | To reproduce in Atom, paste this code into a file:
```
class Card {
Card.foo();
}
```
Then on the second line, delete the "C" from "Card.foo".
The following message should appear, but doesn't:
test.dart:2: Invalid constructor name
If you reanalyze sources then the error will appear.
(Found via a stress test.)
| 1.0 | "Invalid constructor name" error doesn't appear after incremental analysis - To reproduce in Atom, paste this code into a file:
```
class Card {
Card.foo();
}
```
Then on the second line, delete the "C" from "Card.foo".
The following message should appear, but doesn't:
test.dart:2: Invalid constructor name
If you reanalyze sources then the error will appear.
(Found via a stress test.)
| defect | invalid constructor name error doesn t appear after incremental analysis to reproduce in atom paste this code into a file class card card foo then on the second line delete the c from card foo the following message should appear but doesn t test dart invalid constructor name if you reanalyze sources then the error will appear found via a stress test | 1 |
26,327 | 4,676,686,703 | IssuesEvent | 2016-10-07 12:50:42 | phingofficial/phing-issues-test | https://api.github.com/repos/phingofficial/phing-issues-test | opened | PEAR packages are containing the wrong files (Trac #47) | defect Incomplete Migration Migrated from Trac | Migrated from https://www.phing.info/trac/ticket/47
```json
{
"status": "closed",
"changetime": "2009-03-22T00:02:30",
"description": "The PEAR package Phing build script is using the phing.home reserved property, which points to the currently used Phing -- and not the new Phing package.",
"reporter": "hans",
"cc": "",
"resolution": "fixed",
"_ts": "1237680150000000",
"component": "",
"summary": "PEAR packages are containing the wrong files",
"priority": "major",
"keywords": "",
"version": "2.2.0RC1",
"time": "2006-07-10T13:53:28",
"milestone": "2.2.0",
"owner": "hans",
"type": "defect"
}
```
| 1.0 | PEAR packages are containing the wrong files (Trac #47) - Migrated from https://www.phing.info/trac/ticket/47
```json
{
"status": "closed",
"changetime": "2009-03-22T00:02:30",
"description": "The PEAR package Phing build script is using the phing.home reserved property, which points to the currently used Phing -- and not the new Phing package.",
"reporter": "hans",
"cc": "",
"resolution": "fixed",
"_ts": "1237680150000000",
"component": "",
"summary": "PEAR packages are containing the wrong files",
"priority": "major",
"keywords": "",
"version": "2.2.0RC1",
"time": "2006-07-10T13:53:28",
"milestone": "2.2.0",
"owner": "hans",
"type": "defect"
}
```
| defect | pear packages are containing the wrong files trac migrated from json status closed changetime description the pear package phing build script is using the phing home reserved property which points to the currently used phing and not the new phing package reporter hans cc resolution fixed ts component summary pear packages are containing the wrong files priority major keywords version time milestone owner hans type defect | 1 |
243,875 | 20,595,068,184 | IssuesEvent | 2022-03-05 11:05:04 | Chia-Network/chia-blockchain | https://api.github.com/repos/Chia-Network/chia-blockchain | closed | [Bug] Confirmed Offer cause XCH stuck in pending state | bug stale-issue Pending regression test | ### What happened?
I accepted one offer about sell 6.25 XCH for 500 USDS. The offer is confirmed and I received the USDS. But there are 18.75 XCH stuck in the pending state forever. It's not the first time, usually delete the wallet database and wallet peers database then resync will solve this.
The original coin has 25 XCH, the new coin has 18.75 XCH (25-6.25), but it didn't collect by the wallet. I have attached suspicious logs.
### Version
light wallet 1.2.11265
### What platform are you using?
Windows
### What ui mode are you using?
CLI
### Relevant log output
```shell
2022-02-05T11:32:40.126 wallet chia.wallet.wallet : WARNING Record: {'additions': [{'amount': 18750000000000,
'parent_coin_info': '0x5f582438f924e0643795c65bcd862edc8c5533e19669e75dd4c34619ce07f5b0',
'puzzle_hash': '0xa747b85d200b57a2b6ddda60ae38299bc914bc3d9a7c61d165ea7cd08770cd9b'}],
'amount': 6250000000000,
'confirmed': False,
'confirmed_at_height': 0,
'created_at_time': 1644089551,
'fee_amount': 0,
'memos': [],
'name': '0x9b839f4e020ae69da7b7f127bb2fe5c6d229b82fbad7ae2c4ea76e6b1e386398',
'removals': [{'amount': 25000000000000,
'parent_coin_info': '0x781edbef25ad68dcc338b8857de6e31f3be307e7f001968bde580b0b30a1f95e',
'puzzle_hash': '0x84beb46a8fb40f883fdba21a411993f250ee9333e782a9d551d9993ef36f7b14'}],
'sent': 0,
'sent_to': [],
'spend_bundle': None,
'to_puzzle_hash': '0x0101010101010101010101010101010101010101010101010101010101010101',
'trade_id': '0x332256313b60d6783120105603f87d8e4e321e3f3030a65632ecc8fb7b6bd80e',
'type': 5,
'wallet_id': 1} not in mempool, []
```
| 1.0 | [Bug] Confirmed Offer cause XCH stuck in pending state - ### What happened?
I accepted one offer about sell 6.25 XCH for 500 USDS. The offer is confirmed and I received the USDS. But there are 18.75 XCH stuck in the pending state forever. It's not the first time, usually delete the wallet database and wallet peers database then resync will solve this.
The original coin has 25 XCH, the new coin has 18.75 XCH (25-6.25), but it didn't collect by the wallet. I have attached suspicious logs.
### Version
light wallet 1.2.11265
### What platform are you using?
Windows
### What ui mode are you using?
CLI
### Relevant log output
```shell
2022-02-05T11:32:40.126 wallet chia.wallet.wallet : WARNING Record: {'additions': [{'amount': 18750000000000,
'parent_coin_info': '0x5f582438f924e0643795c65bcd862edc8c5533e19669e75dd4c34619ce07f5b0',
'puzzle_hash': '0xa747b85d200b57a2b6ddda60ae38299bc914bc3d9a7c61d165ea7cd08770cd9b'}],
'amount': 6250000000000,
'confirmed': False,
'confirmed_at_height': 0,
'created_at_time': 1644089551,
'fee_amount': 0,
'memos': [],
'name': '0x9b839f4e020ae69da7b7f127bb2fe5c6d229b82fbad7ae2c4ea76e6b1e386398',
'removals': [{'amount': 25000000000000,
'parent_coin_info': '0x781edbef25ad68dcc338b8857de6e31f3be307e7f001968bde580b0b30a1f95e',
'puzzle_hash': '0x84beb46a8fb40f883fdba21a411993f250ee9333e782a9d551d9993ef36f7b14'}],
'sent': 0,
'sent_to': [],
'spend_bundle': None,
'to_puzzle_hash': '0x0101010101010101010101010101010101010101010101010101010101010101',
'trade_id': '0x332256313b60d6783120105603f87d8e4e321e3f3030a65632ecc8fb7b6bd80e',
'type': 5,
'wallet_id': 1} not in mempool, []
```
| non_defect | confirmed offer cause xch stuck in pending state what happened i accepted one offer about sell xch for usds the offer is confirmed and i received the usds but there are xch stuck in the pending state forever it s not the first time usually delete the wallet database and wallet peers database then resync will solve this the original coin has xch the new coin has xch but it didn t collect by the wallet i have attached suspicious logs version light wallet what platform are you using windows what ui mode are you using cli relevant log output shell wallet chia wallet wallet warning record additions amount parent coin info puzzle hash amount confirmed false confirmed at height created at time fee amount memos name removals amount parent coin info puzzle hash sent sent to spend bundle none to puzzle hash trade id type wallet id not in mempool | 0 |
798,402 | 28,262,999,174 | IssuesEvent | 2023-04-07 02:16:46 | AY2223S2-CS2103T-T17-3/tp | https://api.github.com/repos/AY2223S2-CS2103T-T17-3/tp | closed | [PE-D][Tester C] Command mentioned in UG is identified as invalid - add_tag command | priority.High severity.Medium | 1. The UG mentions an `add_tag` command to add a tag to a particular person. But the app identifies the command to be invalid.


<!--session: 1680242328181-7b9dbbfc-1eca-4f3d-b104-b61e0a09d562-->
<!--Version: Web v3.4.7-->
-------------
Labels: `severity.High` `type.FunctionalityBug`
original: Flash161203/ped#3 | 1.0 | [PE-D][Tester C] Command mentioned in UG is identified as invalid - add_tag command - 1. The UG mentions an `add_tag` command to add a tag to a particular person. But the app identifies the command to be invalid.


<!--session: 1680242328181-7b9dbbfc-1eca-4f3d-b104-b61e0a09d562-->
<!--Version: Web v3.4.7-->
-------------
Labels: `severity.High` `type.FunctionalityBug`
original: Flash161203/ped#3 | non_defect | command mentioned in ug is identified as invalid add tag command the ug mentions an add tag command to add a tag to a particular person but the app identifies the command to be invalid labels severity high type functionalitybug original ped | 0 |
7,072 | 7,853,654,829 | IssuesEvent | 2018-06-20 18:08:00 | odoo/odoo | https://api.github.com/repos/odoo/odoo | closed | Improve request: Leaves per hour | Services Wishlist | Impacted versions: All
Improvement request for V11. Be able to do the leave request per hour. The request per day is undo-able. At least in the Netherlands. We have employees who are working part-time and all have a different amount of leaves. When 1 hour leave for a 8 day job, the leave request is now 0,125. For a 6 hour job it is 0,166.
It just don't make sense for employees. Leave request per hour would be much more easy.
Hope this request can be done for Odoo V11. | 1.0 | Improve request: Leaves per hour - Impacted versions: All
Improvement request for V11. Be able to do the leave request per hour. The request per day is undo-able. At least in the Netherlands. We have employees who are working part-time and all have a different amount of leaves. When 1 hour leave for a 8 day job, the leave request is now 0,125. For a 6 hour job it is 0,166.
It just don't make sense for employees. Leave request per hour would be much more easy.
Hope this request can be done for Odoo V11. | non_defect | improve request leaves per hour impacted versions all improvement request for be able to do the leave request per hour the request per day is undo able at least in the netherlands we have employees who are working part time and all have a different amount of leaves when hour leave for a day job the leave request is now for a hour job it is it just don t make sense for employees leave request per hour would be much more easy hope this request can be done for odoo | 0 |
861 | 3,163,218,913 | IssuesEvent | 2015-09-20 02:40:08 | ShannonPaige/headcount | https://api.github.com/repos/ShannonPaige/headcount | opened | DR - find_by_name | spec requirement | find_by_name - returns either nil or an instance of District having done a case insensitive search | 1.0 | DR - find_by_name - find_by_name - returns either nil or an instance of District having done a case insensitive search | non_defect | dr find by name find by name returns either nil or an instance of district having done a case insensitive search | 0 |
45,881 | 5,756,919,760 | IssuesEvent | 2017-04-26 01:37:39 | magneticstain/Inquisition | https://api.github.com/repos/magneticstain/Inquisition | opened | Review Code Coverage | testing | After completion of the final new addition to Anatomize.py - Issue #10 - and writing of unit tests in Issue #13, we should review our code coverage during build tests and improve where possible. | 1.0 | Review Code Coverage - After completion of the final new addition to Anatomize.py - Issue #10 - and writing of unit tests in Issue #13, we should review our code coverage during build tests and improve where possible. | non_defect | review code coverage after completion of the final new addition to anatomize py issue and writing of unit tests in issue we should review our code coverage during build tests and improve where possible | 0 |
44,445 | 12,158,138,943 | IssuesEvent | 2020-04-26 02:08:23 | SarahAbigayle/SE310SE420FinalProjectNonGUI | https://api.github.com/repos/SarahAbigayle/SE310SE420FinalProjectNonGUI | closed | If no journal or a not yet created journal is selected, the program fails | defect | File: MainActivity.java
Line #: 71, 77, 85
These lines ask what journal that the user would like to edit. If the user enters a journal that does not yet exist (i.e. is out of bounds of the vector), or enters nothing, then the program ends with an out of bounds exception
| 1.0 | If no journal or a not yet created journal is selected, the program fails - File: MainActivity.java
Line #: 71, 77, 85
These lines ask what journal that the user would like to edit. If the user enters a journal that does not yet exist (i.e. is out of bounds of the vector), or enters nothing, then the program ends with an out of bounds exception
| defect | if no journal or a not yet created journal is selected the program fails file mainactivity java line these lines ask what journal that the user would like to edit if the user enters a journal that does not yet exist i e is out of bounds of the vector or enters nothing then the program ends with an out of bounds exception | 1 |
8,313 | 21,933,494,437 | IssuesEvent | 2022-05-23 11:54:17 | OasisLMF/OasisPlatform | https://api.github.com/repos/OasisLMF/OasisPlatform | opened | Call the oasislmf.json compatability | Bug Scalable Architecture | ## Desc
Models with older `oasislmf.json` configuration variables will fail when running in the new platform.
This is because the json is loaded directly and passed in programmatically. Add a check in the worker to convert any outdated
key names before passing the conf values to the manage function:
https://github.com/OasisLMF/OasisLMF/blob/master/oasislmf/utils/inputs.py#L39-L42
https://github.com/OasisLMF/OasisLMF/blob/master/oasislmf/_data/config_compatibility_profile.json
| 1.0 | Call the oasislmf.json compatability - ## Desc
Models with older `oasislmf.json` configuration variables will fail when running in the new platform.
This is because the json is loaded directly and passed in programmatically. Add a check in the worker to convert any outdated
key names before passing the conf values to the manage function:
https://github.com/OasisLMF/OasisLMF/blob/master/oasislmf/utils/inputs.py#L39-L42
https://github.com/OasisLMF/OasisLMF/blob/master/oasislmf/_data/config_compatibility_profile.json
| non_defect | call the oasislmf json compatability desc models with older oasislmf json configuration variables will fail when running in the new platform this is because the json is loaded directly and passed in programmatically add a check in the worker to convert any outdated key names before passing the conf values to the manage function | 0 |
68,421 | 21,664,355,179 | IssuesEvent | 2022-05-07 01:05:50 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Sending an invite to a user on a HS which is down should display a more explicit error | T-Defect P1 S-Tolerable | Created by @ Amandine:matrix.org.
Currently fails with 'Service Temporarily Unavailable'. Need to know which service...
| 1.0 | Sending an invite to a user on a HS which is down should display a more explicit error - Created by @ Amandine:matrix.org.
Currently fails with 'Service Temporarily Unavailable'. Need to know which service...
| defect | sending an invite to a user on a hs which is down should display a more explicit error created by amandine matrix org currently fails with service temporarily unavailable need to know which service | 1 |
403,051 | 27,398,993,943 | IssuesEvent | 2023-02-28 22:19:13 | FusionAuth/fusionauth-issues | https://api.github.com/repos/FusionAuth/fusionauth-issues | opened | Add 'related docs' to docs pages | documentation devrel | ## Add 'related docs' to docs pages
### Description
For someone interested in diving deeply into a topic, or just following a thread, it would be helpful to provide links to related docs on all docs pages. We'd like to do this programatically.
- [ ] Tag pages with keywords
- [ ] Build a 'related pages' component that searches for related pages by tag(s) and displays the results in some kind of callout or list
Is this a question about how to use FusionAuth? Please consider posting on the [FusionAuth forum](https://fusionauth.io/community/forum/) instead.
### Community guidelines
All issues filed in this repository must abide by the [FusionAuth community guidelines](https://fusionauth.io/community/forum/topic/1000/code-of-conduct).
| 1.0 | Add 'related docs' to docs pages - ## Add 'related docs' to docs pages
### Description
For someone interested in diving deeply into a topic, or just following a thread, it would be helpful to provide links to related docs on all docs pages. We'd like to do this programatically.
- [ ] Tag pages with keywords
- [ ] Build a 'related pages' component that searches for related pages by tag(s) and displays the results in some kind of callout or list
Is this a question about how to use FusionAuth? Please consider posting on the [FusionAuth forum](https://fusionauth.io/community/forum/) instead.
### Community guidelines
All issues filed in this repository must abide by the [FusionAuth community guidelines](https://fusionauth.io/community/forum/topic/1000/code-of-conduct).
| non_defect | add related docs to docs pages add related docs to docs pages description for someone interested in diving deeply into a topic or just following a thread it would be helpful to provide links to related docs on all docs pages we d like to do this programatically tag pages with keywords build a related pages component that searches for related pages by tag s and displays the results in some kind of callout or list is this a question about how to use fusionauth please consider posting on the instead community guidelines all issues filed in this repository must abide by the | 0 |
112,108 | 11,759,262,707 | IssuesEvent | 2020-03-13 16:55:27 | hyperledger-labs/blockchain-automation-framework | https://api.github.com/repos/hyperledger-labs/blockchain-automation-framework | closed | Add readthedocs readme for Quorum network.yaml | Release 0.4.0.0 documentation quorum | ### Description
As a user I want to access Quroum network.yaml readme on readthedocs so that write the configuration file without errors.
### Acceptance Criteria
1. Similar structure as existing readme for Fabric (docs/source/operations/fabric_networkyaml.md)
2. Add link in docs/source/operationalguide.rst
3. Add link in docs/source/faq.md
4. Add link in docs/source/gettingstarted.md
5. Add link in docs/source/operations/setting_dlt.md
| 1.0 | Add readthedocs readme for Quorum network.yaml - ### Description
As a user I want to access Quroum network.yaml readme on readthedocs so that write the configuration file without errors.
### Acceptance Criteria
1. Similar structure as existing readme for Fabric (docs/source/operations/fabric_networkyaml.md)
2. Add link in docs/source/operationalguide.rst
3. Add link in docs/source/faq.md
4. Add link in docs/source/gettingstarted.md
5. Add link in docs/source/operations/setting_dlt.md
| non_defect | add readthedocs readme for quorum network yaml description as a user i want to access quroum network yaml readme on readthedocs so that write the configuration file without errors acceptance criteria similar structure as existing readme for fabric docs source operations fabric networkyaml md add link in docs source operationalguide rst add link in docs source faq md add link in docs source gettingstarted md add link in docs source operations setting dlt md | 0 |
284,243 | 8,736,780,093 | IssuesEvent | 2018-12-11 20:32:18 | aowen87/TicketTester | https://api.github.com/repos/aowen87/TicketTester | closed | CSG meshing broken for a large problem. | bug likelihood medium priority reviewed severity high | Greg Greenman had a large CSG problem and he tried both the adaptive and multipass methods and neither version ever finished. He increased the number of processors to 64 for both models and this didn't make it any better. He reported that it is slightly larger than a problem that works ok.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1793
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Urgent
Subject: CSG meshing broken for a large problem.
Assigned to: Eric Brugger
Category:
Target version: 2.7.3
Author: Eric Brugger
Start: 04/01/2014
Due date:
% Done: 100
Estimated time: 24.0
Created: 04/01/2014 07:44 pm
Updated: 04/25/2014 07:28 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: 2.7.1
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Greg Greenman had a large CSG problem and he tried both the adaptive and multipass methods and neither version ever finished. He increased the number of processors to 64 for both models and this didn't make it any better. He reported that it is slightly larger than a problem that works ok.
Comments:
I committed revisions 23224 and 23226 to the 2.7 RC and trunk with thefollowing changes:1) I modified the multipass discretization of CSG meshes to process each domain independently if the number total number of boundary surfaces is above the internal limit. This will make if possible to handle larger CSG meshes. I also added the class vtkCSGFixedLengthBitField in order to make it easier to change the number of bits used by the multipass discretization. I added a description of the change to the release notes. This resolves #1793.A visit_vtk/full/vtkCSGFixedLengthBitField.hM avt/Database/Database/avtTransformManager.CM resources/help/en_US/relnotes2.7.3.htmlM visit_vtk/full/vtkBinaryPartitionVolumeFromVolume.CM visit_vtk/full/vtkBinaryPartitionVolumeFromVolume.hM visit_vtk/full/vtkCSGGrid.CM visit_vtk/full/vtkCSGGrid.hM visit_vtk/full/vtkMultiSplitter.CM visit_vtk/full/vtkMultiSplitter.hM visit_vtk/full/vtkVisItSplitter.CM visit_vtk/full/vtkVisItSplitter.hM visit_vtk/full/vtkVolumeFromCSGVolume.CM visit_vtk/full/vtkVolumeFromCSGVolume.h
| 1.0 | CSG meshing broken for a large problem. - Greg Greenman had a large CSG problem and he tried both the adaptive and multipass methods and neither version ever finished. He increased the number of processors to 64 for both models and this didn't make it any better. He reported that it is slightly larger than a problem that works ok.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1793
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: Urgent
Subject: CSG meshing broken for a large problem.
Assigned to: Eric Brugger
Category:
Target version: 2.7.3
Author: Eric Brugger
Start: 04/01/2014
Due date:
% Done: 100
Estimated time: 24.0
Created: 04/01/2014 07:44 pm
Updated: 04/25/2014 07:28 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: 2.7.1
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Greg Greenman had a large CSG problem and he tried both the adaptive and multipass methods and neither version ever finished. He increased the number of processors to 64 for both models and this didn't make it any better. He reported that it is slightly larger than a problem that works ok.
Comments:
I committed revisions 23224 and 23226 to the 2.7 RC and trunk with thefollowing changes:1) I modified the multipass discretization of CSG meshes to process each domain independently if the number total number of boundary surfaces is above the internal limit. This will make if possible to handle larger CSG meshes. I also added the class vtkCSGFixedLengthBitField in order to make it easier to change the number of bits used by the multipass discretization. I added a description of the change to the release notes. This resolves #1793.A visit_vtk/full/vtkCSGFixedLengthBitField.hM avt/Database/Database/avtTransformManager.CM resources/help/en_US/relnotes2.7.3.htmlM visit_vtk/full/vtkBinaryPartitionVolumeFromVolume.CM visit_vtk/full/vtkBinaryPartitionVolumeFromVolume.hM visit_vtk/full/vtkCSGGrid.CM visit_vtk/full/vtkCSGGrid.hM visit_vtk/full/vtkMultiSplitter.CM visit_vtk/full/vtkMultiSplitter.hM visit_vtk/full/vtkVisItSplitter.CM visit_vtk/full/vtkVisItSplitter.hM visit_vtk/full/vtkVolumeFromCSGVolume.CM visit_vtk/full/vtkVolumeFromCSGVolume.h
| non_defect | csg meshing broken for a large problem greg greenman had a large csg problem and he tried both the adaptive and multipass methods and neither version ever finished he increased the number of processors to for both models and this didn t make it any better he reported that it is slightly larger than a problem that works ok redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority urgent subject csg meshing broken for a large problem assigned to eric brugger category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity very serious found in version impact expected use os all support group any description greg greenman had a large csg problem and he tried both the adaptive and multipass methods and neither version ever finished he increased the number of processors to for both models and this didn t make it any better he reported that it is slightly larger than a problem that works ok comments i committed revisions and to the rc and trunk with thefollowing changes i modified the multipass discretization of csg meshes to process each domain independently if the number total number of boundary surfaces is above the internal limit this will make if possible to handle larger csg meshes i also added the class vtkcsgfixedlengthbitfield in order to make it easier to change the number of bits used by the multipass discretization i added a description of the change to the release notes this resolves a visit vtk full vtkcsgfixedlengthbitfield hm avt database database avttransformmanager cm resources help en us htmlm visit vtk full vtkbinarypartitionvolumefromvolume cm visit vtk full vtkbinarypartitionvolumefromvolume hm visit vtk full vtkcsggrid cm visit vtk full vtkcsggrid hm visit vtk full vtkmultisplitter cm visit vtk full vtkmultisplitter hm visit vtk full vtkvisitsplitter cm visit vtk full vtkvisitsplitter hm visit vtk full vtkvolumefromcsgvolume cm visit vtk full vtkvolumefromcsgvolume h | 0 |
72,420 | 8,736,142,384 | IssuesEvent | 2018-12-11 18:42:13 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | Allow Cursor customization in a TextFormField | a: fidelity a: text input f: material design framework waiting for PR to land (fixed) | As a follow up to this PR which allowed a cursor in a TextField to be customized.
https://github.com/flutter/flutter/pull/19317
I think this functionality also needs to be enabled for TextFormFields
## Steps to Reproduce
Use any TextFormField
## Logs
```
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel dev, v0.7.0, on Mac OS X 10.13.4 17E202, locale en-JP)
[✓] Android toolchain - develop for Android devices (Android SDK 28.0.0-rc2)
[✓] iOS toolchain - develop for iOS devices (Xcode 9.4.1)
[✓] Android Studio (version 3.1)
[✓] Android Studio (version 3.1)
[✓] Android Studio (version 3.1)
[✓] Connected devices (2 available)
• No issues found!
Process finished with exit code 0
```
| 1.0 | Allow Cursor customization in a TextFormField - As a follow up to this PR which allowed a cursor in a TextField to be customized.
https://github.com/flutter/flutter/pull/19317
I think this functionality also needs to be enabled for TextFormFields
## Steps to Reproduce
Use any TextFormField
## Logs
```
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel dev, v0.7.0, on Mac OS X 10.13.4 17E202, locale en-JP)
[✓] Android toolchain - develop for Android devices (Android SDK 28.0.0-rc2)
[✓] iOS toolchain - develop for iOS devices (Xcode 9.4.1)
[✓] Android Studio (version 3.1)
[✓] Android Studio (version 3.1)
[✓] Android Studio (version 3.1)
[✓] Connected devices (2 available)
• No issues found!
Process finished with exit code 0
```
| non_defect | allow cursor customization in a textformfield as a follow up to this pr which allowed a cursor in a textfield to be customized i think this functionality also needs to be enabled for textformfields steps to reproduce use any textformfield logs doctor summary to see all details run flutter doctor v flutter channel dev on mac os x locale en jp android toolchain develop for android devices android sdk ios toolchain develop for ios devices xcode android studio version android studio version android studio version connected devices available • no issues found process finished with exit code | 0 |
112,644 | 17,095,327,159 | IssuesEvent | 2021-07-09 01:01:29 | billmcchesney1/vulnerable-rust | https://api.github.com/repos/billmcchesney1/vulnerable-rust | opened | CVE-2021-32715 (Low) detected in hyper-0.13.5.crate | security vulnerability | ## CVE-2021-32715 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hyper-0.13.5.crate</b></p></summary>
<p>A fast and correct HTTP library.</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/hyper/0.13.5/download">https://crates.io/api/v1/crates/hyper/0.13.5/download</a></p>
<p>
Dependency Hierarchy:
- :x: **hyper-0.13.5.crate** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
hyper is an HTTP library for rust. hyper's HTTP/1 server code had a flaw that incorrectly parses and accepts requests with a `Content-Length` header with a prefixed plus sign, when it should have been rejected as illegal. This combined with an upstream HTTP proxy that doesn't parse such `Content-Length` headers, but forwards them, can result in "request smuggling" or "desync attacks". The flaw exists in all prior versions of hyper prior to 0.14.10, if built with `rustc` v1.5.0 or newer. The vulnerability is patched in hyper version 0.14.10. Two workarounds exist: One may reject requests manually that contain a plus sign prefix in the `Content-Length` header or ensure any upstream proxy handles `Content-Length` headers with a plus sign prefix.
<p>Publish Date: 2021-07-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32715>CVE-2021-32715</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32715">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32715</a></p>
<p>Release Date: 2021-07-07</p>
<p>Fix Resolution: hyper - 0.14.10</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Crate","packageName":"hyper","packageVersion":"0.13.5","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"hyper:0.13.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hyper - 0.14.10"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32715","vulnerabilityDetails":"hyper is an HTTP library for rust. hyper\u0027s HTTP/1 server code had a flaw that incorrectly parses and accepts requests with a `Content-Length` header with a prefixed plus sign, when it should have been rejected as illegal. This combined with an upstream HTTP proxy that doesn\u0027t parse such `Content-Length` headers, but forwards them, can result in \"request smuggling\" or \"desync attacks\". The flaw exists in all prior versions of hyper prior to 0.14.10, if built with `rustc` v1.5.0 or newer. The vulnerability is patched in hyper version 0.14.10. Two workarounds exist: One may reject requests manually that contain a plus sign prefix in the `Content-Length` header or ensure any upstream proxy handles `Content-Length` headers with a plus sign prefix.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32715","cvss3Severity":"low","cvss3Score":"3.1","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-32715 (Low) detected in hyper-0.13.5.crate - ## CVE-2021-32715 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hyper-0.13.5.crate</b></p></summary>
<p>A fast and correct HTTP library.</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/hyper/0.13.5/download">https://crates.io/api/v1/crates/hyper/0.13.5/download</a></p>
<p>
Dependency Hierarchy:
- :x: **hyper-0.13.5.crate** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
hyper is an HTTP library for rust. hyper's HTTP/1 server code had a flaw that incorrectly parses and accepts requests with a `Content-Length` header with a prefixed plus sign, when it should have been rejected as illegal. This combined with an upstream HTTP proxy that doesn't parse such `Content-Length` headers, but forwards them, can result in "request smuggling" or "desync attacks". The flaw exists in all prior versions of hyper prior to 0.14.10, if built with `rustc` v1.5.0 or newer. The vulnerability is patched in hyper version 0.14.10. Two workarounds exist: One may reject requests manually that contain a plus sign prefix in the `Content-Length` header or ensure any upstream proxy handles `Content-Length` headers with a plus sign prefix.
<p>Publish Date: 2021-07-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32715>CVE-2021-32715</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32715">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-32715</a></p>
<p>Release Date: 2021-07-07</p>
<p>Fix Resolution: hyper - 0.14.10</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Crate","packageName":"hyper","packageVersion":"0.13.5","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"hyper:0.13.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hyper - 0.14.10"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32715","vulnerabilityDetails":"hyper is an HTTP library for rust. hyper\u0027s HTTP/1 server code had a flaw that incorrectly parses and accepts requests with a `Content-Length` header with a prefixed plus sign, when it should have been rejected as illegal. This combined with an upstream HTTP proxy that doesn\u0027t parse such `Content-Length` headers, but forwards them, can result in \"request smuggling\" or \"desync attacks\". The flaw exists in all prior versions of hyper prior to 0.14.10, if built with `rustc` v1.5.0 or newer. The vulnerability is patched in hyper version 0.14.10. Two workarounds exist: One may reject requests manually that contain a plus sign prefix in the `Content-Length` header or ensure any upstream proxy handles `Content-Length` headers with a plus sign prefix.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32715","cvss3Severity":"low","cvss3Score":"3.1","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_defect | cve low detected in hyper crate cve low severity vulnerability vulnerable library hyper crate a fast and correct http library library home page a href dependency hierarchy x hyper crate vulnerable library found in base branch master vulnerability details hyper is an http library for rust hyper s http server code had a flaw that incorrectly parses and accepts requests with a content length header with a prefixed plus sign when it should have been rejected as illegal this combined with an upstream http proxy that doesn t parse such content length headers but forwards them can result in request smuggling or desync attacks the flaw exists in all prior versions of hyper prior to if built with rustc or newer the vulnerability is patched in hyper version two workarounds exist one may reject requests manually that contain a plus sign prefix in the content length header or ensure any upstream proxy handles content length headers with a plus sign prefix publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hyper isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree hyper isminimumfixversionavailable true minimumfixversion hyper basebranches vulnerabilityidentifier cve vulnerabilitydetails hyper is an http library for rust hyper http server code had a flaw that incorrectly parses and accepts requests with a content length header with a prefixed plus sign when it should have been rejected as illegal this combined with an upstream http proxy that doesn parse such content length headers but forwards them can result in request smuggling or desync attacks the flaw exists in all prior versions of hyper prior to if built with rustc or newer the vulnerability is patched in hyper version two workarounds exist one may reject requests manually that contain a plus sign prefix in the content length header or ensure any upstream proxy handles content length headers with a plus sign prefix vulnerabilityurl | 0 |
49,693 | 13,187,252,178 | IssuesEvent | 2020-08-13 02:49:44 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | [iceprod] propagate the error code for failing IceProd tasks to pass it correctly to the batch system (Trac #1882) | Incomplete Migration Migrated from Trac defect iceprod | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1882">https://code.icecube.wisc.edu/ticket/1882</a>, reported by gmerino and owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-11-03T21:01:05",
"description": "It seems that some IceProd tasks that fail in some of the final steps like contacting the DB, or staging out the output data, or other, are not reporting this error with a non-zero error code at the end of the process. As of now, they are happily exiting with exit code = 0 after the failure.\n\nReturning some non-zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them.",
"reporter": "gmerino",
"cc": "david.schultz",
"resolution": "fixed",
"_ts": "1478206865363965",
"component": "iceprod",
"summary": "[iceprod] propagate the error code for failing IceProd tasks to pass it correctly to the batch system",
"priority": "normal",
"keywords": "",
"time": "2016-10-04T07:51:37",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [iceprod] propagate the error code for failing IceProd tasks to pass it correctly to the batch system (Trac #1882) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1882">https://code.icecube.wisc.edu/ticket/1882</a>, reported by gmerino and owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-11-03T21:01:05",
"description": "It seems that some IceProd tasks that fail in some of the final steps like contacting the DB, or staging out the output data, or other, are not reporting this error with a non-zero error code at the end of the process. As of now, they are happily exiting with exit code = 0 after the failure.\n\nReturning some non-zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them.",
"reporter": "gmerino",
"cc": "david.schultz",
"resolution": "fixed",
"_ts": "1478206865363965",
"component": "iceprod",
"summary": "[iceprod] propagate the error code for failing IceProd tasks to pass it correctly to the batch system",
"priority": "normal",
"keywords": "",
"time": "2016-10-04T07:51:37",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
| defect | propagate the error code for failing iceprod tasks to pass it correctly to the batch system trac migrated from json status closed changetime description it seems that some iceprod tasks that fail in some of the final steps like contacting the db or staging out the output data or other are not reporting this error with a non zero error code at the end of the process as of now they are happily exiting with exit code after the failure n nreturning some non zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them reporter gmerino cc david schultz resolution fixed ts component iceprod summary propagate the error code for failing iceprod tasks to pass it correctly to the batch system priority normal keywords time milestone owner juancarlos type defect | 1 |
69,802 | 22,679,136,625 | IssuesEvent | 2022-07-04 08:19:09 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | No spaces listed when trying to change room join rules to space members | T-Defect | ### Steps to reproduce

### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
Linux
### Application version
Element version: 1.10.15 Olm version: 3.2.8
### How did you install the app?
Flatpak
### Homeserver
matrix.org
### Will you send logs?
Yes | 1.0 | No spaces listed when trying to change room join rules to space members - ### Steps to reproduce

### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
Linux
### Application version
Element version: 1.10.15 Olm version: 3.2.8
### How did you install the app?
Flatpak
### Homeserver
matrix.org
### Will you send logs?
Yes | defect | no spaces listed when trying to change room join rules to space members steps to reproduce outcome what did you expect what happened instead operating system linux application version element version olm version how did you install the app flatpak homeserver matrix org will you send logs yes | 1 |
1,188 | 2,601,756,240 | IssuesEvent | 2015-02-24 00:33:06 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | closed | getMinerals and getGeysers always return 0 | auto-migrated Component-Logic Priority-Critical Type-Defect Usability | ```
What steps will reproduce the problem?
1. use bwapi 3.1.1 beta
2. run exampleAIclient.exe (the one that came with bwapi)
3. launch singleplayer custom game "astral balance"
What is the expected output? What do you see instead?
i expect to see the workers harvesting minerals, but they just sit there.
after adding code to display the size of the returned mineral and geyser sets,
they show 0. Tried with complete map info, still no result. tried calling
bwapiclient::update about 5 times, then testing mineral count, still zero.
What version of the product are you using? On what operating system?
3.1.1 beta on W7P and XP Pro sp3.
Please provide any additional information below.
```
-----
Original issue reported on code.google.com by `dpershouse@gmail.com` on 21 Oct 2010 at 2:46 | 1.0 | getMinerals and getGeysers always return 0 - ```
What steps will reproduce the problem?
1. use bwapi 3.1.1 beta
2. run exampleAIclient.exe (the one that came with bwapi)
3. launch singleplayer custom game "astral balance"
What is the expected output? What do you see instead?
i expect to see the workers harvesting minerals, but they just sit there.
after adding code to display the size of the returned mineral and geyser sets,
they show 0. Tried with complete map info, still no result. tried calling
bwapiclient::update about 5 times, then testing mineral count, still zero.
What version of the product are you using? On what operating system?
3.1.1 beta on W7P and XP Pro sp3.
Please provide any additional information below.
```
-----
Original issue reported on code.google.com by `dpershouse@gmail.com` on 21 Oct 2010 at 2:46 | defect | getminerals and getgeysers always return what steps will reproduce the problem use bwapi beta run exampleaiclient exe the one that came with bwapi launch singleplayer custom game astral balance what is the expected output what do you see instead i expect to see the workers harvesting minerals but they just sit there after adding code to display the size of the returned mineral and geyser sets they show tried with complete map info still no result tried calling bwapiclient update about times then testing mineral count still zero what version of the product are you using on what operating system beta on and xp pro please provide any additional information below original issue reported on code google com by dpershouse gmail com on oct at | 1 |
7,324 | 2,610,363,497 | IssuesEvent | 2015-02-26 19:57:30 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | ALL THE BLOGS FROM TUMBLR ARE COMING BACK WITH FAILED AUTHENTICATION | auto-migrated Priority-Medium tumblr Type-Defect | ```
What's the problem?
Tumblr is not working anymore with the posts because they made some
changes...on their api...
What browser are you using?
chrome
What version of ScribeFire are you running?
4
```
-----
Original issue reported on code.google.com by `Pitat...@gmail.com` on 8 Sep 2012 at 4:11
* Merged into: #766 | 1.0 | ALL THE BLOGS FROM TUMBLR ARE COMING BACK WITH FAILED AUTHENTICATION - ```
What's the problem?
Tumblr is not working anymore with the posts because they made some
changes...on their api...
What browser are you using?
chrome
What version of ScribeFire are you running?
4
```
-----
Original issue reported on code.google.com by `Pitat...@gmail.com` on 8 Sep 2012 at 4:11
* Merged into: #766 | defect | all the blogs from tumblr are coming back with failed authentication what s the problem tumblr is not working anymore with the posts because they made some changes on their api what browser are you using chrome what version of scribefire are you running original issue reported on code google com by pitat gmail com on sep at merged into | 1 |
50,434 | 13,187,489,958 | IssuesEvent | 2020-08-13 03:34:48 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | genie 2.6.4 does not link on systems that use "--as-needed" as a default linker option (Trac #685) | Migrated from Trac defect tools/ports | This patch adds "--no-as-needed" to the linker flags in order to make Genie compile on those systems. Fixing this the right way would require a major re-work of the (horribly bad) Genie Make system.
The attached file is a replacement for the already existing src_make_Make.include.patch patch in the /files subdirectory of the genie 2.6.4 port. Making patches of patches looks awful, so I provide a full version of the file. Replacing the patch and bumping the Portfile version number should fix the problem.
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/685
, reported by claudio.kopper and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-06-03T02:06:37",
"description": "This patch adds \"--no-as-needed\" to the linker flags in order to make Genie compile on those systems. Fixing this the right way would require a major re-work of the (horribly bad) Genie Make system.\n\nThe attached file is a replacement for the already existing src_make_Make.include.patch patch in the /files subdirectory of the genie 2.6.4 port. Making patches of patches looks awful, so I provide a full version of the file. Replacing the patch and bumping the Portfile version number should fix the problem.",
"reporter": "claudio.kopper",
"cc": "",
"resolution": "fixed",
"_ts": "1338689197000000",
"component": "tools/ports",
"summary": "genie 2.6.4 does not link on systems that use \"--as-needed\" as a default linker option",
"priority": "normal",
"keywords": "genie as-needed linker flag",
"time": "2012-06-02T23:05:51",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | genie 2.6.4 does not link on systems that use "--as-needed" as a default linker option (Trac #685) - This patch adds "--no-as-needed" to the linker flags in order to make Genie compile on those systems. Fixing this the right way would require a major re-work of the (horribly bad) Genie Make system.
The attached file is a replacement for the already existing src_make_Make.include.patch patch in the /files subdirectory of the genie 2.6.4 port. Making patches of patches looks awful, so I provide a full version of the file. Replacing the patch and bumping the Portfile version number should fix the problem.
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/685
, reported by claudio.kopper and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-06-03T02:06:37",
"description": "This patch adds \"--no-as-needed\" to the linker flags in order to make Genie compile on those systems. Fixing this the right way would require a major re-work of the (horribly bad) Genie Make system.\n\nThe attached file is a replacement for the already existing src_make_Make.include.patch patch in the /files subdirectory of the genie 2.6.4 port. Making patches of patches looks awful, so I provide a full version of the file. Replacing the patch and bumping the Portfile version number should fix the problem.",
"reporter": "claudio.kopper",
"cc": "",
"resolution": "fixed",
"_ts": "1338689197000000",
"component": "tools/ports",
"summary": "genie 2.6.4 does not link on systems that use \"--as-needed\" as a default linker option",
"priority": "normal",
"keywords": "genie as-needed linker flag",
"time": "2012-06-02T23:05:51",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | genie does not link on systems that use as needed as a default linker option trac this patch adds no as needed to the linker flags in order to make genie compile on those systems fixing this the right way would require a major re work of the horribly bad genie make system the attached file is a replacement for the already existing src make make include patch patch in the files subdirectory of the genie port making patches of patches looks awful so i provide a full version of the file replacing the patch and bumping the portfile version number should fix the problem migrated from reported by claudio kopper and owned by nega json status closed changetime description this patch adds no as needed to the linker flags in order to make genie compile on those systems fixing this the right way would require a major re work of the horribly bad genie make system n nthe attached file is a replacement for the already existing src make make include patch patch in the files subdirectory of the genie port making patches of patches looks awful so i provide a full version of the file replacing the patch and bumping the portfile version number should fix the problem reporter claudio kopper cc resolution fixed ts component tools ports summary genie does not link on systems that use as needed as a default linker option priority normal keywords genie as needed linker flag time milestone owner nega type defect | 1 |
10,627 | 2,622,177,740 | IssuesEvent | 2015-03-04 00:17:25 | byzhang/leveldb | https://api.github.com/repos/byzhang/leveldb | closed | leveldb for iOS does not compile with Xcode 4.3 | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Install Xcode 4.3
2. make PLATFORM=IOS
What is the expected output? What do you see instead?
Expected output:
leveldb compiles
Actual output:
2012-05-02 21:59:53.852 defaults[69258:303]
The domain/default pair of (/Developer/Platforms/iPhoneOS.platform/version,
CFBundleShortVersionString) does not exist
2012-05-02 21:59:53.872 defaults[69260:303]
The domain/default pair of (/Developer/Platforms/iPhoneOS.platform/version,
CFBundleShortVersionString) does not exist
mkdir -p ios-x86/db/
2012-05-02 21:59:53.889 defaults[69272:303]
The domain/default pair of (/Developer/Platforms/iPhoneOS.platform/version,
CFBundleShortVersionString) does not exist
2012-05-02 21:59:53.905 defaults[69273:303]
The domain/default pair of (/Developer/Platforms/iPhoneOS.platform/version,
CFBundleShortVersionString) does not exist
/Developer/Platforms/iPhoneSimulator.platform/Developer/usr/bin/c++ -I.
-I./include -fno-builtin-memcmp -DOS_MACOSX -DLEVELDB_PLATFORM_POSIX -O2
-DNDEBUG -isysroot
/Developer/Platforms/iPhoneSimulator.platform/Developer/SDKs/iPhoneSimulator.sdk
-arch i686 -c db/builder.cc -o ios-x86/db/builder.o
make: /Developer/Platforms/iPhoneSimulator.platform/Developer/usr/bin/c++: No
such file or directory
make: *** [db/builder.o] Error 1
make: *** Waiting for unfinished jobs....
What version of the product are you using? On what operating system?
leveldb git @ 85584d497, Mac OS X 10.7.3, Xcode 4.3.2
Please provide any additional information below.
The attached patch resolves the problem.
(Note: I'm a Facebook employee. I previously contributed to Google
libphonenumber, and I filed a Corporate CLA with Google for libphonenumber
which should apply here.)
```
Original issue reported on code.google.com by `bgertzfi...@gmail.com` on 3 May 2012 at 5:04
Attachments:
* [leveldb-xcode-4.3.patch](https://storage.googleapis.com/google-code-attachments/leveldb/issue-86/comment-0/leveldb-xcode-4.3.patch)
| 1.0 | leveldb for iOS does not compile with Xcode 4.3 - ```
What steps will reproduce the problem?
1. Install Xcode 4.3
2. make PLATFORM=IOS
What is the expected output? What do you see instead?
Expected output:
leveldb compiles
Actual output:
2012-05-02 21:59:53.852 defaults[69258:303]
The domain/default pair of (/Developer/Platforms/iPhoneOS.platform/version,
CFBundleShortVersionString) does not exist
2012-05-02 21:59:53.872 defaults[69260:303]
The domain/default pair of (/Developer/Platforms/iPhoneOS.platform/version,
CFBundleShortVersionString) does not exist
mkdir -p ios-x86/db/
2012-05-02 21:59:53.889 defaults[69272:303]
The domain/default pair of (/Developer/Platforms/iPhoneOS.platform/version,
CFBundleShortVersionString) does not exist
2012-05-02 21:59:53.905 defaults[69273:303]
The domain/default pair of (/Developer/Platforms/iPhoneOS.platform/version,
CFBundleShortVersionString) does not exist
/Developer/Platforms/iPhoneSimulator.platform/Developer/usr/bin/c++ -I.
-I./include -fno-builtin-memcmp -DOS_MACOSX -DLEVELDB_PLATFORM_POSIX -O2
-DNDEBUG -isysroot
/Developer/Platforms/iPhoneSimulator.platform/Developer/SDKs/iPhoneSimulator.sdk
-arch i686 -c db/builder.cc -o ios-x86/db/builder.o
make: /Developer/Platforms/iPhoneSimulator.platform/Developer/usr/bin/c++: No
such file or directory
make: *** [db/builder.o] Error 1
make: *** Waiting for unfinished jobs....
What version of the product are you using? On what operating system?
leveldb git @ 85584d497, Mac OS X 10.7.3, Xcode 4.3.2
Please provide any additional information below.
The attached patch resolves the problem.
(Note: I'm a Facebook employee. I previously contributed to Google
libphonenumber, and I filed a Corporate CLA with Google for libphonenumber
which should apply here.)
```
Original issue reported on code.google.com by `bgertzfi...@gmail.com` on 3 May 2012 at 5:04
Attachments:
* [leveldb-xcode-4.3.patch](https://storage.googleapis.com/google-code-attachments/leveldb/issue-86/comment-0/leveldb-xcode-4.3.patch)
| defect | leveldb for ios does not compile with xcode what steps will reproduce the problem install xcode make platform ios what is the expected output what do you see instead expected output leveldb compiles actual output defaults the domain default pair of developer platforms iphoneos platform version cfbundleshortversionstring does not exist defaults the domain default pair of developer platforms iphoneos platform version cfbundleshortversionstring does not exist mkdir p ios db defaults the domain default pair of developer platforms iphoneos platform version cfbundleshortversionstring does not exist defaults the domain default pair of developer platforms iphoneos platform version cfbundleshortversionstring does not exist developer platforms iphonesimulator platform developer usr bin c i i include fno builtin memcmp dos macosx dleveldb platform posix dndebug isysroot developer platforms iphonesimulator platform developer sdks iphonesimulator sdk arch c db builder cc o ios db builder o make developer platforms iphonesimulator platform developer usr bin c no such file or directory make error make waiting for unfinished jobs what version of the product are you using on what operating system leveldb git mac os x xcode please provide any additional information below the attached patch resolves the problem note i m a facebook employee i previously contributed to google libphonenumber and i filed a corporate cla with google for libphonenumber which should apply here original issue reported on code google com by bgertzfi gmail com on may at attachments | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.